Information Retrieval Evaluation (Synthesis Lectures on Information Concepts, Retrieval, and Services)
暫譯: 資訊檢索評估(資訊概念、檢索與服務綜合講座)
Donna Harman
- 出版商: Morgan & Claypool
- 出版日期: 2011-06-03
- 售價: $1,450
- 貴賓價: 9.5 折 $1,378
- 語言: 英文
- 頁數: 120
- 裝訂: Paperback
- ISBN: 1598299719
- ISBN-13: 9781598299717
海外代購書籍(需單獨結帳)
相關主題
商品描述
The lecture starts with a discussion of the early evaluation of information retrieval systems, starting with the Cranfield testing in the early 1960s, continuing with the Lancaster "user" study for MEDLARS, and presenting the various test collection investigations by the SMART project and by groups in Britain. The emphasis in this chapter is on the how and the why of the various methodologies developed. The second chapter covers the more recent "batch" evaluations, examining the methodologies used in the various open evaluation campaigns such as TREC, NTCIR (emphasis on Asian languages), CLEF (emphasis on European languages), INEX (emphasis on semi-structured data), etc. Here again the focus is on the how and why, and in particular on the evolving of the older evaluation methodologies to handle new information access techniques. This includes how the test collection techniques were modified and how the metrics were changed to better reflect operational environments. The final chapters look at evaluation issues in user studies -- the interactive part of information retrieval, including a look at the search log studies mainly done by the commercial search engines. Here the goal is to show, via case studies, how the high-level issues of experimental design affect the final evaluations.
Table of Contents: Introduction and Early History / "Batch" Evaluation Since 1992 / Interactive Evaluation / Conclusion
商品描述(中文翻譯)
評估在資訊檢索中一直扮演著重要角色,早期的先驅如 Cyril Cleverdon 和 Gerard Salton 為當今大多數評估方法奠定了基礎。在大多數人類語言技術剛剛發展的時期,檢索社群非常幸運地擁有如此扎實的評估範式。本次講座的目標是解釋這些評估方法的來源,以及它們如何持續適應當今搜尋引擎世界的巨大變化。
講座首先討論資訊檢索系統的早期評估,從1960年代初的 Cranfield 測試開始,接著是 Lancaster 對 MEDLARS 的「使用者」研究,並介紹 SMART 專案及英國各組織的各種測試集合調查。本章的重點在於各種方法論的發展過程及其原因。第二章涵蓋了較近期的「批次」評估,檢視在各種公開評估活動中使用的方法論,例如 TREC、NTCIR(重點在亞洲語言)、CLEF(重點在歐洲語言)、INEX(重點在半結構化數據)等。這裡同樣著重於方法論的發展及其原因,特別是如何使舊有的評估方法論演變以應對新的資訊存取技術。這包括測試集合技術的修改以及指標的變更,以更好地反映操作環境。最後幾章探討使用者研究中的評估問題——資訊檢索的互動部分,包括對商業搜尋引擎主要進行的搜尋日誌研究的探討。這裡的目標是通過案例研究展示實驗設計的高層次問題如何影響最終評估。
目錄:介紹與早期歷史 / 自1992年以來的「批次」評估 / 互動評估 / 結論