SIGIR ’94


Book Description

Information retrieval (IR) is becoming an increasingly important area as scientific, business and government organisations take up the notion of "information superhighways" and make available their full text databases for searching. Containing a selection of 35 papers taken from the 17th Annual SIGIR Conference held in Dublin, Ireland in July 1994, the book addresses basic research and provides an evaluation of information retrieval techniques in applications. Topics covered include text categorisation, indexing, user modelling, IR theory and logic, natural language processing, statistical and probabilistic models of information retrieval systems, routing, passage retrieval, and implementation issues.




First Text Retrieval Conference (TREC-1)


Book Description

Held in Gaithersburg, MD, Nov. 4-6, 1992. Evaluates new technologies in information retrieval. Numerous graphs, tables and charts.










Information Retrieval Evaluation


Book Description

Evaluation has always played a major role in information retrieval, with the early pioneers such as Cyril Cleverdon and Gerard Salton laying the foundations for most of the evaluation methodologies in use today. The retrieval community has been extremely fortunate to have such a well-grounded evaluation paradigm during a period when most of the human language technologies were just developing. This lecture has the goal of explaining where these evaluation methodologies came from and how they have continued to adapt to the vastly changed environment in the search engine world today. The lecture starts with a discussion of the early evaluation of information retrieval systems, starting with the Cranfield testing in the early 1960s, continuing with the Lancaster "user" study for MEDLARS, and presenting the various test collection investigations by the SMART project and by groups in Britain. The emphasis in this chapter is on the how and the why of the various methodologies developed. The second chapter covers the more recent "batch" evaluations, examining the methodologies used in the various open evaluation campaigns such as TREC, NTCIR (emphasis on Asian languages), CLEF (emphasis on European languages), INEX (emphasis on semi-structured data), etc. Here again the focus is on the how and why, and in particular on the evolving of the older evaluation methodologies to handle new information access techniques. This includes how the test collection techniques were modified and how the metrics were changed to better reflect operational environments. The final chapters look at evaluation issues in user studies -- the interactive part of information retrieval, including a look at the search log studies mainly done by the commercial search engines. Here the goal is to show, via case studies, how the high-level issues of experimental design affect the final evaluations. Table of Contents: Introduction and Early History / "Batch" Evaluation Since 1992 / Interactive Evaluation / Conclusion




Search Result Diversification


Book Description

This primer reviews the published literature on search result diversification. In particular, it discusses the motivations for diversifying the search results for an ambiguous query and provides a formal definition of the search result diversification problem. In addition, it describes the most successful approaches in the literature for producing and evaluating diversity in multiple search domains.







Big Data Analytics


Book Description

This book constitutes the refereed proceedings of the 6th International Conference on Big Data analytics, BDA 2018, held in Warangal, India, in December 2018. The 29 papers presented in this volume were carefully reviewed and selected from 93 submissions. The papers are organized in topical sections named: big data analytics: vision and perspectives; financial data analytics and data streams; web and social media data; big data systems and frameworks; predictive analytics in healthcare and agricultural domains; and machine learning and pattern mining.