Knowledge Acquisition Passage Retrieval: Corpus, Ranking Models, and Evaluation Resources

Open Access
Authors
Publication date 2024
Host editors
  • L. Goeuriot
  • P. Mulhem
  • G. Quénot
  • D. Schwab
  • G.M. Di Nunzio
  • L. Soulier
  • P. Galuščáková
  • A. García Seco de Herrera
  • G. Faggioli
  • N. Ferro
Book title Experimental IR Meets Multilinguality, Multimodality, and Interaction
Book subtitle 15th International Conference of the CLEF Association, CLEF 2024, Grenoble, France, September 9–12, 2024 : proceedings
ISBN
  • 9783031717352
ISBN (electronic)
  • 9783031717369
Series Lecture Notes in Computer Science
Event 2024 Conference and Labs of the Evaluation Forum
Volume | Issue number I
Pages (from-to) 74-87
Number of pages 14
Publisher Cham: Springer
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Knowledge acquisition passage retrieval is a task that captures search in a learning or educational setting, where users seek to find key educational information within their field of interest. Traditional relevance assessments used in ad-hoc retrieval tasks tend to focus on topical relevance, often overlooking other factors such as the “informativeness” of the retrieved educational content in relation to the user’s knowledge acquisition needs. This paper presents a new test collection for the knowledge acquisition passage retrieval (KAPR) task, constructed using the data and production systems of a large academic publisher containing: First, a set of search requests covering key educational topics/concepts across different science domains. Second, a large corpus of passages extracted from review (survey) articles published in over 2, 700 journals as well as the content of 43, 000 books published in a wide range of science domains. Third, relevance assessments on both topical relevance as well as informativeness, reflecting the task-specific relevance. This resource enables direct evaluation of the user’s utility of the retrieved content and provides a comparative analysis with traditional topical relevance. Our findings indicate a strong correlation between relevance and informativeness, although the distribution of these labels varies per domain.
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-3-031-71736-9_3
Downloads
978-3-031-71736-9_3 (Final published version)
Permalink to this page
Back