This paper presents a preliminary experimentation study using the CLEF 2017 eHealth Task 2 collection for evaluating the effectiveness of different indexing methodologies of documents and query parsing techniques. Furthermore, it is an attempt to advance and share the efforts of observing the characteristics and helpfulness of various methodologies for indexing PubMed documents and for different topic parsing techniques to produce queries. For this purpose, my research includes experimentation with different document indexing methodologies, by utilising existing tools, such as the Lucene4IR (L4IR) information retrieval system, the Technology Assisted Reviews for Empirical Medicine tool for parsing topics of the CLEF collection and the TREC evaluation tool to appraise system's performance. The results showed that including a greater number of fields to the PubMed indexer of L4IR is a decisive factor for the retrieval effectiveness of L4IR.
翻译:本文件介绍了利用2017年CLEF电子保健任务2收集的资料进行的初步实验研究,以评价文件和查询技术的不同索引方法的有效性;此外,还试图推动和分享努力,以观察各种方法的特性和有用性,为PubMed文件和不同专题分类技术编制索引,以产生查询;为此目的,我的研究包括试验不同的文件索引方法,方法是利用现有工具,如Luceen4IR(L4IR)信息检索系统、用于CLEF收集专题分类的“经验医学技术辅助审查”工具以及TREC评估系统业绩的评估工具;结果显示,将更多领域纳入L4IR的PubMed索引器是L4IR检索有效性的决定性因素。