Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset
碩士 === 輔仁大學 === 資訊管理學系碩士班 === 102 === In this study, we used TREC-6 document and topic sets which, through Text Retrieval Conference, served as our data set. We used different information retrieval models (i.e., TF_IDF, BM25, and Hiemstra_LM) and different tags (with title “T” and with title and des...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2014
|
Online Access: | http://ndltd.ncl.edu.tw/handle/hypstr |
id |
ndltd-TW-102FJU00396018 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-102FJU003960182019-05-15T21:23:34Z http://ndltd.ncl.edu.tw/handle/hypstr Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset Terrier 資訊檢索平台實作與評估互動式資訊搜尋:以TREC 6為例 CHIOU, SHAO-SYUAN 邱紹軒 碩士 輔仁大學 資訊管理學系碩士班 102 In this study, we used TREC-6 document and topic sets which, through Text Retrieval Conference, served as our data set. We used different information retrieval models (i.e., TF_IDF, BM25, and Hiemstra_LM) and different tags (with title “T” and with title and description “T+D”) by Terrier IR Command Line Mode to retrieve documents from TREC-6. Our aim was to determine which model would achieve higher precision under different experimental settings. We then tested the query expansion function to investigate the effects of the length of topics for different IR models. Furthermore, we created an interactive information retrieval (IR) interface using a modified query expansion class provided by Terrier IR, and we selected two topics from the TREC-6 document and topic sets. The study involved 18 users who were recruited as our evaluation subjects. We aimed to investigate (1) the effectiveness of the three models in terms of the number of retrieved relevant documents by subject as well as average precision; (2) the effectiveness of the interactive IR with the help of term suggestions or without term suggestions. We also recorded and analyzed the subjects’ search behaviors using Morae software. The results indicate that in Command Line Mode, when we used title with description “T+D” but no query expansion, the language model of Hiemstra_LM showed the best performance. Interestingly, when we used query expansion, the Hiemstra_LM model demonstrated the worst performance. This suggests that this model as a Terrier IR tool is not fit to process topics with more words. In our proposed interactive IR interface, regardless of whether term suggestions were given, subjects who used the Hiemstra_LM model achieved the best performance. Moreover, when subjects used the interface with the term suggestion function, they were also able to find a greater number of relevant documents compared to the interface without term suggestion. Finally, according to our recording videos, we found that when using the interface without the term suggestion function, the Hiemstra_LM model let users find relevant documents more quickly and accurately. In conclusion, the Hiemstra_LM model is capable of building good index terms and suggesting proper terms for helping users achieve better interactive IR. Wu, I-Chin 吳怡瑾 2014 學位論文 ; thesis 83 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 輔仁大學 === 資訊管理學系碩士班 === 102 === In this study, we used TREC-6 document and topic sets which, through Text Retrieval Conference, served as our data set. We used different information retrieval models (i.e., TF_IDF, BM25, and Hiemstra_LM) and different tags (with title “T” and with title and description “T+D”) by Terrier IR Command Line Mode to retrieve documents from TREC-6. Our aim was to determine which model would achieve higher precision under different experimental settings. We then tested the query expansion function to investigate the effects of the length of topics for different IR models. Furthermore, we created an interactive information retrieval (IR) interface using a modified query expansion class provided by Terrier IR, and we selected two topics from the TREC-6 document and topic sets. The study involved 18 users who were recruited as our evaluation subjects. We aimed to investigate (1) the effectiveness of the three models in terms of the number of retrieved relevant documents by subject as well as average precision; (2) the effectiveness of the interactive IR with the help of term suggestions or without term suggestions. We also recorded and analyzed the subjects’ search behaviors using Morae software.
The results indicate that in Command Line Mode, when we used title with description “T+D” but no query expansion, the language model of Hiemstra_LM showed the best performance. Interestingly, when we used query expansion, the Hiemstra_LM model demonstrated the worst performance. This suggests that this model as a Terrier IR tool is not fit to process topics with more words. In our proposed interactive IR interface, regardless of whether term suggestions were given, subjects who used the Hiemstra_LM model achieved the best performance. Moreover, when subjects used the interface with the term suggestion function, they were also able to find a greater number of relevant documents compared to the interface without term suggestion. Finally, according to our recording videos, we found that when using the interface without the term suggestion function, the Hiemstra_LM model let users find relevant documents more quickly and accurately. In conclusion, the Hiemstra_LM model is capable of building good index terms and suggesting proper terms for helping users achieve better interactive IR.
|
author2 |
Wu, I-Chin |
author_facet |
Wu, I-Chin CHIOU, SHAO-SYUAN 邱紹軒 |
author |
CHIOU, SHAO-SYUAN 邱紹軒 |
spellingShingle |
CHIOU, SHAO-SYUAN 邱紹軒 Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset |
author_sort |
CHIOU, SHAO-SYUAN |
title |
Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset |
title_short |
Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset |
title_full |
Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset |
title_fullStr |
Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset |
title_full_unstemmed |
Interactive Information Retrieval Evaluations by Implementation of Terrier IR Platform: TREC 6 Dataset |
title_sort |
interactive information retrieval evaluations by implementation of terrier ir platform: trec 6 dataset |
publishDate |
2014 |
url |
http://ndltd.ncl.edu.tw/handle/hypstr |
work_keys_str_mv |
AT chioushaosyuan interactiveinformationretrievalevaluationsbyimplementationofterrierirplatformtrec6dataset AT qiūshàoxuān interactiveinformationretrievalevaluationsbyimplementationofterrierirplatformtrec6dataset AT chioushaosyuan terrierzīxùnjiǎnsuǒpíngtáishízuòyǔpínggūhùdòngshìzīxùnsōuxúnyǐtrec6wèilì AT qiūshàoxuān terrierzīxùnjiǎnsuǒpíngtáishízuòyǔpínggūhùdòngshìzīxùnsōuxúnyǐtrec6wèilì |
_version_ |
1719113330314969088 |