Affiliation:
1. Kobenhavns Universitet Copenhagen , Copenhagen , Denmark
2. Aalborg Universitet , Aalborg , Denmark
Abstract
Abstract
When submitting queries to information retrieval (IR) systems, users often have the option of specifying which, if any, of the query terms are heavily dependent on each other and should be treated as a fixed phrase, for instance by placing them between quotes.In addition to such cases where users specify term dependence, automatic ways also exist for IR systems to detect dependent terms in queries. Most IR systems use both user and algorithmic approaches. It is not however clear whether and to what extent user-defined term dependence agrees with algorithmic estimates of term dependence, nor which of the two may fetch higher performance gains. Simply put, is it better to trust users or the system to detect term dependence in queries? To answer this question, we experiment with 101 crowdsourced search engine users and 334 queries (52 train and 282 test TREC queries) and we record 10 assessments per query. We find that (i) user assessments of term dependence differ significantly from algorithmic assessments of term dependence (their overlap is approximately 30%); (ii) there is little agreement among users about term dependence in queries, and this disagreement increases as queries become longer; (iii) the potential retrieval gain that can be fetched by treating term dependence (both user- and system-defined) over a bag of words baseline is reserved to a small subset (approximately 8%) of the queries, and is much higher for low-depth than deep precision measures. Points (ii) and (iii) constitute novel insights into term dependence.
Subject
Geology,Ocean Engineering,Water Science and Technology
Reference35 articles.
1. Bendersky, M., Croft, W. B., & Smith, D. A. (2011, June). Joint annotation of search queries. Paper presented at the Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies-Volume 1 (pp. 102-111). Association for Computational Linguistics.
2. Bergsma, S., & Wang, Q. I. (2007). Learning noun phrase query segmentation. Paper presented at the Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, June 28-30, 2007, Prague, Czech Republic (pp.819-826). DBLP.
3. Blanco, R., Halpin, H., Herzig, D. M., Mika, P., Pound, J., Thompson, H. S., & Tran Duc, T. (2011, July). Repeatable and reliable search system evaluation using crowdsourcing. In Proceedings of the 34th international ACM SIGIR conference on Research and development in Information Retrieval (pp. 923-932). ACM. http://dx.doi.org/10.1145/2009916.2010039
4. Fagan, J. L. (1989). The effectiveness of a nonsyntactic approach to automatic phrase indexing for document retrieval.Journal of the American Society for Information Science, 40(2), 115-132. http://dx.doi.org/10.1002/(SICI)1097-4571(198903)40:2%3C115::AID-ASI6%3E3.0.CO;2-B
5. Fenichel, C. H. (1981). Online searching: Measures that discriminate among users with different types of experiences. Journal of the Association for Information Science and Technology, 32(1), 23-32. http://dx.doi.org/10.1002/asi.4630320104
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献