Affiliation:
1. Tianjin University, Tianjin, China
2. Tianjin University and The Open University, Milton Keynes, United Kingdom
3. Lanzhou University, Lanzhou, Gansu, China
Abstract
In many research and application areas, such as information retrieval and machine learning, we often encounter dealing with a probability distribution that is mixed by one distribution that is relevant to our task in hand and the other that is irrelevant and that we want to get rid of. Thus, it is an essential problem to separate the irrelevant distribution from the mixture distribution. This article is focused on the application in Information Retrieval, where relevance feedback is a widely used technique to build a refined query model based on a set of feedback documents. However, in practice, the relevance feedback set, even provided by users explicitly or implicitly, is often a mixture of relevant and irrelevant documents. Consequently, the resultant query model (typically a term distribution) is often a mixture rather than a true relevance term distribution, leading to a negative impact on the retrieval performance. To tackle this problem, we recently proposed a Distribution Separation Method (DSM), which aims to approximate the true relevance distribution by separating a seed irrelevance distribution from the mixture one. While it achieved a promising performance in an empirical evaluation with simulated explicit irrelevance feedback data, it has not been deployed in the scenario where one should automatically obtain the irrelevance feedback data. In this article, we propose a substantial extension of the basic DSM from two perspectives: developing a further regularization framework and deploying DSM in the automatic irrelevance feedback scenario. Specifically, in order to avoid the output distribution of DSM drifting away from the true relevance distribution when the quality of seed irrelevant distribution (as the input to DSM) is not guaranteed, we propose a DSM regularization framework to constrain the estimation for the relevance distribution. This regularization framework includes three algorithms, each corresponding to a regularization strategy incorporated in the objective function of DSM. In addition, we exploit DSM in automatic (i.e., pseudo) irrelevance feedback, by automatically detecting the seed irrelevant documents via three different document reranking methods. We have carried out extensive experiments based on various TREC datasets, in order to systematically evaluate the proposed methods. The experimental results demonstrate the effectiveness of our proposed approaches in comparison with various strong baselines.
Funder
Chinese National Program on Key Basic Research Project
Natural Science Foundation of China
Publisher
Association for Computing Machinery (ACM)
Subject
Artificial Intelligence,Theoretical Computer Science
Reference42 articles.
1. Nasreen Abdul-Jaleel James Allan W. Bruce Croft Fernando Diaz Leah Larkey Xiaoyan Li Donald Metzler Mark D. Smucker Trevor Strohman Howard Turtle and Courtney Wade. 2004. Umass at trec 2004: Novelty and hard. In TREC’04. Nasreen Abdul-Jaleel James Allan W. Bruce Croft Fernando Diaz Leah Larkey Xiaoyan Li Donald Metzler Mark D. Smucker Trevor Strohman Howard Turtle and Courtney Wade. 2004. Umass at trec 2004: Novelty and hard. In TREC’04.
2. Fast Outlier Detection in High Dimensional Spaces
3. Paul N. Bennett Ryen W. White Wei Chu Susan T. Dumais Peter Bailey Fedor Borisyuk and Xiaoyuan Cui. 2012. Modeling the impact of short-and long-term behavior on search personalization. In SIGIR’12. ACM 185--194. 10.1145/2348283.2348312 Paul N. Bennett Ryen W. White Wei Chu Susan T. Dumais Peter Bailey Fedor Borisyuk and Xiaoyuan Cui. 2012. Modeling the impact of short-and long-term behavior on search personalization. In SIGIR’12. ACM 185--194. 10.1145/2348283.2348312
4. Chris Buckley and Gerard Salton. 1995. Optimization of relevance feedback weights. In SIGIR’95. ACM 351--357. 10.1145/215206.215383 Chris Buckley and Gerard Salton. 1995. Optimization of relevance feedback weights. In SIGIR’95. ACM 351--357. 10.1145/215206.215383
5. From ranknet to lambdarank to lambdamart: An overview;Burges Christopher J. C.;Learning,2010
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Modeling Queries with Contextual Snippets for Information Retrieval;ACM Transactions on Intelligent Systems and Technology;2018-07-31