Affiliation:
1. IBM Research Labs, Haifa, Israel
2. Technion, Haifa, Israel
Abstract
We study the process in which search engines with segmented indices serve queries. In particular, we investigate the number of result pages that search engines should prepare during the query processing phase.Search engine users have been observed to browse through very few pages of results for queries that they submit. This behavior of users suggests that prefetching many results upon processing an initial query is not efficient, since most of the prefetched results will not be requested by the user who initiated the search. However, a policy that abandons result prefetching in favor of retrieving just the first page of search results might not make optimal use of system resources either.We argue that for a certain behavior of users, engines should prefetch a constant number of result pages per query. We define a concrete query processing model for search engines with segmented indices, and analyze the cost of such prefetching policies. Based on these costs, we show how to determine the constant that optimizes the prefetching policy. Our results are mostly applicable to
local index
partitions of the inverted files, but are also applicable to processing short queries in
global index
architectures.
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications
Cited by
10 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Web Search Result Caching and Prefetching;Encyclopedia of Database Systems;2018
2. A machine learning approach for result caching in web search engines;Information Processing & Management;2017-07
3. Web Search Result Caching and Prefetching;Encyclopedia of Database Systems;2016
4. MetaSurfer: a new metasearch engine based on FAHP and modified EOWA operator;International Journal of System Assurance Engineering and Management;2014-10-22
5. Second Chance;ACM Transactions on the Web;2013-12