Meta-evaluation of online and offline web search evaluation metrics
Chen, Ye; Zhou, Ke; Liu, Yiqun; Zhang, Min; Ma, Shaoping
As in most information retrieval (IR) studies, evaluation plays an essential part in Web search research. Both offline and online evaluation metrics are adopted in measuring the performance of search engines. Offline metrics are usually based on relevance judgments of query-document pairs from assessors while online metrics exploit the user behavior data, such as clicks, collected from search engines to compare search algorithms. Although both types of IR evaluation metrics have achieved success, to what extent can they predict user satisfaction still remains under-investigated. To shed light on this research question, we meta-evaluate a series of existing online and offline metrics to study how well they infer actual search user satisfaction in different search scenarios. We find that both types of evaluation metrics significantly correlate with user satisfaction while they reflect satisfaction from different perspectives for different search tasks. Offline metrics better align with user satisfaction in homogeneous search (i.e. ten blue links) whereas online metrics outperform when vertical results are federated. Finally, we also propose to incorporate mouse hover information into existing online evaluation metrics, and empirically show that they better align with search user satisfaction than click-based online metrics.
|Publication Date||Aug 7, 2017|
|Peer Reviewed||Peer Reviewed|
|Book Title||Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval - SIGIR '17|
|APA6 Citation||Chen, Y., Zhou, K., Liu, Y., Zhang, M., & Ma, S. (2017). Meta-evaluation of online and offline web search evaluation metrics. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval - SIGIR '17doi:10.1145/3077136.3080804|
|Copyright Statement||Copyright information regarding this work can be found at the following address: http://eprints.nottingh.../end_user_agreement.pdf|
Copyright information regarding this work can be found at the following address: http://eprints.nottingham.ac.uk/end_user_agreement.pdf
You might also like
Detecting collusive spamming activities in community question answering
Does document relevance affect the searcher's perception 0f time?
Palimpsest: improving assisted curation of loco-specific literature
Predicting pre-click quality for native advertisements
Incorporating non-sequential behavior into click models