Published January 1, 2016 | Version v1
Conference paper Open

Risk-Sensitive Evaluation and Learning to Rank using Multiple Baselines

  • 1. Sitki Kocman Univ Mugla, Mugla, Turkey
  • 2. Univ Glasgow, Glasgow, Lanark, Scotland

Description

A robust retrieval system ensures that user experience is not damaged by the presence of poorly-performing queries. Such robustness can be measured by risk-sensitive evaluation measures, which assess the extent to which a system performs worse than a given baseline system. However, using a particular, single system as the baseline suffers from the fact that retrieval performance highly varies among IR systems across topics. Thus, a single system would in general fail in providing enough information about the real baseline performance for every topic under consideration, and hence it would in general fail in measuring the real risk associated with any given system. Based upon the Chi-squared statistic, we propose a new measure Z(Risk) that exhibits more promise since it takes into account multiple baselines when measuring risk, and a derivative measure called GeoRisk, which enhances Z(Risk) by also taking into account the overall magnitude of effectiveness. This paper demonstrates the bene fits of Z(Risk) and GeoRisk upon TREC data, and how to exploit GeoRisk for risk-sensitive learning to rank, thereby making use of multiple baselines within the learning objective function to obtain effective yet risk-averse/robust ranking systems. Experiments using 10,000 topics from the MSLR learning to rank dataset demonstrate the efficacy of the proposed Chi-square statistic-based objective function.

Files

bib-54385b34-b129-4a04-99a8-7ecac2fdf304.txt

Files (240 Bytes)

Name Size Download all
md5:124b97bc38dc209b472e50b6f1cde697
240 Bytes Preview Download