User:GModena (WMF)/Notes/MLR
Appearance
Literature
- LTR workshop at SIGIR 2015
- https://staff.fnwi.uva.nl/m.derijke/talks-etc/online-learning-to-rank-tutorial/
- Discusses "Good" vs "Bad" abandonment signals
- Beyond DCG: User Behavior as a Predictor of a Successful Search, A Hassan, R Jones, KL Klinkner WSDM'10
- https://www.wsdm-conference.org/2010/proceedings/docs/p221.pdf
- Web search engines are traditionally evaluated in terms of the relevance of web pages to individual queries. However, relevance of web pages does not tell the complete picture, since an individual query may represent only a piece of the user’s information need and users may have different information needs underlying the same queries. We address the problem of predicting user search goal success by modeling user behavior. We show empirically that user behavior alone can give an accurate picture of the success of the user’s web search goals, without considering the relevance of the documents displayed. In fact, our experiments show that models using user behavior are more predictive of goal success than those using document relevance. We build novel sequence models incorporating time distributions for this task and our experiments show that the sequence and time distribution models are more accurate than static models based on user behavior, or predictions based on document relevance.
- Potential Good Abandonment Prediction, Yandex (2012)
- https://wwwconference.org/wp-content/uploads/2025/01/p485.pdf
- Abandonment rate is one of the most broadly used online user satisfaction metrics. In this paper we discuss the notion of potential good abandonment, i.e. queries that may potentially result in user satisfaction without the need to click on search results (if search engine result page contains enough details to satisfy the user information need). We show, that we can train a classifier which is able to distinguish between potential good and bad abandonments with rather good results compared to our baseline. As a case study we show how to apply these ideas to IR evaluation and introduce a new metric for A/B-testing — Bad Abandonment Rate
- A General Framework for Counterfactual Learning-to-Rank, Aman Agarwal, Kenta Takatsu, Ivan Zaitsev, and Thorsten Joachims. 2019.
- https://www.cs.cornell.edu/people/tj/publications/agarwal_etal_19b.pdf
- Implicit feedback (e.g., click, dwell time) is an attractive source of training data for Learning-to-Rank, but its naive use leads to learning results that are distorted by presentation bias. For the special case of optimizing average rank for linear ranking functions, however, the recently developed SVM-PropRank method has shown that counterfactual inference techniques can be used to provably overcome the distorting efect of presentation bias. Going beyond this special case, this paper provides a general and theoretically rigorous framework for counterfactual learning-to-rank that enables unbiased training for a broad class of additive ranking metrics (e.g., Discounted Cumulative Gain (DCG)) as well as a broad class of models (e.g., deep networks). Speciically, we derive a relaxation for propensity-weighted rank-based metrics which is subdiferentiable and thus suitable for gradient-based optimization. We demonstrate the efectiveness of this general approach by instantiating two new learning methods. One is a new type of unbiased SVM that optimizes DCG ś called SVM PropDCG ś, and we show how the resulting optimization problem can be solved via the Convex Concave Procedure (CCP). The other is Deep PropDCG, where the ranking function can be an arbitrary deep network. In addition to the theoretical support, we empirically ind that SVM PropDCG signiicantly outperforms existing linear rankers in terms of DCG. Moreover, the ability to train non-linear ranking functions via Deep PropDCG further improves performance.