Gradient descent optimization of smoothed information retrieval metrics

被引:168
作者
Chapelle, Olivier [1 ]
Wu, Mingrui [1 ]
机构
[1] Yahoo Labs, Santa Clara, CA USA
来源
INFORMATION RETRIEVAL | 2010年 / 13卷 / 03期
关键词
Learning to rank; Gradient descent; Annealing;
D O I
10.1007/s10791-009-9110-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
080201 [机械制造及其自动化];
摘要
Most ranking algorithms are based on the optimization of some loss functions, such as the pairwise loss. However, these loss functions are often different from the criteria that are adopted to measure the quality of the web page ranking results. To overcome this problem, we propose an algorithm which aims at directly optimizing popular measures such as the Normalized Discounted Cumulative Gain and the Average Precision. The basic idea is to minimize a smooth approximation of these measures with gradient descent. Crucial to this kind of approach is the choice of the smoothing factor. We provide various theoretical analysis on that choice and propose an annealing algorithm to iteratively minimize a less and less smoothed approximation of the measure of interest. Results on the Letor benchmark datasets show that the proposed algorithm achieves state-of-the-art performances.
引用
收藏
页码:216 / 235
页数:20
相关论文
共 33 条
[1]
[Anonymous], ADV NEURAL INFORM PR
[2]
[Anonymous], SIGIR
[3]
[Anonymous], 2003, Journal of machine learning research
[4]
[Anonymous], 2000, ADV LARGE MARGIN CLA
[5]
[Anonymous], 2008, P 25 INT C MACH LEAR, DOI [10.1145/1390156.1390306, DOI 10.1145/1390156.1390306]
[6]
[Anonymous], 2005, P INT C MACH LEARN
[7]
[Anonymous], P 17 ANN INT ACM SIG
[8]
[Anonymous], 1994, INTRO CONJUGATE GRAD
[9]
[Anonymous], 2002, KDD 2002
[10]
Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690