Abstract

We cast the ranking problem as (1) multiple classification (“Mc”) (2) multiple ordinal classification, which lead to computationally tractable learning algorithms for relevance ranking in Web search. We consider the DCG criterion (discounted cumulative gain), a standard quality measure in information retrieval. Our approach is motivated by the fact that perfect classifications result in perfect DCG scores and the DCG errors are bounded by classification errors. We propose using the Expected Relevance to convert class probabilities into ranking scores. The class probabilities are learned using a gradient boosting tree algorithm. Evaluations on large-scale datasets show that our approach can improve LambdaRank [5] and the regressions-based ranker [6], in terms of the (normalized) DCG scores. An efficient implementation of the boosting tree algorithm is also presented. 1

Keywords

Boosting (machine learning)Gradient boostingLearning to rankRanking (information retrieval)Artificial intelligenceMachine learningComputer scienceBounded functionMathematicsPattern recognition (psychology)Data miningRandom forest

Affiliated Institutions

Related Publications

Publication Info

Year
2007
Type
article
Volume
20
Pages
897-904
Citations
434
Access
Closed

External Links

Citation Metrics

434
OpenAlex

Cite This

Ping Li, Qiang Wu, Christopher J. C. Burges (2007). McRank: Learning to Rank Using Multiple Classification and Gradient Boosting. , 20 , 897-904.