Preference-based learning to rank

Nir Ailon, Mehryar Mohri

Research output: Contribution to journalArticle

Abstract

This paper presents an efficient preference-based ranking algorithm running in two stages. In the first stage, the algorithm learns a preference function defined over pairs, as in a standard binary classification problem. In the second stage, it makes use of that preference function to produce an accurate ranking, thereby reducing the learning problem of ranking to binary classification. This reduction is based on the familiar QuickSort and guarantees an expected pairwise misranking loss of at most twice that of the binary classifier derived in the first stage. Furthermore, in the important special case of bipartite ranking, the factor of two in loss is reduced to one. This improved bound also applies to the regret achieved by our ranking and that of the binary classifier obtained. Our algorithm is randomized, but we prove a lower bound for any deterministic reduction of ranking to binary classification showing that randomization is necessary to achieve our guarantees. This, and a recent result by Balcan et al., who show a regret bound of two for a deterministic algorithm in the bipartite case, suggest a trade-off between achieving low regret and determinism in this context. Our reduction also admits an improved running time guarantee with respect to that deterministic algorithm. In particular, the number of calls to the preference function in the reduction is improved from Ω(n 2) to O(nlog∈n). In addition, when the top k ranked elements only are required (k≪n), as in many applications in information extraction or search engine design, the time complexity of our algorithm can be further reduced to O(klog∈k+n). Our algorithm is thus practical for realistic applications where the number of points to rank exceeds several thousand.

Original languageEnglish (US)
Pages (from-to)189-211
Number of pages23
JournalMachine Learning
Volume80
Issue number2-3
DOIs
StatePublished - Sep 2010

Fingerprint

Classifiers
Search engines

Keywords

  • Learning to rank
  • Machine learning reductions
  • ROC

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software

Cite this

Preference-based learning to rank. / Ailon, Nir; Mohri, Mehryar.

In: Machine Learning, Vol. 80, No. 2-3, 09.2010, p. 189-211.

Research output: Contribution to journalArticle

Ailon, Nir ; Mohri, Mehryar. / Preference-based learning to rank. In: Machine Learning. 2010 ; Vol. 80, No. 2-3. pp. 189-211.
@article{877000f06211478a880f5a60fe5ce188,
title = "Preference-based learning to rank",
abstract = "This paper presents an efficient preference-based ranking algorithm running in two stages. In the first stage, the algorithm learns a preference function defined over pairs, as in a standard binary classification problem. In the second stage, it makes use of that preference function to produce an accurate ranking, thereby reducing the learning problem of ranking to binary classification. This reduction is based on the familiar QuickSort and guarantees an expected pairwise misranking loss of at most twice that of the binary classifier derived in the first stage. Furthermore, in the important special case of bipartite ranking, the factor of two in loss is reduced to one. This improved bound also applies to the regret achieved by our ranking and that of the binary classifier obtained. Our algorithm is randomized, but we prove a lower bound for any deterministic reduction of ranking to binary classification showing that randomization is necessary to achieve our guarantees. This, and a recent result by Balcan et al., who show a regret bound of two for a deterministic algorithm in the bipartite case, suggest a trade-off between achieving low regret and determinism in this context. Our reduction also admits an improved running time guarantee with respect to that deterministic algorithm. In particular, the number of calls to the preference function in the reduction is improved from Ω(n 2) to O(nlog∈n). In addition, when the top k ranked elements only are required (k≪n), as in many applications in information extraction or search engine design, the time complexity of our algorithm can be further reduced to O(klog∈k+n). Our algorithm is thus practical for realistic applications where the number of points to rank exceeds several thousand.",
keywords = "Learning to rank, Machine learning reductions, ROC",
author = "Nir Ailon and Mehryar Mohri",
year = "2010",
month = "9",
doi = "10.1007/s10994-010-5176-9",
language = "English (US)",
volume = "80",
pages = "189--211",
journal = "Machine Learning",
issn = "0885-6125",
publisher = "Springer Netherlands",
number = "2-3",

}

TY - JOUR

T1 - Preference-based learning to rank

AU - Ailon, Nir

AU - Mohri, Mehryar

PY - 2010/9

Y1 - 2010/9

N2 - This paper presents an efficient preference-based ranking algorithm running in two stages. In the first stage, the algorithm learns a preference function defined over pairs, as in a standard binary classification problem. In the second stage, it makes use of that preference function to produce an accurate ranking, thereby reducing the learning problem of ranking to binary classification. This reduction is based on the familiar QuickSort and guarantees an expected pairwise misranking loss of at most twice that of the binary classifier derived in the first stage. Furthermore, in the important special case of bipartite ranking, the factor of two in loss is reduced to one. This improved bound also applies to the regret achieved by our ranking and that of the binary classifier obtained. Our algorithm is randomized, but we prove a lower bound for any deterministic reduction of ranking to binary classification showing that randomization is necessary to achieve our guarantees. This, and a recent result by Balcan et al., who show a regret bound of two for a deterministic algorithm in the bipartite case, suggest a trade-off between achieving low regret and determinism in this context. Our reduction also admits an improved running time guarantee with respect to that deterministic algorithm. In particular, the number of calls to the preference function in the reduction is improved from Ω(n 2) to O(nlog∈n). In addition, when the top k ranked elements only are required (k≪n), as in many applications in information extraction or search engine design, the time complexity of our algorithm can be further reduced to O(klog∈k+n). Our algorithm is thus practical for realistic applications where the number of points to rank exceeds several thousand.

AB - This paper presents an efficient preference-based ranking algorithm running in two stages. In the first stage, the algorithm learns a preference function defined over pairs, as in a standard binary classification problem. In the second stage, it makes use of that preference function to produce an accurate ranking, thereby reducing the learning problem of ranking to binary classification. This reduction is based on the familiar QuickSort and guarantees an expected pairwise misranking loss of at most twice that of the binary classifier derived in the first stage. Furthermore, in the important special case of bipartite ranking, the factor of two in loss is reduced to one. This improved bound also applies to the regret achieved by our ranking and that of the binary classifier obtained. Our algorithm is randomized, but we prove a lower bound for any deterministic reduction of ranking to binary classification showing that randomization is necessary to achieve our guarantees. This, and a recent result by Balcan et al., who show a regret bound of two for a deterministic algorithm in the bipartite case, suggest a trade-off between achieving low regret and determinism in this context. Our reduction also admits an improved running time guarantee with respect to that deterministic algorithm. In particular, the number of calls to the preference function in the reduction is improved from Ω(n 2) to O(nlog∈n). In addition, when the top k ranked elements only are required (k≪n), as in many applications in information extraction or search engine design, the time complexity of our algorithm can be further reduced to O(klog∈k+n). Our algorithm is thus practical for realistic applications where the number of points to rank exceeds several thousand.

KW - Learning to rank

KW - Machine learning reductions

KW - ROC

UR - http://www.scopus.com/inward/record.url?scp=77955661709&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=77955661709&partnerID=8YFLogxK

U2 - 10.1007/s10994-010-5176-9

DO - 10.1007/s10994-010-5176-9

M3 - Article

VL - 80

SP - 189

EP - 211

JO - Machine Learning

JF - Machine Learning

SN - 0885-6125

IS - 2-3

ER -