72
Advanced Topics in Information Retrieval Learning to Rank Vinay Setty Jannik Strötgen [email protected] [email protected] ATIR – July 14, 2016

Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

Advanced Topics in Information Retrieval

Learning to Rank

Vinay Setty Jannik Strötgen

[email protected] [email protected]

ATIR – July 14, 2016

Page 2: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Before we startoral exams

July 28, the full dayif you have any temporal constraints, let us know

Q-A sessions – suggestionThursday, July 21: Vinay and “his topics”Monday, July 25: Jannik and “his topics”

c© Jannik Strötgen – ATIR-10 2 / 72

Page 3: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

Advanced Topics in Information Retrieval

Learning to Rank

Vinay Setty Jannik Strötgen

[email protected] [email protected]

ATIR – July 14, 2016

Page 4: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

The Beginning of LeToR

learning to rank (LeToR)

builds on established methods from Machine Learningallows different targets derived from different kinds of userinputactive area of research for past 10 – 15 yearsearly work already (end of) 1980s (e.g., Fuhr 1989)

c© Jannik Strötgen – ATIR-10 4 / 72

Page 5: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

The Beginning of LeToR

why wasn’t LeToR successful earlier?

IR and ML communities were not very connectedsometimes ideas take timelimited training data

– it was hard to gather (real-world) test collection queries andrelevance judgments that are representative of real user needsand judgments on returned documents

– this changed in academia and industry

poor machine learning techniquesinsufficient customization to IR problemnot enough features for ML to show value

c© Jannik Strötgen – ATIR-10 5 / 72

Page 6: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

The Beginning of LeToR

standard ranking based onterm frequency / inverse document frequencyOkapi BM25language models...

traditional ranking functions in IR exploit very few features

standard approach to combine different featuresnormalize features (zero mean, unit standard deviation)feature combination function (typically: weighted sum)tune weights (either manually or exhaustively via grid search)

traditional ranking functions easy to tune

c© Jannik Strötgen – ATIR-10 6 / 72

Page 7: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Why learning to rank nowadays?

c© Jannik Strötgen – ATIR-10 7 / 72

Page 8: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Why learning to rank?

modern systems use a huge number of features(especially Web search engines)

textual relevance (e.g., using LM, Okapi BM25)proximity of query keywords in document contentlink-based importance (e.g., determined using PageRank)depth of URL (top-level page vs. leaf page)spamminess (e.g., determine using SpamRank)host importance (e.g., determined using host-level PageRank)readability of contentlocation and time of the userlocation and time of documents...

c© Jannik Strötgen – ATIR-10 8 / 72

Page 9: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Why learning to rank?

high creativity in feature engineering taskquery word in color on page?number of images on page?URL contains ~?number of (out) links on a page?page edit recencypage length

learning to rank makes combining features more systematic

c© Jannik Strötgen – ATIR-10 9 / 72

Page 10: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Outline I

1 LeToR Framework

2 Modeling Approaches

3 Gathering User Feedback

4 Evaluating Learning to Rank

5 Learning-to-Rank for Temporal IR

6 Learning-to-Rank – Beyond Search

c© Jannik Strötgen – ATIR-10 10 / 72

Page 11: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Outline

1 LeToR Framework

2 Modeling Approaches

3 Gathering User Feedback

4 Evaluating Learning to Rank

5 Learning-to-Rank for Temporal IR

6 Learning-to-Rank – Beyond Search

c© Jannik Strötgen – ATIR-10 11 / 72

Page 12: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LeToR Frameworkquery

documents

user

learningmethod

rankedresults

open issueshow do we model the problem?is it a regression or classification problem?what about our prediction target?

c© Jannik Strötgen – ATIR-10 12 / 72

Page 13: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LeToR Frameworkquery

documents

user

learningmethod

rankedresults

scoring as functiondifferent input signals (features) xiwith weights αi

score(d,q) = f (x1, ..., xm, α1, ..., αm)

whereweights are learnedfeatures derived from d, q, and context

c© Jannik Strötgen – ATIR-10 13 / 72

Page 14: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Outline

1 LeToR Framework

2 Modeling Approaches

3 Gathering User Feedback

4 Evaluating Learning to Rank

5 Learning-to-Rank for Temporal IR

6 Learning-to-Rank – Beyond Search

c© Jannik Strötgen – ATIR-10 14 / 72

Page 15: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Classification – Regression

classification exampledataset of 〈q,d , r〉 triplesr: relevance (binary or multiclass)d: document represented by feature vectortrain ML model to predict class r of a d-q pairdecide relevant if score is above threshold

classification problemsresult in an unordered set of classes

c© Jannik Strötgen – ATIR-10 15 / 72

Page 16: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Classification – Regression

classification problemsresult in an unordered set of classes

regression problemsmap to real values

ordinal regression problemsresult in ordered set of classes

c© Jannik Strötgen – ATIR-10 16 / 72

Page 17: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LeToR Modeling

LeToR can be modeled in three ways:

pointwise: predict goodness of individual documentspairwise: predict users’ relative preference for pairs ofdocumentslistwise: predict goodness of entire query results

each has advantages and disadvantagesfor each concrete approaches exist

in-depth discussion of concrete approaches by Liu 2009

c© Jannik Strötgen – ATIR-10 17 / 72

Page 18: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Pointwise Modeling

( , )query documentyes / no

(−∞,+∞)

x f (x , θ) y

pointwise approachespredict for every document based on its feature vector xdocument goodness y (e.g., label or measure of engagement)training determines the parameter θ based on a loss function(e.g., root-mean-square error)

main disadvantageas input is single document, relative order between documentscannot be naturally considered in the learning process

c© Jannik Strötgen – ATIR-10 18 / 72

Page 19: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Pairwise Modeling

query document 1 document 2( , , ) {-1, +1}

x f (x , θ) y

pairwise approachespredict for every pair of documents based on feature vector xusers’ relative preference regarding the documents(+1 shows preference for document 1; -1 for document 2)training determines the parameter θ based on a loss function(e.g., the number of inverted pairs)

advantage: models relative ordermain disadvantages:

no distinction between excellent–bad and fair–badsensitive to noisy labels (1 wrong label, many mislabeled pairs)

c© Jannik Strötgen – ATIR-10 19 / 72

Page 20: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Listwise Modeling

query doc. 1 . . . doc. k( , , , ) (−∞,+∞)

x f (x , θ) y

listwise approachespredict for ranked list of documents based on feature vector xeffectiveness of ranked list y (e.g., MAP or nDCG)training determines the parameter θ based on a loss function

advantage: positional information visible to loss functiondisadvantage: high training complexity, ...

c© Jannik Strötgen – ATIR-10 20 / 72

Page 21: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Typical Learning-to-Rank Pipeline

learning to rankis typically deployed as a re-ranking step(infeasible to apply it to entire document collection)

query top-K results top-k results user1 2

step 1Determine a top-K result (K ~1,000) using a proven baselineretrieval method (e.g., Okapi BM25 + PageRank)

step 2Re-rank documents from top-K using learning to rankapproach, then return top-k (k ~100) to user

c© Jannik Strötgen – ATIR-10 21 / 72

Page 22: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Outline

1 LeToR Framework

2 Modeling Approaches

3 Gathering User Feedback

4 Evaluating Learning to Rank

5 Learning-to-Rank for Temporal IR

6 Learning-to-Rank – Beyond Search

c© Jannik Strötgen – ATIR-10 22 / 72

Page 23: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Gathering User Feedback

independent of pointwise, pairwise, or listwise modelingsome input from the user is required

to determine prediction target y

two types of user inputexplicit user input (e.g., relevance assessments)implicit user input (e.g., by analyzing their behavior)

c© Jannik Strötgen – ATIR-10 23 / 72

Page 24: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Relevance Assessmentsprocedure

construct a collection of (difficult) queriespool results from different baselinesgather graded relevance assessments from humanassessors

problemshard to represent query workload within 50, 500, 5K queriesdifficult for queries that require personalization or localizationexpensive, time-consuming, and subject to Web dynamics

c© Jannik Strötgen – ATIR-10 24 / 72

Page 25: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Clicks

track user behavior and measure their engagement with resultsclick-through rate of document when shown for querydwell time, i.e., how much time user spent on document

problemsposition bias (consider only first result shown)spurious clicks (consider only clicks with dwell time abovethreshold)feedback loop (add some randomness to results)

reliability of click dataJoachims et al. (2007) and Radlinski & Joachims (2005)

c© Jannik Strötgen – ATIR-10 25 / 72

Page 26: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Skips

user behavior tells us more:skips in addition to clicks as a source of implicit feedback

top 5: d7 d1 d3 d9 d8click

no click

skip previous: d1 > d7 and d9 > d3 (user prefers d1 over d7)skip above: d1 > d7 and d9 > d3, d9 > d7

user study (Joachims et al., 2007): derived relative preferences

are less biased than measures merely based on clicksshow moderate agreement with explicit relevance assessments

c© Jannik Strötgen – ATIR-10 26 / 72

Page 27: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Outline

1 LeToR Framework

2 Modeling Approaches

3 Gathering User Feedback

4 Evaluating Learning to Rank

5 Learning-to-Rank for Temporal IR

6 Learning-to-Rank – Beyond Search

c© Jannik Strötgen – ATIR-10 27 / 72

Page 28: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Learning to Rank – Evaluation

Several benchmark datasets have been released to allow for acomparison of different learning-to-rank methods

LETOR 2.0, 3.0, 4.0 (2007–2009) by Microsoft Research Asiabased on publicly available document collectionscome with precomputed low-level featuresand relevance assessments

Yahoo! Learning to Rank Challenge (2010) by Yahoo! Labscomes with precomputed low-level featuresand relevance assessments

Microsoft Learning to Rank Datasets by Microsoft Research U.S.

comes with precomputed low-level featuresand relevance assessments

c© Jannik Strötgen – ATIR-10 28 / 72

Page 29: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Features

Yahoo! Featuresqueries, ulrs and features descriptions are not given, only the

feature values!

feature engineering critical for any commercial search enginereleasing queries, URLs leads to risk of reverse engineeringreasonable consideration, but prevent IR researchers fromstudying what feature are most effective

LETOR / Microsoft Featureseach query-url pair is represented by a 136-dimensional vector

c© Jannik Strötgen – ATIR-10 29 / 72

Page 30: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 30 / 72

Page 31: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 31 / 72

Page 32: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 32 / 72

Page 33: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 33 / 72

Page 34: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 34 / 72

Page 35: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 35 / 72

Page 36: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 36 / 72

Page 37: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 37 / 72

Page 38: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 38 / 72

Page 39: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

LETOR Features

c© Jannik Strötgen – ATIR-10 39 / 72

Page 40: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Learning to Rank – Starting Point

all detailshttp://research.microsoft.com/en-us/um/beijing/projects/letor/

https://www.microsoft.com/en-us/research/project/mslr/

datasetsdataset descriptionspartitioned in subsets (for cross-validation)evaluation scripts, significance test scriptsfeature list

everything required to get started is available

c© Jannik Strötgen – ATIR-10 40 / 72

Page 41: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Outline

1 LeToR Framework

2 Modeling Approaches

3 Gathering User Feedback

4 Evaluating Learning to Rank

5 Learning-to-Rank for Temporal IR

6 Learning-to-Rank – Beyond Search

c© Jannik Strötgen – ATIR-10 41 / 72

Page 42: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Learning-to-Rank for Temporal IR

Kanhabua & Nørvåg (2012)learning-to-rank approach for time-sensitive queries

standard temporal IR approachesmixture model linearly combining textual similarity and temporalsimilarityprobabilistic model generating a query from the textual andtemporal part of document independent

learning-to-rank approachtwo classes of features: entity features and time featuresboth derived from annotations (NER, temporal tagging)

c© Jannik Strötgen – ATIR-10 42 / 72

Page 43: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Learning-to-Rank for Temporal IR

document modela document collection over timedocument is composed of a bag of words and time

– publication date– temporal expressions mentioned in document

annotated document composed of– set of named entities– set of temporal expressions– set of annotated sentences

temporal query modelq = {qtext ,qtime}qtime might be explicit or implicit

c© Jannik Strötgen – ATIR-10 43 / 72

Page 44: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Learning-to-Rank for Temporal IR

learning-to-ranka wide range of temporal featuresa wide range of entity featuresmodels trained using labeled query/document pairsdocuments ranked according to weighted sum of its featurescores

experimentsshow improvement of baselines and other time-aware models(many queries also contained entities, news corpus)

c© Jannik Strötgen – ATIR-10 44 / 72

Page 45: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Outline

1 LeToR Framework

2 Modeling Approaches

3 Gathering User Feedback

4 Evaluating Learning to Rank

5 Learning-to-Rank for Temporal IR

6 Learning-to-Rank – Beyond Search

c© Jannik Strötgen – ATIR-10 45 / 72

Page 46: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Learning-to-Rank – Beyond Search

learning to rank is applicable beyond web search

Example: matching in eharmony.com

Slides by Vaclav Petricek:http://www.slideshare.net/VaclavPetricek/data-science-of-love

basic idea:standard approach is search-based, filter out non-matcheseharmony approach is learning to rank: suggest potentialmatches

c© Jannik Strötgen – ATIR-10 46 / 72

Page 47: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Matching in eHarmony.com

starting point in the 1990sdistinguish marriages that work well and those that don’t

step 1: compatibility matchingbased on 150 questions: personality, values, attitudes, beliefsimportant attributes for the long termpredict marital satisfaction

even if people are compatible,they might not be interested to talk to each other

c© Jannik Strötgen – ATIR-10 47 / 72

Page 48: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Matching in eHarmony.com

step 2: affinity matchingbased on other features: distance, height difference, zoom levelof photopredict probability of message exchange

howeverwho should be introduced to whom and when?

match distributionbased on graph optimization problem (constrained max flow)

c© Jannik Strötgen – ATIR-10 48 / 72

Page 49: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 49 / 72

Page 50: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

blue: happy marriagesred: distressed marriages

is that person arguing withanything you say?

relation betweenobstreperousness andmarriage happiness

c© Jannik Strötgen – ATIR-10 50 / 72

Page 51: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 51 / 72

Page 52: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Matching in eHarmony.com

even if people are compatible,they might not be interested to talk to each other

c© Jannik Strötgen – ATIR-10 52 / 72

Page 53: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 53 / 72

Page 54: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 54 / 72

Page 55: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 55 / 72

Page 56: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 56 / 72

Page 57: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

self-reportedattractiveness

people who report sameattractiveness match better

c© Jannik Strötgen – ATIR-10 57 / 72

Page 58: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 58 / 72

Page 59: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

zoom size mattersonly face: doesn’t tell muchno face: someone is hiding

ratio: face / pic size

c© Jannik Strötgen – ATIR-10 59 / 72

Page 60: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 60 / 72

Page 61: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 61 / 72

Page 62: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Matching in eHarmony.com

howeverwho should be introduced to whom and when?

c© Jannik Strötgen – ATIR-10 62 / 72

Page 63: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 63 / 72

Page 64: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 64 / 72

Page 65: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

although many arecompatible

not all should be suggested

c© Jannik Strötgen – ATIR-10 65 / 72

Page 66: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 66 / 72

Page 67: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 67 / 72

Page 68: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

c© Jannik Strötgen – ATIR-10 68 / 72

Page 69: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

optimization problemgoal: maximize two way

communication(highest chance that both are

interested)

c© Jannik Strötgen – ATIR-10 69 / 72

Page 70: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Summary

Learning to Rankprovides systematic ways to combine features

modelingpointwise: predict goodness of individual documentpairwise: predict relative preference for document pairslistwise: predict effectiveness of ranked list of documents

explicit and implicit user inputsinclude relevance assessments, clicks, and skips

Thank you for your attention!

c© Jannik Strötgen – ATIR-10 70 / 72

Page 71: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

References

Fuhr (1989): Optimum Polynomial Retrieval Functions basedon the the Probability Ranking Principle, ACM TOIS 7(3).Liu (2009): Learning to Rank for Information Retrieval,Foundations and Trends in Information Retrieval 3(3):225â331.Joachims et al. (2007): Evaluating the Accuracy of ImplicitFeedback from Clicks and Query Reformulations in WebSearch, ACM TOIS 25(2).Radlinski & Joachims (2005): Query Chains: Learning to Rankfrom Implicit Feedback, KDD.Kanhabua & Nørvåg (2012): Learning to Rank Search Resultsfor Time-Sensitive Queries, CIKM.

c© Jannik Strötgen – ATIR-10 71 / 72

Page 72: Learning to Rankresources.mpi-inf.mpg.de/departments/d5/teaching/... · Clicks track user behavior and measure their engagement with results click-through rate of document when shown

LeToR Framework Modeling User Feedback Evaluation Time Beyond Search

Thankssome slides / examples are taken from / similar to those of:

Klaus Berberich, Saarland University, previous ATIR lecture

Manning, Raghavan, Schütze: Introduction to Information Retrieval(including slides to the book)

and the eharmony.com slides by Vaclav Petricek:http://www.slideshare.net/VaclavPetricek/data-science-of-love

c© Jannik Strötgen – ATIR-10 72 / 72