SlideShare a Scribd company logo
Portfolio Theory of Information
Retrieval
Jun Wang and Jianhan Zhu
jun.wang@cs.ucl.ac.uk
Department of Computer Science
University College London, UK
Portfolio Theory of Information Retrieval – p. 1/22
Outline
Research Problem
An Analogy: Stock Selection In Financial Markets
Portfolio Theory of Information Retrieval
Evaluations on ad hoc text retrieval
Conclusions
Portfolio Theory of Information Retrieval – p. 2/22
Two-stage Process in IR
Stage 1 calculates the relevance between the given
user information need (query) and each of the
documents
Producing a “best guess” at the relevance, e.g. the
BM25 and the language modelling approaches
Stage 2 presents (normally rank) relevant documents
The probability ranking principle (PRP) states that
“the system should rank documents in order of
decreasing probability of relevance” [Cooper(1971)]
Under certain assumptions, the overall effectiveness,
e.g., expected Precision, is maximized
[Robertson(1977)]
Portfolio Theory of Information Retrieval – p. 3/22
Ranking Under Uncertainty
the PRP ignores [Gordon and Lenk(1991)]:
there is uncertainty when we calculate relevance
scores, e.g., due to limited sample size, and
relevance scores of documents are correlated
0 0.2 0.4 0.6 0.8 1
0
1
2
3
4
5
6
Document A
Document B
Uncertainty of the relevance scores Correlations of relevance scores
Portfolio Theory of Information Retrieval – p. 4/22
An Example
Suppose we have query apple, and two classes of users
U1: Apple_Computers and U2: Apple_Fruit; U1 has
twice as many members as U2
An IR system retrieved three documents d1, d2 and d3;
their probabilities of relevance are as follows:
UserClass d1: Apple_Comp. d2: Apple_Comp. d3: Apple_Fruit
Apple Computers 1 1 0
Apple Fruit 0 0 1
p(r) 2/3 2/3 1/3
The PRP is not good
({d1Apple_Comp., d2Apple_Comp., d3Apple_Fruit}) as
user group U2 (Apple Fruit) has to reject two documents
before reaching the one it wants [Robertson(1977)]
Portfolio Theory of Information Retrieval – p. 5/22
A Little History of Probabilistic Ranking
1960s [Maron and Kuhns(1960)] mentioned the
two-stage process implicitly
1970s [Cooper(1971)] examined the PRP explicitly
well discussed in [Robertson(1977)] and Stirling’s
thesis [Stirling(1977)]
1991 [Gordon and Lenk(1991)] studied its limitations
1998 [Carbonell and Goldstein(1998)] proposed
diversity-based reranking (MMR)
2006 The “less is more” model
[Chen and Karger(2006)]
maximize the probability of finding a relevant
documents in top-n ranked list, where a < n
Portfolio Theory of Information Retrieval – p. 6/22
Our View of the Ranking Problem (1)
We argue that ranking under uncertainty is not just
about picking individual relevant documents, but about
choosing the right combination of relevant
document - the Portfolio Effect
There is a similar scenario in financial markets:
Two observations:
The future returns of stocks cannot be estimated
with absolute certainty
The future returns are correlated
Portfolio Theory of Information Retrieval – p. 7/22
Our View of the Ranking Problem (2)
The analogy:
According to the PRP, one might first rank stocks and
then choose the top-n most “profitable” stocks
Such a principle that essentially maximizes the
expected future return was, however, rejected by
Markowitz in Modern Portfolio Theory [Markowitz(1952)]
Portfolio Theory of Information Retrieval – p. 8/22
Our View of the Ranking Problems (3)
Markowitz’ approach is based on the analysis of the
expected return (mean) of a portfolio and its variance
(or standard deviation) of return. The latter serves as a
measure of risk
4 6 8 10 12 14
0
0.5
1
1.5
2
2.5
3
Standard Deviation
Return
Efficient Frontier
Google
Coca− Cola
0 20 40 60 80 100 120
0
0.2
0.4
0.6
0.8
1
α (Risk Preference)
PortfolioPercentage
Google
Coca− Cola
Efficient Frontier Percentage in the Portfolio
Portfolio Theory of Information Retrieval – p. 9/22
Portfolio Theory in IR (1)
Objective: find an optimal ranked list (consisting of n
documents from rank 1 to n) that has the maximum
effectiveness in response to the given information need
Define effectiveness: consider the weighted average of
the relevance scores in the ranked list:
Rn ≡
n
i=1
wiri
where Rn denotes the overall relevance of a ranked list.
Variable wi, where n
i=1 wi = 1, differentiates the
importance of rank positions. ri is the relevance score
of a document in the list, where i = {1, ..., n}, for each of
the rank positions
Portfolio Theory of Information Retrieval – p. 10/22
Portfolio Theory in IR (2)
Weight wi is similar to the discount factors that have
been applied to IR evaluation in order to penalize
late-retrieved relevant documents
[Järvelin and Kekäläinen(2002)]
It can be easily shown that when w1 > w2... > wn, the
maximum value of Rn gives the ranking order
r1 > r2... > rn
This follows immediately that maximizing R – by which
the document with highest relevance score is retrieved
first, the document with next highest is retrieved
second, etc. – is equivalent to the PRP
Portfolio Theory of Information Retrieval – p. 11/22
Portfolio Theory in IR (3)
During retrieval, the overall relevance Rn cannot be
calculated with certainty
Quantify a ranked list based on its expectation (mean
E[Rn] ) and its variance (V ar(Rn)):
E[Rn] =
n
i=1
wiE[ri]
V ar(Rn) =
n
i=1
n
j=1
wiwjci,j
where ci,j is the (co)variance of the relevance scores
between the two documents at position i and j. E[ri] is
the expected relevance score, determined by a point
estimate from the specific retrieval model
Portfolio Theory of Information Retrieval – p. 12/22
Portfolio Theory in IR (4)
What to be optimized?
1. Maximize the mean E[Rn] regardless of its variance
2. Minimize the variance V ar(Rn) regardless of its mean
3. Minimize the variance for a specified mean t
(parameter): min V ar(Rn), subject to E[Rn] = t
4. Maximize the mean for a specified variance h
(parameter): max E[Rn], subject to V ar(Rn) = h
5. Maximize the mean and minimize the variance by using
a specified risk preference parameter b:
max On = E[Rn] − bV ar(Rn)
Portfolio Theory of Information Retrieval – p. 13/22
Portfolio Theory in IR (5)
The Efficient Frontier:
0 0.05 0.1 0.15 0.2 0.25
2
2.5
3
3.5
4
4.5
5
5.5
6
Variance
ExpectedRelevance
Objective function: On = E[Rn] − bV ar(Rn) where b is a
parameter adjusting the risk level
Portfolio Theory of Information Retrieval – p. 14/22
Portfolio Theory in IR (6)
Our solution provides a mathematical model of rank
diversification
Suppose we have two documents. Their relevance
scores are 0.5 and 0.9 respectively
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5
0.4
0.5
0.6
0.7
0.8
0.9
1
Relevance
Standard Deviation
ρ = −1
ρ = −0.5
ρ = 0
ρ = 1
Portfolio Theory of Information Retrieval – p. 15/22
Evaluations on Ad hoc Text Retrieval (1)
We experimented on Ad hoc and sub topic retrieval
Calculation of the Mean and Variance:
Mean: posterior mean of the chosen text retrieval
model
Covariance matrix:
- largely missing in IR modelling
- formally, should be determined by the second
moment of the relevance scores (model
parameters), e.g., applying the Bayesian paradigm
- can be approximated by the covariance with
respect to their term occurrences
Portfolio Theory of Information Retrieval – p. 16/22
Evaluations on Ad hoc Text Retrieval (2)
Impact of parameter b on different evaluation metrics
(a) Mean Reciprocal Rank (MRR) (b) Mean Average Precision (MAP)
(a) positive b: “invest” into different docs. increases the
chance of early returning the first rel. docs
(b) negative b: “invest” in “similar” docs (big variance)
might hurt the MRR but on average increases the
performance of the entire ranked list
Portfolio Theory of Information Retrieval – p. 17/22
Evaluations on Ad hoc Text Retrieval (3)
The impact of the parameter b on a risk-sensitive
metric, k-call [Chen and Karger(2006)]
10-call: ambitious, return 10 rel. docs.
1-call: conservative, return at least one rel. doc.
Positive b when k is small (1 and 2). diversifying
reduces the risk of not returning any rel docs.
Negative b as k increases. Taking risk increases the
chance of finding more rel. docs.
Portfolio Theory of Information Retrieval – p. 18/22
Evaluations on Ad hoc Text Retrieval (4)
Comparison with the PRP (Linear smoothing LM)
Measures CSIRO WT10g Robust Robust hard TREC8
MRR 0.869 0.558 0.592 0.393 0.589
0.843 0.492 0.549 0.352 0.472
+3.08% +13.41%* +7.83%* +11.65%* +24.79%*
MAP 0.41 0.182 0.204 0.084 0.212
0.347 0.157 0.185 0.078 0.198
+18.16%*+15.92%*+10.27%* +7.69%* +7.07%*
NDCG 0.633 0.433 0.421 0.271 0.452
0.587 0.398 0.396 0.252 0.422
+7.88%* +8.82%* +6.25%* +7.55%* +7.05%*
NDCG@10 0.185 0.157 0.175 0.081 0.149
0.170 0.141 0.169 0.078 0.140
+8.96%* +11.23%* +3.80% +3.90% +6.36%*
NDCG@100 0.377 0.286 0.314 0.169 0.305
0.355 0.262 0.292 0.159 0.287
+6.25%* +9.27%* +7.55%* +6.58%* +6.34%*
Portfolio Theory of Information Retrieval – p. 19/22
Evaluations on Ad hoc Text Retrieval (5)
Comparison with diversity-based reranking, the MMR
[Carbonell and Goldstein(1998)]
In each cell, the first line shows the performance of our approach, and the second line shows
the performance of the MMR method and gain of our method over the MMR method.
Models Dirichlet Jelinek-Mercer BM25
sub-MRR 0.014 0.011 0.009
0.012 (+16.67%*)0.009 (+22.22%*)0.007 (+28.57%*)
sub-Recall@5 0.324 0.255 0.275
0.304 (+6.58%*) 0.234 (+8.97%*) 0.27 (+1.85%)
sub-Recall@10 0.381 0.366 0.352
0.362 (+5.25%) 0.351 (+4.27%) 0.344 (+2.33%)
sub-Recall@20 0.472 0.458 0.464
0.455 (+3.74%) 0.41 (+11.71%*) 0.446 (+4.04%)
sub-Recall@100 0.563 0.582 0.577
0.558 (+0.90%) 0.55 (+5.82%*) 0.558 (+3.41%)
Portfolio Theory of Information Retrieval – p. 20/22
Conclusions
We have presented a new ranking theory for IR
The benefit of diversification is well quantified
Is it a unified theory to explain risk and reward in IR?
Portfolio Theory of Information Retrieval – p. 21/22
References
[Cooper(1971)] William S. Cooper. The inadequacy of probability of usefulness as a ranking
criterion for retrieval system output. University of California, Berkeley, 1971.
[Robertson(1977)] S. E. Robertson. The probability ranking principle in IR. Journal of
Documentation, pages 294–304, 1977.
[Gordon and Lenk(1991)] Michael D. Gordon and Peter Lenk. A utility theoretic examination
of the probability ranking principle in information retrieval. JASIS, 42(10):703–714, 1991.
[Maron and Kuhns(1960)] M. E. Maron and J. L. Kuhns. On relevance, probabilistic indexing
and information retrieval. J. ACM, 7(3), 1960.
[Stirling(1977)] Keith H. Stirling. The Effect of Document Ranking on Retrieval System
Performance: A Search for an Optimal Ranking Rule. PhD thesis, UC, Berkeley, 1977.
[Carbonell and Goldstein(1998)] Jaime Carbonell and Jade Goldstein. The use of MMR,
diversity-based reranking for reordering documents and producing summaries. In SIGIR,
1998.
[Chen and Karger(2006)] Harr Chen and David R. Karger. Less is more: probabilistic
models for retrieving fewer relevant documents. In SIGIR, 2006.
[Markowitz(1952)] H Markowitz. Portfolio selection. Journal of Finance, 1952.
[Järvelin and Kekäläinen(2002)] Kalervo Järvelin and Jaana Kekäläinen. Cumulated
gain-based evaluation of IR techniques. ACM Trans. Inf. Syst., 2002.
Portfolio Theory of Information Retrieval – p. 22/22

More Related Content

What's hot (20)

DOCX
Statistics for management
Vinay Aradhya
 
PPTX
Lu2 introduction to statistics
LamineKaba6
 
PDF
Statistics assignment help
Cheap Assignment Help UK
 
DOC
Statistics Assignments 090427
amykua
 
PPTX
Data What Type Of Data Do You Have V2.1
TimKasse
 
PDF
The pertinent single-attribute-based classifier for small datasets classific...
IJECEIAES
 
PDF
Dmml report final
sarthakkhare3
 
PDF
Statistics And Probability Tutorial | Statistics And Probability for Data Sci...
Edureka!
 
PDF
Optimizing transformation for linearity between online
Alexander Decker
 
PDF
Exploratory data analysis data visualization
Dr. Hamdan Al-Sabri
 
PDF
Machine learning meetup
QuantUniversity
 
PDF
5. Llinking employers and employees responses
BEYOND4.0
 
PDF
Mathematical Econometrics
jonren
 
PDF
2. Joint analysis - TNO
BEYOND4.0
 
PPTX
AI Algorithms
Dr. C.V. Suresh Babu
 
PDF
A second order confirmatory factor analysis of composite
Alexander Decker
 
DOCX
Statistics for managers
sonia gupta
 
DOC
Business Development Analysis
Manpreet Chandhok
 
DOCX
SMU DRIVE SPRING 2017 MBA 103- Statistics for Management solved free assignment
rahul kumar verma
 
PDF
IRJET- Online Course Recommendation System
IRJET Journal
 
Statistics for management
Vinay Aradhya
 
Lu2 introduction to statistics
LamineKaba6
 
Statistics assignment help
Cheap Assignment Help UK
 
Statistics Assignments 090427
amykua
 
Data What Type Of Data Do You Have V2.1
TimKasse
 
The pertinent single-attribute-based classifier for small datasets classific...
IJECEIAES
 
Dmml report final
sarthakkhare3
 
Statistics And Probability Tutorial | Statistics And Probability for Data Sci...
Edureka!
 
Optimizing transformation for linearity between online
Alexander Decker
 
Exploratory data analysis data visualization
Dr. Hamdan Al-Sabri
 
Machine learning meetup
QuantUniversity
 
5. Llinking employers and employees responses
BEYOND4.0
 
Mathematical Econometrics
jonren
 
2. Joint analysis - TNO
BEYOND4.0
 
AI Algorithms
Dr. C.V. Suresh Babu
 
A second order confirmatory factor analysis of composite
Alexander Decker
 
Statistics for managers
sonia gupta
 
Business Development Analysis
Manpreet Chandhok
 
SMU DRIVE SPRING 2017 MBA 103- Statistics for Management solved free assignment
rahul kumar verma
 
IRJET- Online Course Recommendation System
IRJET Journal
 

Viewers also liked (7)

PDF
Weinan Zhang's KDD15 Talk: Statistical Arbitrage Mining for Display Advertising
Jun Wang
 
PDF
On Search, Personalisation and Real-time Advertising
Jun Wang
 
PDF
A Brief Introduction of Real-time Bidding Display Advertising and Evaluation ...
Jun Wang
 
PDF
Wsdm17 value-at-risk-bidding
Jun Wang
 
PDF
Wsdm2015
Jun Wang
 
PPTX
Statistical Information Retrieval Modelling: from the Probability Ranking Pr...
Jun Wang
 
PDF
Deep Learning
Jun Wang
 
Weinan Zhang's KDD15 Talk: Statistical Arbitrage Mining for Display Advertising
Jun Wang
 
On Search, Personalisation and Real-time Advertising
Jun Wang
 
A Brief Introduction of Real-time Bidding Display Advertising and Evaluation ...
Jun Wang
 
Wsdm17 value-at-risk-bidding
Jun Wang
 
Wsdm2015
Jun Wang
 
Statistical Information Retrieval Modelling: from the Probability Ranking Pr...
Jun Wang
 
Deep Learning
Jun Wang
 
Ad

Similar to Portfolio Theory of Information Retrieval (20)

PDF
fma.ny.presentation
Artur Aiguzhinov
 
PDF
Diversity
Telnet
 
PPT
lectueereerrrrrrtttttrre11-probir(1).ppt
joyaluca2
 
PPT
lecture11-prohdhhdhdhdhdhdhdhdbir(2).ppt
joyaluca2
 
PDF
IR-lec17-probabilistic-ir.pdf
himarusti
 
PPTX
Information Retrieval Evaluation
José Ramón Ríos Viqueira
 
PPT
lecture14-learning-ranking.ppt
VishalKumar725248
 
PPT
lecture14-learning-ranking.ppt
VishalKumar725248
 
PPT
Part 1
butest
 
PDF
The legacy of modern portfolio theory academic essay assignment - www.topgr...
Top Grade Papers
 
PDF
cikm2016_mean_variance_evaluation
João Palotti
 
PPT
probabilistic ranking
FELIX75
 
PDF
An Introduction to Information Retrieval.pdf
Tiffany Daniels
 
PDF
Dynamic Information Retrieval Tutorial - SIGIR 2015
Marc Sloan
 
PPT
lecture10-efficient-scoring.ppmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmt
RAtna29
 
PDF
Probabilistic Information Retrieval
Harsh Thakkar
 
PPTX
Algorithm evaluation using Item Response Theory
CSIRO
 
PPTX
Learn to Rank search results
Ganesh Venkataraman
 
PPTX
lecture13-DTrees-textcat.pptxnnnnnnnnnnnnnnnnnnnnnn
RAtna29
 
PDF
Search: Probabilistic Information Retrieval
Vipul Munot
 
fma.ny.presentation
Artur Aiguzhinov
 
Diversity
Telnet
 
lectueereerrrrrrtttttrre11-probir(1).ppt
joyaluca2
 
lecture11-prohdhhdhdhdhdhdhdhdbir(2).ppt
joyaluca2
 
IR-lec17-probabilistic-ir.pdf
himarusti
 
Information Retrieval Evaluation
José Ramón Ríos Viqueira
 
lecture14-learning-ranking.ppt
VishalKumar725248
 
lecture14-learning-ranking.ppt
VishalKumar725248
 
Part 1
butest
 
The legacy of modern portfolio theory academic essay assignment - www.topgr...
Top Grade Papers
 
cikm2016_mean_variance_evaluation
João Palotti
 
probabilistic ranking
FELIX75
 
An Introduction to Information Retrieval.pdf
Tiffany Daniels
 
Dynamic Information Retrieval Tutorial - SIGIR 2015
Marc Sloan
 
lecture10-efficient-scoring.ppmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmt
RAtna29
 
Probabilistic Information Retrieval
Harsh Thakkar
 
Algorithm evaluation using Item Response Theory
CSIRO
 
Learn to Rank search results
Ganesh Venkataraman
 
lecture13-DTrees-textcat.pptxnnnnnnnnnnnnnnnnnnnnnn
RAtna29
 
Search: Probabilistic Information Retrieval
Vipul Munot
 
Ad

Recently uploaded (20)

PDF
How a Code Plagiarism Checker Protects Originality in Programming
Code Quiry
 
PDF
Julia Furst Morgado The Lazy Guide to Kubernetes with EKS Auto Mode + Karpenter
AWS Chicago
 
PPTX
Lecture 5 - Agentic AI and model context protocol.pptx
Dr. LAM Yat-fai (林日辉)
 
PDF
Productivity Management Software | Workstatus
Lovely Baghel
 
PPTX
python advanced data structure dictionary with examples python advanced data ...
sprasanna11
 
PDF
2025-07-15 EMEA Volledig Inzicht Dutch Webinar
ThousandEyes
 
PDF
CloudStack GPU Integration - Rohit Yadav
ShapeBlue
 
PDF
Bitcoin+ Escalando sin concesiones - Parte 1
Fernando Paredes García
 
PDF
Shuen Mei Parth Sharma Boost Productivity, Innovation and Efficiency wit...
AWS Chicago
 
PDF
CIFDAQ'S Token Spotlight for 16th July 2025 - ALGORAND
CIFDAQ
 
PPTX
Earn Agentblazer Status with Slack Community Patna.pptx
SanjeetMishra29
 
PDF
CIFDAQ Market Insight for 14th July 2025
CIFDAQ
 
PPTX
Building a Production-Ready Barts Health Secure Data Environment Tooling, Acc...
Barts Health
 
PDF
GITLAB-CICD_For_Professionals_KodeKloud.pdf
deepaktyagi0048
 
PDF
Women in Automation Presents: Reinventing Yourself — Bold Career Pivots That ...
DianaGray10
 
PDF
Human-centred design in online workplace learning and relationship to engagem...
Tracy Tang
 
PPTX
Simplifying End-to-End Apache CloudStack Deployment with a Web-Based Automati...
ShapeBlue
 
PPTX
Building and Operating a Private Cloud with CloudStack and LINBIT CloudStack ...
ShapeBlue
 
PPTX
Machine Learning Benefits Across Industries
SynapseIndia
 
PDF
Trading Volume Explained by CIFDAQ- Secret Of Market Trends
CIFDAQ
 
How a Code Plagiarism Checker Protects Originality in Programming
Code Quiry
 
Julia Furst Morgado The Lazy Guide to Kubernetes with EKS Auto Mode + Karpenter
AWS Chicago
 
Lecture 5 - Agentic AI and model context protocol.pptx
Dr. LAM Yat-fai (林日辉)
 
Productivity Management Software | Workstatus
Lovely Baghel
 
python advanced data structure dictionary with examples python advanced data ...
sprasanna11
 
2025-07-15 EMEA Volledig Inzicht Dutch Webinar
ThousandEyes
 
CloudStack GPU Integration - Rohit Yadav
ShapeBlue
 
Bitcoin+ Escalando sin concesiones - Parte 1
Fernando Paredes García
 
Shuen Mei Parth Sharma Boost Productivity, Innovation and Efficiency wit...
AWS Chicago
 
CIFDAQ'S Token Spotlight for 16th July 2025 - ALGORAND
CIFDAQ
 
Earn Agentblazer Status with Slack Community Patna.pptx
SanjeetMishra29
 
CIFDAQ Market Insight for 14th July 2025
CIFDAQ
 
Building a Production-Ready Barts Health Secure Data Environment Tooling, Acc...
Barts Health
 
GITLAB-CICD_For_Professionals_KodeKloud.pdf
deepaktyagi0048
 
Women in Automation Presents: Reinventing Yourself — Bold Career Pivots That ...
DianaGray10
 
Human-centred design in online workplace learning and relationship to engagem...
Tracy Tang
 
Simplifying End-to-End Apache CloudStack Deployment with a Web-Based Automati...
ShapeBlue
 
Building and Operating a Private Cloud with CloudStack and LINBIT CloudStack ...
ShapeBlue
 
Machine Learning Benefits Across Industries
SynapseIndia
 
Trading Volume Explained by CIFDAQ- Secret Of Market Trends
CIFDAQ
 

Portfolio Theory of Information Retrieval

  • 1. Portfolio Theory of Information Retrieval Jun Wang and Jianhan Zhu [email protected] Department of Computer Science University College London, UK Portfolio Theory of Information Retrieval – p. 1/22
  • 2. Outline Research Problem An Analogy: Stock Selection In Financial Markets Portfolio Theory of Information Retrieval Evaluations on ad hoc text retrieval Conclusions Portfolio Theory of Information Retrieval – p. 2/22
  • 3. Two-stage Process in IR Stage 1 calculates the relevance between the given user information need (query) and each of the documents Producing a “best guess” at the relevance, e.g. the BM25 and the language modelling approaches Stage 2 presents (normally rank) relevant documents The probability ranking principle (PRP) states that “the system should rank documents in order of decreasing probability of relevance” [Cooper(1971)] Under certain assumptions, the overall effectiveness, e.g., expected Precision, is maximized [Robertson(1977)] Portfolio Theory of Information Retrieval – p. 3/22
  • 4. Ranking Under Uncertainty the PRP ignores [Gordon and Lenk(1991)]: there is uncertainty when we calculate relevance scores, e.g., due to limited sample size, and relevance scores of documents are correlated 0 0.2 0.4 0.6 0.8 1 0 1 2 3 4 5 6 Document A Document B Uncertainty of the relevance scores Correlations of relevance scores Portfolio Theory of Information Retrieval – p. 4/22
  • 5. An Example Suppose we have query apple, and two classes of users U1: Apple_Computers and U2: Apple_Fruit; U1 has twice as many members as U2 An IR system retrieved three documents d1, d2 and d3; their probabilities of relevance are as follows: UserClass d1: Apple_Comp. d2: Apple_Comp. d3: Apple_Fruit Apple Computers 1 1 0 Apple Fruit 0 0 1 p(r) 2/3 2/3 1/3 The PRP is not good ({d1Apple_Comp., d2Apple_Comp., d3Apple_Fruit}) as user group U2 (Apple Fruit) has to reject two documents before reaching the one it wants [Robertson(1977)] Portfolio Theory of Information Retrieval – p. 5/22
  • 6. A Little History of Probabilistic Ranking 1960s [Maron and Kuhns(1960)] mentioned the two-stage process implicitly 1970s [Cooper(1971)] examined the PRP explicitly well discussed in [Robertson(1977)] and Stirling’s thesis [Stirling(1977)] 1991 [Gordon and Lenk(1991)] studied its limitations 1998 [Carbonell and Goldstein(1998)] proposed diversity-based reranking (MMR) 2006 The “less is more” model [Chen and Karger(2006)] maximize the probability of finding a relevant documents in top-n ranked list, where a < n Portfolio Theory of Information Retrieval – p. 6/22
  • 7. Our View of the Ranking Problem (1) We argue that ranking under uncertainty is not just about picking individual relevant documents, but about choosing the right combination of relevant document - the Portfolio Effect There is a similar scenario in financial markets: Two observations: The future returns of stocks cannot be estimated with absolute certainty The future returns are correlated Portfolio Theory of Information Retrieval – p. 7/22
  • 8. Our View of the Ranking Problem (2) The analogy: According to the PRP, one might first rank stocks and then choose the top-n most “profitable” stocks Such a principle that essentially maximizes the expected future return was, however, rejected by Markowitz in Modern Portfolio Theory [Markowitz(1952)] Portfolio Theory of Information Retrieval – p. 8/22
  • 9. Our View of the Ranking Problems (3) Markowitz’ approach is based on the analysis of the expected return (mean) of a portfolio and its variance (or standard deviation) of return. The latter serves as a measure of risk 4 6 8 10 12 14 0 0.5 1 1.5 2 2.5 3 Standard Deviation Return Efficient Frontier Google Coca− Cola 0 20 40 60 80 100 120 0 0.2 0.4 0.6 0.8 1 α (Risk Preference) PortfolioPercentage Google Coca− Cola Efficient Frontier Percentage in the Portfolio Portfolio Theory of Information Retrieval – p. 9/22
  • 10. Portfolio Theory in IR (1) Objective: find an optimal ranked list (consisting of n documents from rank 1 to n) that has the maximum effectiveness in response to the given information need Define effectiveness: consider the weighted average of the relevance scores in the ranked list: Rn ≡ n i=1 wiri where Rn denotes the overall relevance of a ranked list. Variable wi, where n i=1 wi = 1, differentiates the importance of rank positions. ri is the relevance score of a document in the list, where i = {1, ..., n}, for each of the rank positions Portfolio Theory of Information Retrieval – p. 10/22
  • 11. Portfolio Theory in IR (2) Weight wi is similar to the discount factors that have been applied to IR evaluation in order to penalize late-retrieved relevant documents [Järvelin and Kekäläinen(2002)] It can be easily shown that when w1 > w2... > wn, the maximum value of Rn gives the ranking order r1 > r2... > rn This follows immediately that maximizing R – by which the document with highest relevance score is retrieved first, the document with next highest is retrieved second, etc. – is equivalent to the PRP Portfolio Theory of Information Retrieval – p. 11/22
  • 12. Portfolio Theory in IR (3) During retrieval, the overall relevance Rn cannot be calculated with certainty Quantify a ranked list based on its expectation (mean E[Rn] ) and its variance (V ar(Rn)): E[Rn] = n i=1 wiE[ri] V ar(Rn) = n i=1 n j=1 wiwjci,j where ci,j is the (co)variance of the relevance scores between the two documents at position i and j. E[ri] is the expected relevance score, determined by a point estimate from the specific retrieval model Portfolio Theory of Information Retrieval – p. 12/22
  • 13. Portfolio Theory in IR (4) What to be optimized? 1. Maximize the mean E[Rn] regardless of its variance 2. Minimize the variance V ar(Rn) regardless of its mean 3. Minimize the variance for a specified mean t (parameter): min V ar(Rn), subject to E[Rn] = t 4. Maximize the mean for a specified variance h (parameter): max E[Rn], subject to V ar(Rn) = h 5. Maximize the mean and minimize the variance by using a specified risk preference parameter b: max On = E[Rn] − bV ar(Rn) Portfolio Theory of Information Retrieval – p. 13/22
  • 14. Portfolio Theory in IR (5) The Efficient Frontier: 0 0.05 0.1 0.15 0.2 0.25 2 2.5 3 3.5 4 4.5 5 5.5 6 Variance ExpectedRelevance Objective function: On = E[Rn] − bV ar(Rn) where b is a parameter adjusting the risk level Portfolio Theory of Information Retrieval – p. 14/22
  • 15. Portfolio Theory in IR (6) Our solution provides a mathematical model of rank diversification Suppose we have two documents. Their relevance scores are 0.5 and 0.9 respectively 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 0.4 0.5 0.6 0.7 0.8 0.9 1 Relevance Standard Deviation ρ = −1 ρ = −0.5 ρ = 0 ρ = 1 Portfolio Theory of Information Retrieval – p. 15/22
  • 16. Evaluations on Ad hoc Text Retrieval (1) We experimented on Ad hoc and sub topic retrieval Calculation of the Mean and Variance: Mean: posterior mean of the chosen text retrieval model Covariance matrix: - largely missing in IR modelling - formally, should be determined by the second moment of the relevance scores (model parameters), e.g., applying the Bayesian paradigm - can be approximated by the covariance with respect to their term occurrences Portfolio Theory of Information Retrieval – p. 16/22
  • 17. Evaluations on Ad hoc Text Retrieval (2) Impact of parameter b on different evaluation metrics (a) Mean Reciprocal Rank (MRR) (b) Mean Average Precision (MAP) (a) positive b: “invest” into different docs. increases the chance of early returning the first rel. docs (b) negative b: “invest” in “similar” docs (big variance) might hurt the MRR but on average increases the performance of the entire ranked list Portfolio Theory of Information Retrieval – p. 17/22
  • 18. Evaluations on Ad hoc Text Retrieval (3) The impact of the parameter b on a risk-sensitive metric, k-call [Chen and Karger(2006)] 10-call: ambitious, return 10 rel. docs. 1-call: conservative, return at least one rel. doc. Positive b when k is small (1 and 2). diversifying reduces the risk of not returning any rel docs. Negative b as k increases. Taking risk increases the chance of finding more rel. docs. Portfolio Theory of Information Retrieval – p. 18/22
  • 19. Evaluations on Ad hoc Text Retrieval (4) Comparison with the PRP (Linear smoothing LM) Measures CSIRO WT10g Robust Robust hard TREC8 MRR 0.869 0.558 0.592 0.393 0.589 0.843 0.492 0.549 0.352 0.472 +3.08% +13.41%* +7.83%* +11.65%* +24.79%* MAP 0.41 0.182 0.204 0.084 0.212 0.347 0.157 0.185 0.078 0.198 +18.16%*+15.92%*+10.27%* +7.69%* +7.07%* NDCG 0.633 0.433 0.421 0.271 0.452 0.587 0.398 0.396 0.252 0.422 +7.88%* +8.82%* +6.25%* +7.55%* +7.05%* NDCG@10 0.185 0.157 0.175 0.081 0.149 0.170 0.141 0.169 0.078 0.140 +8.96%* +11.23%* +3.80% +3.90% +6.36%* NDCG@100 0.377 0.286 0.314 0.169 0.305 0.355 0.262 0.292 0.159 0.287 +6.25%* +9.27%* +7.55%* +6.58%* +6.34%* Portfolio Theory of Information Retrieval – p. 19/22
  • 20. Evaluations on Ad hoc Text Retrieval (5) Comparison with diversity-based reranking, the MMR [Carbonell and Goldstein(1998)] In each cell, the first line shows the performance of our approach, and the second line shows the performance of the MMR method and gain of our method over the MMR method. Models Dirichlet Jelinek-Mercer BM25 sub-MRR 0.014 0.011 0.009 0.012 (+16.67%*)0.009 (+22.22%*)0.007 (+28.57%*) sub-Recall@5 0.324 0.255 0.275 0.304 (+6.58%*) 0.234 (+8.97%*) 0.27 (+1.85%) sub-Recall@10 0.381 0.366 0.352 0.362 (+5.25%) 0.351 (+4.27%) 0.344 (+2.33%) sub-Recall@20 0.472 0.458 0.464 0.455 (+3.74%) 0.41 (+11.71%*) 0.446 (+4.04%) sub-Recall@100 0.563 0.582 0.577 0.558 (+0.90%) 0.55 (+5.82%*) 0.558 (+3.41%) Portfolio Theory of Information Retrieval – p. 20/22
  • 21. Conclusions We have presented a new ranking theory for IR The benefit of diversification is well quantified Is it a unified theory to explain risk and reward in IR? Portfolio Theory of Information Retrieval – p. 21/22
  • 22. References [Cooper(1971)] William S. Cooper. The inadequacy of probability of usefulness as a ranking criterion for retrieval system output. University of California, Berkeley, 1971. [Robertson(1977)] S. E. Robertson. The probability ranking principle in IR. Journal of Documentation, pages 294–304, 1977. [Gordon and Lenk(1991)] Michael D. Gordon and Peter Lenk. A utility theoretic examination of the probability ranking principle in information retrieval. JASIS, 42(10):703–714, 1991. [Maron and Kuhns(1960)] M. E. Maron and J. L. Kuhns. On relevance, probabilistic indexing and information retrieval. J. ACM, 7(3), 1960. [Stirling(1977)] Keith H. Stirling. The Effect of Document Ranking on Retrieval System Performance: A Search for an Optimal Ranking Rule. PhD thesis, UC, Berkeley, 1977. [Carbonell and Goldstein(1998)] Jaime Carbonell and Jade Goldstein. The use of MMR, diversity-based reranking for reordering documents and producing summaries. In SIGIR, 1998. [Chen and Karger(2006)] Harr Chen and David R. Karger. Less is more: probabilistic models for retrieving fewer relevant documents. In SIGIR, 2006. [Markowitz(1952)] H Markowitz. Portfolio selection. Journal of Finance, 1952. [Järvelin and Kekäläinen(2002)] Kalervo Järvelin and Jaana Kekäläinen. Cumulated gain-based evaluation of IR techniques. ACM Trans. Inf. Syst., 2002. Portfolio Theory of Information Retrieval – p. 22/22