week6
week6
Introduction to
Information Retrieval
Introducing Information Retrieval
and Web Search
Introduction to Information Retrieval
Information Retrieval
Information Retrieval (IR) is finding material (usually
documents) of an unstructured nature (usually text)
that satisfies an information need from within large
collections (usually stored on computers).
2
Introduction to Information Retrieval
250
200
150
Unstructured
100 Structured
50
0
Data volume Market Cap
3
Introduction to Information Retrieval
250
200
150
Unstructured
100 Structured
50
0
Data volume Market Cap
4
Introduction to Information Retrieval Sec. 1.1
5
Introduction to Information Retrieval
Info need
Info about removing mice
without killing them
Misformulation?
Query Searc
how trap mice alive
h
Search
engine
Query Results
Collection
refinement
Introduction to Information Retrieval Sec. 1.1
7
Introduction to Information Retrieval
Introduction to
Information Retrieval
Term-document incidence matrices
Introduction to Information Retrieval Sec. 1.1
Antony and Cleopatra Julius Caesar The Tempest Hamlet Othello Macbeth
Antony 1 1 0 0 0 1
Brutus 1 1 0 1 0 0
Caesar 1 1 0 1 1 1
Calpurnia 0 1 0 0 0 0
Cleopatra 1 0 0 0 0 0
mercy 1 0 1 1 1 1
worser 1 0 1 1 1 0
Incidence vectors
So we have a 0/1 vector for each term.
To answer query: take the vectors for Brutus, Caesar
and Calpurnia (complemented) bitwise AND.
110100 AND
110111 AND
101111 = Antony
Antony and Cleopatra
1
Julius Caesar
1
The Tempest
0
Hamlet
0
Othello
0
Macbeth
1
100100 Brutus
Caesar
1
1
1
1
0
0
1
1
0
1
0
1
Calpurnia 0 1 0 0 0 0
Cleopatra 1 0 0 0 0 0
mercy 1 0 1 1 1 1
worser 1 0 1 1 1 0
12
Introduction to Information Retrieval Sec. 1.1
Answers to query
Antony and Cleopatra, Act III, Scene ii
Agrippa [Aside to DOMITIUS ENOBARBUS]: Why, Enobarbus,
When Antony found Julius Caesar dead,
He cried almost to roaring; and he wept
When at Philippi he found Brutus slain.
13
Introduction to Information Retrieval Sec. 1.1
Bigger collections
Consider N = 1 million documents, each with about
1000 words.
Avg 6 bytes/word including spaces/punctuation
6GB of data in the documents.
Say there are M = 500K distinct terms among these.
14
Introduction to Information Retrieval Sec. 1.1
15
Introduction to Information Retrieval
Introduction to
Information Retrieval
The Inverted Index
The key data structure underlying modern IR
Introduction to Information Retrieval Sec. 1.2
Inverted index
For each term t, we must store a list of all documents
that contain t.
Identify each doc by a docID, a document serial number
Can we used fixed-size arrays for this?
Brutus 1 2 4 11 31 45 173 174
Caesar 1 2 4 5 6 16 57 132
Calpurnia 2 31 54 101
Inverted index
We need variable-size postings lists
On disk, a continuous run of postings is normal and best
In memory, can use linked lists or variable length arrays
Some tradeoffs in size/ease of insertion Posting
Dictionary Postings
Sorted by docID (more later on why).
19
Introduction to Information Retrieval Sec. 1.2
Tokenizer
Token stream Friends Romans Countrymen
Linguistic
modules
Modified tokens friend roman countryman
Indexer friend 2 4
roman 1 2
Inverted index
countryman 13 16
Introduction to Information Retrieval
Doc 1 Doc 2
Why frequency?
Will discuss later.
Introduction to Information Retrieval Sec. 1.2
Terms
and
counts IR system
implementation
• How do we
index efficiently?
• How much
storage do we
need?
Pointers 26
Introduction to Information Retrieval
Introduction to
Information Retrieval
Query processing with an inverted index
Introduction to Information Retrieval Sec. 1.3
29
Introduction to Information Retrieval Sec. 1.3
2 4 8 16 32 64 128 Brutus
1 2 3 5 8 13 21 34 Caesar
30
Introduction to Information Retrieval Sec. 1.3
The merge
Walk through the two postings simultaneously, in
time linear in the total number of postings entries
2 4 8 16 32 64 128 Brutus
1 2 3 5 8 13 21 34 Caesar
The merge
Walk through the two postings simultaneously, in
time linear in the total number of postings entries
2 4 8 16 32 64 128 Brutus
2 8
1 2 3 5 8 13 21 34 Caesar
33
Introduction to Information Retrieval
Introduction to
Information Retrieval
Phrase queries and positional indexes
Introduction to Information Retrieval Sec. 2.4
Phrase queries
We want to be able to answer queries such as
“stanford university” – as a phrase
Thus the sentence “I went to university at Stanford”
is not a match.
The concept of phrase queries has proven easily
understood by users; one of the few “advanced search”
ideas that works
Many more queries are implicit phrase queries
For this, it no longer suffices to store only
<term : docs> entries
Introduction to Information Retrieval Sec. 2.4.1
<be: 993427;
1: 7, 18, 33, 72, 86, 231; Which of docs 1,2,4,5
2: 3, 149; could contain “to be
4: 17, 191, 291, 430, 434; or not to be”?
5: 363, 367, …>
Proximity queries
LIMIT! /3 STATUTE /3 FEDERAL /2 TORT
Again, here, /k means “within k words of”.
Clearly, positional indexes can be used for such
queries; biword indexes cannot.
Exercise: Adapt the linear merge of postings to
handle proximity queries. Can you make it work for
any value of k?
This is a little tricky to do correctly and efficiently
See Figure 2.12 of IIR
Introduction to Information Retrieval Sec. 2.4.2
Rules of thumb
A positional index is 2–4 as large as a non-positional
index
Combination schemes
These two approaches can be profitably combined
For particular phrases (“Michael Jackson”, “Britney
Spears”) it is inefficient to keep on merging positional
postings lists
Even more so for phrases like “The Who”
Williams et al. (2004) evaluate a more sophisticated
mixed indexing scheme
A typical web query mixture was executed in ¼ of the time
of using just a positional index
It required 26% more space than having a positional index
alone
Introduction to Information Retrieval
Introduction to
Information Retrieval
Introducing ranked retrieval
Introduction to Information Retrieval Ch. 6
Ranked retrieval
Thus far, our queries have all been Boolean.
Documents either match or don’t.
Good for expert users with precise understanding of
their needs and the collection.
Also good for applications: Applications can easily
consume 1000s of results.
Not good for the majority of users.
Most users incapable of writing Boolean queries (or they
are, but they think it’s too much work).
Most users don’t want to wade through 1000s of results.
This is particularly true of web search.
Introduction to Information Retrieval Ch. 6
Introduction to
Information Retrieval
Scoring with the Jaccard coefficient
Introduction to Information Retrieval Ch. 6
Introduction to
Information Retrieval
Term frequency weighting
Introduction to Information Retrieval Sec. 6.2
Antony 1 1 0 0 0 1
Brutus 1 1 0 1 0 0
Caesar 1 1 0 1 1 1
Calpurnia 0 1 0 0 0 0
Cleopatra 1 0 0 0 0 0
mercy 1 0 1 1 1 1
worser 1 0 1 1 1 0
Antony and Cleopatra Julius Caesar The Tempest Hamlet Othello Macbeth
Antony 157 73 0 0 0 0
Brutus 4 157 0 1 0 0
Caesar 232 227 0 2 1 1
Calpurnia 0 10 0 0 0 0
Cleopatra 57 0 0 0 0 0
mercy 2 0 3 5 5 1
worser 2 0 1 1 1 0
Introduction to Information Retrieval
Term frequency tf
The term frequency tft,d of term t in document d is
defined as the number of times that t occurs in d.
We want to use tf when computing query-document
match scores. But how?
Raw term frequency is not what we want:
A document with 10 occurrences of the term is more
relevant than a document with 1 occurrence of the term.
But not 10 times more relevant.
Relevance does not increase proportionally with
term frequency.
NB: frequency = count in IR
Introduction to Information Retrieval Sec. 6.2
Log-frequency weighting
The log frequency weight of term t in d is
1 + log 10 tf t,d , if tf t,d > 0
w t,d =
0, otherwise
Introduction to
Information Retrieval
(Inverse) Document frequency weighting
Introduction to Information Retrieval Sec. 6.2.1
Document frequency
Rare terms are more informative than frequent terms
Recall stop words
Consider a term in the query that is rare in the
collection (e.g., arachnocentric)
A document containing this term is very likely to be
relevant to the query arachnocentric
→ We want a high weight for rare terms like
arachnocentric.
Introduction to Information Retrieval Sec. 6.2.1
idf weight
dft is the document frequency of t: the number of
documents that contain t
dft is an inverse measure of the informativeness of t
dft ≤ N
We define the idf (inverse document frequency) of t
by
idf t = log10 ( N/df t )
We use log (N/dft) instead of N/dft to “dampen” the effect
of idf.
28
Introduction to Information Retrieval
29
Introduction to Information Retrieval Sec. 6.2.1
Introduction to
Information Retrieval
tf-idf weighting
Introduction to Information Retrieval Sec. 6.2.2
tf-idf weighting
The tf-idf weight of a term is the product of its tf
weight and its idf weight.
w t ,d = (1 + log tf t ,d ) × log10 ( N / df t )
Best known weighting scheme in information retrieval
Note: the “-” in tf-idf is a hyphen, not a minus sign!
Alternative names: tf.idf, tf x idf
Increases with the number of occurrences within a
document
Increases with the rarity of the term in the collection
Introduction to Information Retrieval Sec. 6.2.2
Score(q,d) = tf.idft,d
t ∈q∩d
34
Introduction to Information Retrieval Sec. 6.3
Introduction to
Information Retrieval
The Vector Space Model (VSM)
Introduction to Information Retrieval Sec. 6.3
Documents as vectors
Now we have a |V|-dimensional vector space
Terms are axes of the space
Documents are points or vectors in this space
Very high-dimensional: tens of millions of
dimensions when you apply this to a web search
engine
These are very sparse vectors – most entries are zero
Introduction to Information Retrieval Sec. 6.3
Queries as vectors
Key idea 1: Do the same for queries: represent them
as vectors in the space
Key idea 2: Rank documents according to their
proximity to the query in this space
proximity = similarity of vectors
proximity ≈ inverse of distance
Recall: We do this because we want to get away from
the you’re-either-in-or-out Boolean model
Instead: rank more relevant documents higher than
less relevant documents
Introduction to Information Retrieval Sec. 6.3
Length normalization
A vector can be (length-) normalized by dividing each
of its components by its length – for this we use the
L2 norm:
x 2 = i xi2
cosine(query,document)
Dot product Unit vectors
V
q•d q d q di
i =1 i
cos( q , d ) = = • =
q d
qd
i=1 i
V 2 V
2
q
i =1 i
d
cos(q, d ) = q • d = qi di
V
i=1
for q, d length-normalized.
47
Introduction to Information Retrieval
48
Introduction to Information Retrieval Sec. 6.3
Sensibility jealous 10 7 11
gossip 2 0 6
PaP: Pride and
wuthering 0 0 38
Prejudice, and
WH: Wuthering Term frequencies (counts)
Heights?
cos(SaS,PaP) ≈
0.789 × 0.832 + 0.515 × 0.555 + 0.335 × 0.0 + 0.0 × 0.0 ≈
0.94
cos(SaS,WH) ≈ 0.79
cos(PaP,WH) ≈ 0.69
Why do we have cos(SaS,PaP) > cos(SAS,WH)?
Introduction to Information Retrieval
Introduction to
Information Retrieval
Calculating tf-idf cosine scores
in an IR system
Introduction to Information Retrieval Sec. 6.4
Introduction to
Information Retrieval
Evaluating search engines
Introduction to Information Retrieval Sec. 8.6
Evaluating an IR system
An information need is translated into a query
Relevance is assessed relative to the information
need not the query
E.g., Information need: I’m looking for information on
whether drinking red wine is more effective at
reducing your risk of heart attacks than white wine.
Query: wine red white heart attack effective
You evaluate whether the doc addresses the
information need, not whether it has these words
63
Introduction to Information Retrieval Sec. 8.4
Recall/Precision
R P
1 R 0.1 1.0
2 N 0.1 0.5 Top 10 returned
3 N 0.1 0.33 docs ranked top-down
4 R 0.2 0.5
5 R 0.3 0.6
6 N Assume 10 rel docs
…
7 R in collection
8 N
9 N
10 N
Introduction to Information Retrieval Sec. 8.4
R N N R R N R R …
P: 1.0 0.5 0.6 0.58 0.62
AP=avg(1.0,0.5,0.6,0.58,0.62,…)
MAP=avg(AP1, AP2, AP3, …) 66