Slides of the presentation given at ACM SAC 2018 for the following paper:
Daniel Valcarce, Javier Parapar, Alvaro Barreiro: LiMe: Linear Methods for Pseudo-Relevance Feedback. SAC 2018: 678-687.
http://dx.doi.org/10.1145/3167132.3167207
LiMe: Linear Methods for Pseudo-Relevance Feedback [SAC '18 Slides]
1. ACM SAC 2018, Pau, France
LiMe: Linear Methods for Pseudo-Relevance Feedback
Daniel Valcarce Javier Parapar Álvaro Barreiro
@dvalcarce @jparapar @AlvaroBarreiroG
Information Retrieval Lab
University of A Coruña
Spain
4. Pseudo-Relevance Feedback (I)
Pseudo-Relevance Feedback provides an automatic method for
query expansion:
First retrieval with the original query
◦ Top retrieved documents are assumed to be relevant
(pseudo-relevant set)
Expand the query with terms from the pseudo-relevant set
Second retrieval with the expanded query
◦ The expanded query usually performs better than the
original one
3
14. LiMe
LiMe, as a PRF technique:
models the PRF task a matrix decomposition problem
employs linear methods to provide a solution
is able to learn inter-term similarities
jointly models the query and the pseudo-relevant set
admits different feature schemes to represent documents
and queries
is agnostic to the retrieval model
6
15. Notation
Some notation:
A user prompts a query Q
The collection C is composed of documents D
V denotes the vocabulary and is formed of terms t
We denote the pseudo-relevant set by F
And the extended pseudo-relevant set by F {Q} ∪ F
◦ Its cardinal is m |F | |F| + 1
◦ And its vocabulary VF has size n |VF | ≤ |V|
7
16. LiMe: Matrix Formulation
Let X ∈ Rm×n be the extended pseudo-relevant set matrix, we
aim to find a inter-term similarity matrix W ∈ Rn×n
+ such that:
X X × W
Q
D1
. . .
Dm−1 m×n
Q
D1
. . .
Dm−1 m×n
×
w11 · · · w1n
...
...
...
wn1 · · · wnn n×n
s.t. diag(W) 0, W ≥ 0
8
18. LiMe: Feature Schemes (I)
How do we fill matrix X
Q
D1
. . .
Dm−1 m×n
?
xij
s(tj , Q) if i 1 and f (tj , Q) > 0,
s(tj , Di−1) if i > 1 and f (tj , Di−1) > 0,
0 otherwise
s(t, D): weighting function of term t in D (or Q)
f (t, D): #occurrences of term t in D (or Q)
9
19. LiMe: Feature Schemes (II)
We tested two well-known Information Retrieval weighting
functions:
TF
stf(w, D) 1 + log2 f (w, D)
TF-IDF
stf-idf(w, D) 1 + log2 f (w, D) × log2
|C|
df(w)
10
21. LiMe: Optimization Problem
W∗
arg min
W
1
2
X − X W 2
F + β1 W 1,1 +
β2
2
W 2
F
s.t. diag(W) 0, W ≥ 0
(1)
Bound constrained least squares optimization problem with
elastic net ( 1 and 2 regularization) penalty:
ìw∗
·j arg min
ìw·j
1
2
ìx·j − X ìw·j
2
2
+ β1 ìw·j 1
+
β2
2
ìw·j
2
2
s.t. wjj 0, ìw·j ≥ 0
(2)
11
22. LiMe: Query Expansion
To expand the original query, we reconstruct the first row of X:
Q
1×n
Q
1×n
×
w11 · · · w1n
...
...
...
wn1 · · · wnn n×n
ˆx1· ìx1· × W∗
(3)
12
23. LiMe: Query Expansion
To expand the original query, we reconstruct the first row of X:
Q
1×n
Q
1×n
×
w11 · · · w1n
...
...
...
wn1 · · · wnn n×n
ˆx1· ìx1· × W∗
(3)
We compute a probabilistic estimate of a term tj given the
feedback model θF:
p(tj |θF)
ˆx1j
tv ∈VF
ˆx1v
if tj ∈ VF ,
0 otherwise
(4)
12
24. LiMe: Second retrieval
The second retrieval is performed interpolating the original
query model with the feedback model:
p(t|θQ) (1 − α) p(t|θQ) + α p(t|θF) (5)
The hyperparameter α controls the interpolation
This is a standard procedure in state-of-the-art PRF
techniques
13
28. Evaluation Metrics
We produce a ranking of 1000 documents per query:
MAP Mean Average Precision
nDCG Normalised Discounted Cumulative Gain
RI Robustness Index:
#topics improved−#topics degraded
#topics
17
36. Conclusions
LiMe:
is a PRF technique that shows state-of-the-art performance
can be plugged on top of any retrieval model
accepts different feature schemes
models inter-term similarities
25
37. Future work
Alternative feature schemes based on:
retrieval features
query logs
Explore connection with Translation Models which also rely on
inter-term similarities:
learnt from training data [Berger & Lafferty, SIGIR ’99]
based on mutual information [Karimzadehgan & Zhai,
SIGIR ’10]
26