Multi Target Prediction
Multi Target Prediction
Krzysztof Dembczyński
Intelligent Decision Support Systems Laboratory (IDSS)
Poznań University of Technology, Poland
1 / 102
Multi-target prediction
2 / 102
Image annotation/retrieval
3 / 102
Multi-label classification
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 1 1 0
x2 2.0 2.5 0 1 0
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 0 1 1
x 4.0 2.5 ? ? ?
4 / 102
Multi-label classification
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 1 1 0
x2 2.0 2.5 0 1 0
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 0 1 1
x 4.0 2.5 1 1 0
4 / 102
Ecology
5 / 102
Multi-variate regression
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 14 0.3 9
x2 2.0 2.5 15 1.1 4.5
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 19 0.9 2
x 4.0 2.5 ? ? ?
6 / 102
Multi-variate regression
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 14 0.3 9
x2 2.0 2.5 15 1.1 4.5
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 19 0.9 2
x 4.0 2.5 18 0.5 1
6 / 102
Label ranking
X1 X2 Y1 Y2 Ym
x1 5.0 4.5 1 3 2
x2 2.0 2.5 2 1 3
.. .. .. .. ..
. . . . .
xn 3.0 3.5 3 1 2
x 4.0 2.5 ? ? ?
1
E. Hüllermeier, J. Fürnkranz, W. Cheng, and K. Brinker. Label ranking by learning pairwise
preferences. Artificial Intelligence, 172:1897–1916, 2008
7 / 102
Label ranking
X1 X2 Y1 Y2 Ym
x1 5.0 4.5 1 3 2
x2 2.0 2.5 2 1 3
.. .. .. .. ..
. . . . .
xn 3.0 3.5 3 1 2
x 4.0 2.5 1 2 3
1
E. Hüllermeier, J. Fürnkranz, W. Cheng, and K. Brinker. Label ranking by learning pairwise
preferences. Artificial Intelligence, 172:1897–1916, 2008
7 / 102
Multi-task learning
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 14 9
x2 2.0 2.5 1.1
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 2
x 4.0 2.5 ?
8 / 102
Multi-task learning
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 14 9
x2 2.0 2.5 1.1
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 2
x 4.0 2.5 1
8 / 102
Collaborative filtering2
m1 m2 m3 · · · mm
u1 1 ··· 4
u2 3 1 ···
u3 2 5 ···
... ···
un 2 ··· 1
2
D. Goldberg, D. Nichols, B.M. Oki, and D. Terry. Using collaborative filtering to weave and
information tapestry. Communications of the ACM, 35(12):61–70, 1992
9 / 102
Dyadic prediction3
4 5 ··· 7 8 6
10 14 · · · 9 21 12
instances y1 y2 · · · ym y m+1 y m+2
1 1 x1 10 ? ··· 1 ? ?
3 5 x2 0.1 · · · 0 ?
7 0 x3 ? ? ··· 1 ?
1 1 ... ··· 0 ?
3 1 xn 0.9 · · · 1 ? ?
2 3 xn+1 ? ··· ? ?
3 1 xn+2 ? ··· ? ? ?
3
A.K. Menon and C. Elkan. Predicting labels for dyadic data. Data Mining and Knowledge
Discovery, 21(2), 2010
10 / 102
Multi-target prediction
y1 , y2 , . . . , ym
I A multitude of multivariate loss functions defined over the output
vector
`(y, h(x))
• Main question:
I Can we improve over independent models trained for each target?
• Two views:
I The individual-target view
I The joint-target view
11 / 102
The individual target view
12 / 102
The joint target view
13 / 102
The individual and joint target view
14 / 102
Multi-target prediction
the individual
target view
15 / 102
Target interdependences
16 / 102
Target interdependences
• Model similarities:
17 / 102
Target interdependences
I Hierarchies,
I General graphs,
I ...
18 / 102
Target interdependences
19 / 102
Multivariate loss functions
20 / 102
The individual target view
• Learning algorithms
I Pooling.
I Stacking.
I Regularized multi-target learning.
• Problem settings
I Multi-label classification.
I Multivariate regression.
I Multi-task learning.
21 / 102
A starting example
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 1 1 0
x2 2.0 2.5 0 1 0
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 0 1 1
x 4.0 2.5 ? ? ?
22 / 102
A starting example
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 1 1 0
x2 2.0 2.5 0 1 0
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 0 1 1
x 4.0 2.5 1 1 0
22 / 102
Loss functions and optimal predictions
`(yi , ŷi )
• The loss function can be also written over all targets as:
m
X
`(y, ŷ) = `(yi , ŷi )
i=1
X Y
z }| { z }| {
xT1
x11 · · · x1p y1
.. .. .. ..
= . →
. . .
xn1 · · · xnp xTn yn
24 / 102
Single output regression vs. multivariate regression
25 / 102
Pooling
26 / 102
Pooling
27 / 102
Pooling
28 / 102
Pooling
• The critical sample size (dashed line) depends on the model similarity,
which is normally not known!
• To pool or not to pool? Or maybe pooling to some degree?
29 / 102
James-Stein estimator
(m − 2)σ 2
θ̂JS+ = 1− (y − v) + v
ky − vk2
31 / 102
James-Stein estimator
• Works best when the norm of the mean vector is close to zero.5
33 / 102
Mean-regularized multi-target learning6
Target 1
• Simple assumption:
models for different targets
are related to each other.
• Simple solution: the
parameters of these models Target 2 Mean Target 4
• Disadvantage: the
assumption of all target
models being similar might
m m
be invalid for many X 1 X
applications. min kY−XAkF +λ kai − aj k2
A m
i=1 j=1
6
Evgeniou and Pontil. Regularized multi-task learning. In KDD 2004
34 / 102
Multi-target prediction methods
i.e., the output space (possibly along with the feature space) is first
transformed, and than univariate (regression) methods are then
trained on the new output variables h−1 (y, x).
35 / 102
Stacking applied to multi-target prediction: general principle8
Level 2 h1 h2 h3 h4
Level 1 f1 f2 f3 f4
8
W. Cheng and E. Hüllermeier. Combining instance-based learning and logistic regression for
multilabel classification. Machine Learning, 76(2-3):211–225, 2009
36 / 102
Multivariate regression methods
9
L. Breiman and J. Friedman. Predicting multivariate responses in multiple linear regression. J.
R. Stat. Soc., Ser. B, 69:3–54, 1997
10
A. Izenman. Reduced-rank regression for the multivariate linear model. J. Multivar. Anal.,
5:248–262, 1975
11
A. an der Merwe and J.V. Zidek. Multivariate regression analysis and canonical variates. Cana-
dian Journal of Statistics, 8:27–39, 1980
37 / 102
Multivariate regression methods
38 / 102
The joint target view
• Learning algorithms
I Reduction algorithms.
I Conditional random fields (CRFs).
I Structured support vector machines (SSVMs).
I Probabilistic classifier chains (PCCs).
• Problem settings
I Hamming and subset 0/1 loss minimization.
I Multilabel ranking.
I F-measure maximization.
39 / 102
A starting example
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 1 1 0
x2 2.0 2.5 0 1 0
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 0 1 1
x 4.0 2.5 ? ? ?
40 / 102
A starting example
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 1 1 0
x2 2.0 2.5 0 1 0
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 0 1 1
x 4.0 2.5 1 1 0
40 / 102
Two basic approaches
X1 X2 Y1 Y2 ... Ym
x1 5.0 4.5 1 1 0
x2 2.0 2.5 0 1 0
.. .. .. .. .. ..
. . . . . .
xn 3.0 3.5 0 1 1
41 / 102
Synthetic data
1.0
●●
●●
●●
● ●●●● ●●●●●
●●●
●●●
●●●●
●●
●● ● ●●● ●
●●
●●●
● ●● ● ●
●●● ● ● ●
● ●●●
● ●
● ●● ●●● ●● ●● ●
●●●● ●●
●●
●●
● ●●●● ●●●●●
●●●
●●●
●●●●
●●
●● ● ●●● ●
●●
●●●
● ●● ● ●
●●● ● ● ●
● ●●●
● ●
● ●● ●●● ●● ●● ●
●●●●
● ● ● ● ●●●●
●●● ●
●●
●● ● ●● ● ● ●● ●● ●●●●● ● ●
● ● ● ● ● ● ●●●●
●●● ●
●●
●● ● ●● ● ● ●● ●● ●●●●● ● ●
● ●
●●
● ●
●●●
●
●
●
●
●●
●
●
●●
● ●●
●●●
●●
●
●
●●●
● ●●
●
●
●
●
●
●
●
●●
●●●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●●●●
●●●
●
●● ●
●●
●
●
●
●●●
● ●●●●
●
●●●●
●
●
●●●●
●
●●
●●●●●●
●●
●
● ●●
●●●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●●●
● ●●
● ●
●●●
●
●
●
●
●●
●
●
●●
● ●●
●●●
●●
●
●
●●●
● ●●
●
●
●
●
●
●
●
●●
●●●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●●●●
●●●
●
●● ●
●●
●
●
●
●●●
● ●●●●
●
●●●●
●
●
●●●●
●
●●
●●●●●●
●●
●
● ●●
●●●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●●●
●
●
● ● ●● ●
●● ●● ●
●●●
● ●● ● ●
● ●●●●● ●
●●●
●● ●●● ● ● ●●●●
●●● ● ●●● ● ●● ●●●●
●
●●
●●
●
●●●●● ●
●●
●
●●
●●
● ●
● ● ●● ●
●● ●● ●
●●●
● ●● ● ●
● ●●●●● ●
●●●
●● ●●● ● ● ●●●●
●●● ● ●●● ● ●● ●●●●
●
●●
●●
●
●●●●● ●
●●
●
●●
●●
●
●●●●●●● ●●●●●
●● ●●●● ●● ●●
● ●
● ●
●●
● ●● ● ● ●●
●●● ● ● ●● ● ●
● ●
●● ●
● ●
●
● ● ● ●●
●●● ●
●●● ●●●●●●● ●●●●●
●● ●●●● ●● ●●
● ●
● ●
●●
● ●● ● ● ●●
●●● ● ●●● ● ●
● ●
●● ●
● ●
●
● ● ● ●●
●●● ●
●●●
●●●●● ●●● ●●● ●
●●
●
●●● ●
●●●
●●●●● ●
●●●●●● ●●
●
●
● ●
●●● ●
● ●●
●●●●●●● ●●
● ●
●●●●●●●●
●●
●●●●●● ●●
●●●
● ●● ●
● ●●●●● ●●● ●●● ●
●●
●
●●● ●
●●●
●●●●● ●
●●●●●● ●●
●
●
● ●
●●● ●
● ●●
●●● ●
●●● ●●
● ●
●●●●●●●●
●●
●●●●●● ●●
●●●
● ●● ●
●
●●● ●●● ●●● ●●●
● ●
●●
●
●
●
● ●●
●●●
●●●●
●●
●●
●●
●
●
●●
●
●●●●
●
●
● ●
●●●
● ●
● ● ●●●● ●
● ●●
●●●
●
●● ●●
●● ●
●●● ● ●
●
● ●
● ●●● ●●● ●●● ●●● ●●●
● ●
●●
●
●
●
● ●●
●●●
●●●●
●●
●●
●●
●
●
●●
●
●●●●
●
●
● ●
●●●
● ●
● ● ●●● ●●
● ●● ●
●● ●●
●● ●
●●● ● ●
●
● ●
● ●●●
●
●●●
●
●●●
●●
●
●●
●●
●
●● ●●
●
●● ●●
●●●
●●
●
●●●
●● ●●
●
●
●
●
●●●
● ●●●
● ●
●●●
●●
●●
●●
● ●
●
●
●●
●●
●●●●
●
●●
●●
●●●
● ●
●●● ●●●● ●●●●●
● ●●●●●●●
● ●●
●●●
●
●
●●
●
●●●
●
● ●
●●●
●
●●●
●●
●
●●
●●
●
●● ●●
●● ●●
●●●
●●
●
●●●
●● ●●
●
●
●
●
●●●
● ●●●
● ●
●●●
●●
●●
●●
● ●
●
●
●●
●●
●●●●
●
●●
●●
●●●
● ●
●
●●●
●● ●●●● ●●●●●
● ●●●●●●●
● ●●
●●●
●
●
●●
●
●●●
●
●
●●
●● ●
●●●●● ●●●
● ●
●●
●●
●
●
● ● ●●●
●
●●●●●●●● ●●
●●
●● ●
●●
●
●●
●●●●● ●● ●
●
●●
●
●●
●●●
● ●●● ●
● ●
●●●
●
●●●● ●●
●●
●●●●●●●
●●●●
●
●
●●●●
●
●
● ● ●●
●● ●
●●●●● ●●
●●
● ●
●●
●●
●
●
● ● ●●●
●
●●●●●●●● ●●
●●
●● ●
●●
●
●●
●●●●● ●● ●
●
●●
●
●●●
●●
● ●●● ●
● ●
●●●
●
●●●● ●●
●●
●●●●●●●
●●●●
●
●
●●●●
●
●
● ●
0.5
0.5
● ● ● ● ●●
● ● ● ● ● ●●●
●● ●● ● ● ●● ●●●●
● ● ●●● ● ●●●
● ●
● ● ● ● ● ●●
● ● ●
● ● ● ●●●
●● ● ● ● ●● ●●●●
● ● ●●● ● ●●●
● ●
●
●●
● ●●
●
●● ●● ●●●●
● ●
●●●●
●
●●●●●●
●
●
●
●
●●● ●●●●●●
●● ●●
●●
●●
●
●●
● ● ●
●●
●●
●
●●●
●●
●● ● ●
●●● ●
● ●
● ● ●● ●●
●
● ●●
●
● ●
●●● ●●
● ●●
●
●● ●● ●●●●
● ●
●●●●
●
●●●●●●
●
●
●
●
●●● ●●●●●●
●● ●
●●
●●
●●
●
●●
● ● ●
●●
●●
●
●●●
●
●
●● ● ●
●●● ●
● ●
● ● ●● ●●
●
● ●●
●
● ●
●●●
●
●●●
● ●
●●
●
●●●
●●
●●●
● ●●
●● ●
●●
●●
●
●●
●
●
●
●●●●●●
●
●
●
●
●
●●
●●●●
●
●
●
●
●●●
●●●
●
●●
●●●
●●
●●●●●●●
●
●●●●
●
●●
●
●●●
● ●●
●
● ●
●
●●●●
●●
●
●
●
●
●●
●
●●●●●●
●●
●
●
●
●
●
●●
●
●
●●
●●●
● ●
● ●
●●●
● ●
●●
●
●●●
●●
●●●
● ●●
●● ●
●●
●●
●
●●
●
●
●
●●●●●●
●
●
●
●
●
●●
●●●●
●
●
●
●
●●●
●●●
●
●●
●●●
●●
●●●●●●●●
●
●●●
●
●●
●
●●●
● ●●
●
● ●
●
●●●●
●●
●
●
●
●
●●
●
●●●●●●
●●
●
●
●
●
●
●●
●
●
●●
●●●
● ●
●
●
● ●●
●●●
●
●● ●●● ● ●
●
●● ●
●●●
●●●●●●●
●●●●
●●
●
●●● ●
●
●
●
●
●● ●
●
●
●
●
●● ● ●●
● ●
●
●●
●●
●●
●●
●●● ●
● ●
●●●●●
●●●
●● ●●
●
●
●●● ●●
●
● ●
● ● ●
●●● ●
● ●●
●●●
●
●● ●●● ● ●
●
●● ●
●●●
●●●●●●●
●●●●
●●
●
●●● ●
●
●
●
●
●● ●
●
●
●
●
●● ● ●●
● ●
●
●●
●●
●●●
●
●●● ●
● ●
●●●●●
●●●
●● ●●
●
●
●●● ●●
●
● ●
● ● ●
●●●
●●● ● ●●●
● ●
●● ●●●
● ●●●●●●●● ●
● ● ●●● ● ●●● ●●●● ● ●●●
●●●● ● ●● ●●● ● ●●●
● ●● ●●
● ●●●●●●●● ●
● ● ●●● ● ●●● ●●●● ● ●●●
●●●● ● ●●
●●
●●
●
● ●●
●●
●●●●●●●
● ●
●●●●●
●
● ● ●
●
●●
●
●
●
●●
●●
● ●●●
●●●
●●●●●●●
●
●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●●
●
●
●
●
●
●
●
●●
●
●●●
●●●●
●
●●●●
●●
●
●
●● ●●●
●
●
●●
●
●
●
●●
● ●
●
●
●●
●
●
●●
●
●
●
●
●●
●●
● ●
●
●
●
●
●●
●
●
●
●●
●●
●
●
●
●●
●●
●
● ●●
●●
●●●●●●●
● ●
●●●●●
●
● ● ●
●
●●
●
●
●
●
●
●
●●
● ●●●
●
●●●
●●●●●●●
●
●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●●
●
●
●
●
●
●
●
●●
●
● ●
●●
●
●●
●
●●●●
●●
●
●
●● ●●●
●
●
●●
●
●
●
●●
● ●
●
●
●●
●
●
●●
●
●
●
●
●●
●●
● ●
●
●
●
●
●●
●
●
●
●●
●●
●
●
●
●
● ●
●●
●
●
●●● ●
● ●●●● ●
●
●
●●●●
●
● ●
●●●
●●
●● ●
● ●●
● ●●
●● ●●●● ● ●
●● ●
●
●●
●●● ●●
●●
● ●
●●● ● ●
●●●●●●● ●●
●●●●
●●
● ●
●● ●
● ●
●●
●
●
●●● ●
● ●●●● ●
●
●
●●●●
●
● ●
●●●
●●
●● ●
● ●●
● ●●
●● ●●●● ● ●
●● ●
●●
●●
●● ●●
●●
● ●
●●● ● ●
●●●●●●● ●●
●●●●
●●
● ●
●●
●●● ●●● ●● ● ●●●● ●●● ●
●●●●● ●●●●● ●●● ● ● ● ●
●●●●●●
● ●● ● ●●●● ●●● ●● ●● ● ●●●● ●●● ●
●●●●● ●●●●● ●●● ● ● ● ●
●●●●●● ●● ● ●●●●
●●
●●●●●
●● ●
● ● ● ●●
●
●●
●●●●●
●
●●
●
●●●●●● ●●●
●●●
●● ●●● ● ●●●
●●● ●
● ●●
●●●●
●●●●●
●●●●●
●● ●●●●●●●● ●
●
●●
● ●●
●●●●●
●● ●
●
● ● ● ●●
●
●●
●●●●●
●
●●
●
●●●●●● ●●●
●●●
●● ●●● ● ●●●
● ●
● ●
● ●●
●●●●●
●●●●
●●●●●
●
●● ●●●●●●●● ●
●
●●
●
●
● ●●
●
●●●
● ● ●● ●●
●●●
●●● ●
●●
●
●●
●●● ●
● ●
●
●●●
● ●●●
●
●● ●● ●●● ●
●●●
● ●
●●●●
●●
●
●
● ●●●●●
●●
●● ●●●
●●●●●●● ●
●
●
● ●● ● ● ●
● ●●
●
●●●
● ● ●● ●●
●●●
●●● ●
●●
●
●●
●●● ●
● ●
●
●●●
● ●●●
●
●● ●● ●●● ●
●●●
● ●
●●●●
●●
●
●
● ●●●●●
●●
●● ●●●
●●●●●●● ●
●
●
● ●● ● ●
● ● ●● ● ● ● ● ● ● ● ●●● ● ● ●● ● ● ● ● ● ● ● ●●●
0.0
●●
0.0
● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ●●
●
●
●
●
●
●
●●
●
●●
●●●
●● ● ●
●
●
●●●
●
●
●●
●
●
●
●●●
●
●●●
●●●●
●
●●
●
● ●
●●●
●
●
●
●
●● ●
●●
●●
● ●
●●
●● ●●●●
●
●●
●●
● ●●●
● ●
● ●●●
●●
●●●
●
●●
●●●
●
●
●●●
● ●
●
●●● ●
●
●●
●
●
●
●●
●●
●
●●●●
●
●●
●
●
●
●
●
●
●●
●
●●
●●●
●● ● ●
●
●
●●●
●
●
●●
●
●
●
●●●
●
●●●
●●●●
●
●●
●
● ●
●●●
●
●
●
●
●● ●
●●
●●
● ●
●●
●● ●●●●
●
●●
●●
● ●●●
● ●
● ●●●
●●
●●●
●
●●
●●●
●
●
●●●
● ●
●
●●● ●
●
●●
●
●
●
●●
●●
●
●●●●
●
●●
●
●
●
●●
●●●
●
●
●
●●
● ●●●
●
●
●
●●●
●●●●
●
●
●
●●
●
● ● ●●●●
●
●
●●●
●
●●●●●
●●●
●●●
●
●●
●●●
●
●
●●
●●
●
●
●
●
●
●●●●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
● ●
●
●
●
●
●
●
●●
●●●
● ●
●
●
●●●
●●●
●
●
●●
●
●
●
●●
●
●●
●●
●
●●●● ●
●
●
●●
●●●
●
●
●
●●
● ●●●
●
●
●
●●●
●●●●
●
●
●
●●
●
● ● ●●●●
●
●
●●●
●
●●●●●
●●●
●●●
●
●●
●●●
●
●
●●
●●
●
●
●
●
●
●●●●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
● ●
●
●
●
●
●
●
●●
●●●
● ●
●
●
●●●
●●●
●
●
●●
●
●
●
●●
●
●●
●●
●
●●●●
● ●● ●
●● ●●●●● ●
● ●
●
●●●● ●●
●●●●● ●●● ●
●
●●
●
● ● ●●● ●●
●● ●●●● ●
●●●● ●●●● ●
●●● ●
●●● ● ● ●●●●●●●
●● ● ● ●● ●
●● ●●●●● ●
● ●
●
●●●● ●●
●●●●● ●●● ●
●
●●
●
● ● ●●● ●●
●● ●●●● ●
●●●● ●●●● ●
●●● ●
●●● ● ● ●●●●●●●
●● ●
●●●
● ● ●
●● ● ●●
●●
●● ●●
● ●●● ● ●
● ● ● ●●● ●●●●●
●● ●
● ●
●●●● ●●
● ●
● ●●●●
●
● ●
●●●●● ●●
● ● ●● ● ●●●
● ● ●
●● ● ●●
●●
●● ●●
● ●●● ● ●
● ● ● ●●● ●●●●●
●● ●●●
●●●● ●●
● ●
● ●●●●
●
● ●
●●●●● ●●
● ● ●● ●
● ●
●●●● ●●
● ●
●● ● ●● ●
●
●● ●
●
●●●●
● ●
● ●
●●●●
●●●
●● ●●
●●
●●●
●●●●
●●● ● ●
●●●● ●
●●●
●●● ●
●●●● ●
●●
●●●
●●●●●●●● ●
●● ●● ●
●●●●●
● ● ● ●
●●●● ●●
● ●
●● ● ●● ●
●
●● ●
●
●●●●
● ●
● ●
●●●●
●●●
●● ●●
●●
●●●
●●●●
●●● ● ●
●●●● ●
●●●
●●● ●
●●●● ●
●●
●●●
●●●●●●●● ●
●● ●● ●
●●●●●
● ●
●
●● ● ●● ●
●
●●● ●●● ● ● ●●●●
● ●● ●● ●●
●● ●
●● ●●●●
● ● ●●
● ● ●● ● ●
●● ● ●● ●
●
●●● ●●● ● ● ●●●●
● ●● ●● ●●
●● ●
●● ●●●●
● ● ●●
● ● ●● ●
●
●●
●●
●
●
●
●●
● ●
●
●●
●
●
● ●●
●●
●●●●
●●
●●
●
●●
●● ●●●
●
●●
●
●● ●
●●
● ●●
●
●
●●●
● ●
●●●● ●● ●
●●
●●●●●● ●
●
● ●
●● ●
●●
●●
●●● ●
●●●●
●
●
●●
●
●●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●●●●
●●
●
●
●
●
●●
●● ●
●●
●●
●
●
●
●●
● ●
●
●●
●
●
● ●●
●●
●●●●
●●
●●
●
●●
●● ●●●
●
●●
●
●● ●
●●
● ●●
●
●
●●●
● ●
●●●● ●● ●
●●
●●●●●● ●
●
● ●
●● ●
●●
●●
●●● ●
●●●●
●
●
●●
●
●●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●●●●
●●
●
●
●
●
●●
●●
●●●
●
●
●●
●
●●
● ●●●● ● ●●●●
●
●●
●
●●●●
●●● ●
●●●
● ●● ●●●●
●●
● ●●
●●
●
●
●
●●●●
● ●●●●
●●●●●
●●
●●
●●
●●
●● ●●
● ●●●●
●●●
●●●●● ● ●
●●
● ●●●●
●
●
●
●●
●
● ●●●
●
●
●●
●
●●
● ●●●● ● ●●●●
●
●●
●
●●●●
●●● ●
●●●
● ●● ●●●●
●●
● ●●
●●
●
●
●
●●●●
● ●●●●
●●●●●●
●●
●●
● ●●
●● ●●
● ●●●●
●●●
●●●●● ● ●
●●
● ●●●●
●
●
●
●●
●
●
−0.5
● ●● ●●
−0.5
●● ●● ●● ●● ●●●●●●●
● ●● ●
●●●●●●
● ● ● ●●● ●● ●
● ●● ●●●
● ●
●
●● ● ●●
● ●
● ● ●●
●●●●●● ●
●
●●● ●● ●
●● ●● ●● ●●●●●●●
● ●● ●
●●●●●●
● ● ● ●●● ●● ●
● ●● ●●●
● ●
●
●● ● ●●
● ●
● ● ●●
●●●●●● ●
●
●●●
●
●●
●
●●
●
●
●●
●
●●
●
●
●
●
●
● ●●
● ●
●
●
●
●
●●
●● ●
●
●
●
●
●
●
● ●
●
●●
●
●●
●●●●●
●●
●
●
●
●
●
●●
●●●
●●●
●●●●
●●
●
●
●
●
●
●
●
● ●●
●
●●●
●
●
●
●●
●●●
●●●
●●●
●●●●
●
●●
●●●●
●
●
●
●
●●
●
●
● ●●
●
●●
● ●
●
●●●
●
●
●●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●●
●
●●
●
● ●
●
●
●●
●
●●
●
●
●●
●
●●
●
●
●
●
●
● ●●
● ●
●
●
●
●
●●
●● ●
●
●
●
●
●
●
● ●
●
●●
●
●●
●●●●●
●●
●
●
●
●
●
●●
●●●
●●●
●●●●
●●
●
●
●
●
●
●
●
● ●●
●
●●●
●
●
●
●●
●●●
●●●
●●●
●●●●
●
●●
●●●●
●
●
●
●
●●
●
●
● ●●
●
●●
● ●
●
●●●
●
●
●●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●●
●
●●
●
● ●
●
● ●●
●
●●● ●● ● ●
●●●●
●●● ●
●●●●
● ●●
● ● ●●
● ●●●●● ● ● ● ●
●●●
●●
● ●●
●
●● ●
● ●●
●● ●●
●
●● ●●● ●●
●●● ●
●●
●
●●●●●
●● ● ●
●● ● ●●
●
●●● ●● ● ●
●●●●
●●● ●
●●●●
● ●●
● ● ●●
● ●●●●● ● ● ● ●
●●●
●●
● ●●
●
●● ●
● ●●
●● ●●
●
●● ●●● ●●
●●● ●●
●
●●●
● ●●
●● ● ●
●●
●●
●●●
●● ●●●●
●●●● ●
● ●●
●● ●
●●
●
●
●● ●●
● ●
●● ● ●● ●●●
● ●
●●
● ●● ●● ● ●
●● ●●
●● ● ●●
●
●● ●
●
● ●● ●●
●●●
●● ●●●●
●●●● ●
● ●●
●● ●
●●
●
●
●● ●●
● ●
●● ● ●● ●●●
● ●
●●
● ●● ●● ● ●
●● ●●
●● ● ●●
●
●● ●
●
● ●●
●●●
●●●● ●●● ●
●●●● ●● ●
●●● ●
●●
●
●● ●●●●●
●●●●●●
●● ●● ●
● ●
●●●●●
● ●● ●●
●● ● ●●
●
●●●
●
● ● ●●● ●● ●
●●●
●
●●● ●●●
●●●● ●●● ●
●●●● ●● ●
●●● ●
●●
●
●● ●●●●●
●●●●●●
●● ●● ●
● ●
●●●●●
● ●● ●●
●● ● ●●
●
●●●
●
● ● ●●● ●● ●
●●●
●
●●●
●●
●●● ●●●●●●● ●
● ●
●●● ●● ●●●● ●
●● ●● ●●●●● ● ●●●●●●●●●●● ●●●
● ● ●● ●●●●●
● ● ●
● ●●
●●● ●●●●●●● ●
● ●
●●● ●● ●●●● ● ●● ●●●●● ● ●●●● ●
●●●●●● ●●●
● ● ●● ●●●●●
● ● ●
●
●●●●●
●● ●
●●
●
●●
●
●●
●
●
●●●●●
●
● ●●●●●● ●
●●●
●● ● ●● ●
● ●●
●●●●● ●
●●●
●
●
● ●
●
●●
●●
● ●
●● ●
●●●
●●
●●
● ●● ●
●●●●●
●●●●
●
● ●●●●●
●● ●
●●
●
●●
●
●●
●
●
●●●●●
●
● ●●●●●●●
● ●
●●●
●● ● ●● ●
● ●●
●●●●● ●
●●●
●
●
● ●
●
●●
●●
● ●
●● ●
●●●
●●
●●
● ●● ●
●●●●●
●●●●
●
●
●
●
●
●●
●
●
●
●
●
●●●●
●
●●
●● ●
●
●
●●
● ●●
●●
● ●
●●
●
●
●
●
●
●
●
●
● ●●● ●
●●
● ●●●●
●
●
●
●
●
●
●
●●
● ●●
●●●●●●●●
●
●●●●● ●●●●● ●
●
●● ●
●●
●●
●●
●
●
●●●
●●
●●
● ●●●●●
●● ●●●
●●
●●
●
●
●
●●●
● ●
●
●
●●
●
●
●
●
●
●●●●
●
●●
●● ●
●
●
●●
● ●●
●●
● ●
●●
●
●
●
●
●
●
●
●
● ●●● ●
●●
● ●●●●
●
●
●
●
●
●
●
●●
● ●●
●●●●●●●●
●
●●●●● ●●●●● ●
●
●● ●
●●
●●
●●
●
●
●●●
●●
●●
● ●●●●●
●● ●●●
●●
●●
●
●
●
●●●
●
●●●●
●
●
●
●●●
● ●●●
●●●●
● ●
●
●●●
● ●
●●●
●
●
●●
●
●
●●
●●
●●
●●
●●
● ●●●
●●
●● ●●
●●●
●
● ●●
●●
●●●●●
●
●●
●●●
● ●●
●●● ●
●●
●
●
●
●●●
●
●
●
●
●
●●●
●
●
●●●
●●●
●●●
●●
● ●●
●●● ●
●
●
●●
●
●
● ●●●●
●
●
●
●●●
● ●●●
●●●●
● ●
●
●●●
● ●
●●●
●
●
●●
●
●
●●
●●
●●
●●
●●
● ●●●
●●
●● ●●
●●●
●
● ●●
●●
●●●●●
●
●●
●●●
● ●●
●●● ●
●●
●
●
●
●●●
●
●
●
●
●
●●●
●
●
●●●
●●●
●●●
●●
● ●●
●●● ●
●
●
●●
●
●
●
●●●● ● ●●● ● ●
● ●● ●
● ●● ●
● ● ● ●● ● ● ● ●● ● ● ●●●● ● ●●● ● ●
● ●● ●
● ●● ●
● ● ● ●● ● ● ● ●● ● ●
−1.0
−1.0
●
● ●
●●
●●●●
●● ●●● ●●
● ●
●●
●● ●●●
●
●●●● ●●● ●
●
●● ● ●
●
●●
●●● ●
●● ● ●
●●●● ●●●●●●
●
●
●●● ● ●●●
●●
● ●
●●●
●●●●
●●
●
● ●●●●
● ●
● ●
●●
●●●●
●● ●●● ●●
● ●
●●
●● ●●●
●
●●●● ●●● ●
●
●● ● ●
●
●●
●●● ●
●● ● ●
●●●● ●●●●●●
●
●
●●● ● ●●●
●●
● ●
●●●
●●●●
●●
●
● ●●●●
●
●
●● ● ● ● ● ● ● ●●● ● ● ●
●● ● ● ● ● ● ● ●●● ● ●
●●
●
●● ●
●●●●●
●
●●
●●●●● ●●● ●
●
● ●●●●●
●●●●●●●●
●●●●●●●●●
● ●● ●●● ●●●● ●●● ●●●
●● ●●●●
●●●● ●●●●● ●
●●
●
●
●
● ●
●
● ●●
●
●● ●
●●●●●
●
●●
●●●●● ●●● ●
●
● ●●●●●
●●●●●●●●
●●●●●●●●●
● ●● ●●● ●●●● ●●● ●●●
●● ●●●●
●●●● ●●●●● ●
●●
●
●
●
● ●
●
●
−1.0 −0.5 0.0 0.5 1.0 −1.0 −0.5 0.0 0.5 1.0
42 / 102
Synthetic data
1.0
●●
●●
●●
● ●●●● ●●●●●
●●●
●●●
●●●●
●●
●● ● ●●● ●
●●
●●●
● ●● ● ●
●●● ● ● ●
● ●●●
● ●
● ●● ●●● ●● ●● ●
●●●● ●●
●●
●●
● ●●●● ●●●●●
●●●
●●●
●●●●
●●
●● ● ●●● ●
●●
●●●
● ●● ● ●
●●● ● ● ●
● ●●●
● ●
● ●● ●●● ●● ●● ●
●●●●
● ● ● ● ●●●●
●●● ●
●●
●● ● ●● ● ● ●● ●● ●●●●● ● ●
● ● ● ● ● ● ●●●●
●●● ●
●●
●● ● ●● ● ● ●● ●● ●●●●● ● ●
● ●
●●
● ●
●●●
●
●
●
●
●●
●
●
●●
● ●●
●●●
●●
●
●
●●●
● ●●
●
●
●
●
●
●
●
●●
●●●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●●●●
●●●
●
●● ●
●●
●
●
●
●●●
● ●●●●
●
●●●●
●
●
●●●●
●
●●
●●●●●●
●●
●
● ●●
●●●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●●●
● ●●
● ●
●●●
●
●
●
●
●●
●
●
●●
● ●●
●●●
●●
●
●
●●●
● ●●
●
●
●
●
●
●
●
●●
●●●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●●●●
●●●
●
●● ●
●●
●
●
●
●●●
● ●●●●
●
●●●●
●
●
●●●●
●
●●
●●●●●●
●●
●
● ●●
●●●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●●●
●
●
● ● ●● ●
●● ●● ●
●●●
● ●● ● ●
● ●●●●● ●
●●●
●● ●●● ● ● ●●●●
●●● ● ●●● ● ●● ●●●●
●
●●
●●
●
●●●●● ●
●●
●
●●
●●
● ●
● ● ●● ●
●● ●● ●
●●●
● ●● ● ●
● ●●●●● ●
●●●
●● ●●● ● ● ●●●●
●●● ● ●●● ● ●● ●●●●
●
●●
●●
●
●●●●● ●
●●
●
●●
●●
●
●●●●●●● ●●●●●
●● ●●●● ●● ●●
● ●
● ●
●●
● ●● ● ● ●●
●●● ● ● ●● ● ●
● ●
●● ●
● ●
●
● ● ● ●●
●●● ●
●●● ●●●●●●● ●●●●●
●● ●●●● ●● ●●
● ●
● ●
●●
● ●● ● ● ●●
●●● ● ●●● ● ●
● ●
●● ●
● ●
●
● ● ● ●●
●●● ●
●●●
●●●●● ●●● ●●● ●
●●
●
●●● ●
●●●
●●●●● ●
●●●●●● ●●
●
●
● ●
●●● ●
● ●●
●●●●●●● ●●
● ●
●●●●●●●●
●●
●●●●●● ●●
●●●
● ●● ●
● ●●●●● ●●● ●●● ●
●●
●
●●● ●
●●●
●●●●● ●
●●●●●● ●●
●
●
● ●
●●● ●
● ●●
●●● ●
●●● ●●
● ●
●●●●●●●●
●●
●●●●●● ●●
●●●
● ●● ●
●
●●● ●●● ●●● ●●●
● ●
●●
●
●
●
● ●●
●●●
●●●●
●●
●●
●●
●
●
●●
●
●●●●
●
●
● ●
●●●
● ●
● ● ●●●● ●
● ●●
●●●
●
●● ●●
●● ●
●●● ● ●
●
● ●
● ●●● ●●● ●●● ●●● ●●●
● ●
●●
●
●
●
● ●●
●●●
●●●●
●●
●●
●●
●
●
●●
●
●●●●
●
●
● ●
●●●
● ●
● ● ●●● ●●
● ●● ●
●● ●●
●● ●
●●● ● ●
●
● ●
● ●●●
●
●●●
●
●●●
●●
●
●●
●●
●
●● ●●
●
●● ●●
●●●
●●
●
●●●
●● ●●
●
●
●
●
●●●
● ●●●
● ●
●●●
●●
●●
●●
● ●
●
●
●●
●●
●●●●
●
●●
●●
●●●
● ●
●●● ●●●● ●●●●●
● ●●●●●●●
● ●●
●●●
●
●
●●
●
●●●
●
● ●
●●●
●
●●●
●●
●
●●
●●
●
●● ●●
●● ●●
●●●
●●
●
●●●
●● ●●
●
●
●
●
●●●
● ●●●
● ●
●●●
●●
●●
●●
● ●
●
●
●●
●●
●●●●
●
●●
●●
●●●
● ●
●
●●●
●● ●●●● ●●●●●
● ●●●●●●●
● ●●
●●●
●
●
●●
●
●●●
●
●
●●
●● ●
●●●●● ●●●
● ●
●●
●●
●
●
● ● ●●●
●
●●●●●●●● ●●
●●
●● ●
●●
●
●●
●●●●● ●● ●
●
●●
●
●●
●●●
● ●●● ●
● ●
●●●
●
●●●● ●●
●●
●●●●●●●
●●●●
●
●
●●●●
●
●
● ● ●●
●● ●
●●●●● ●●
●●
● ●
●●
●●
●
●
● ● ●●●
●
●●●●●●●● ●●
●●
●● ●
●●
●
●●
●●●●● ●● ●
●
●●
●
●●●
●●
● ●●● ●
● ●
●●●
●
●●●● ●●
●●
●●●●●●●
●●●●
●
●
●●●●
●
●
● ●
0.5
0.5
● ● ● ● ●●
● ● ● ● ● ●●●
●● ●● ● ● ●● ●●●●
● ● ●●● ● ●●●
● ●
● ● ● ● ● ●●
● ● ●
● ● ● ●●●
●● ● ● ● ●● ●●●●
● ● ●●● ● ●●●
● ●
●
●●
● ●●
●
●● ●● ●●●●
● ●
●●●●
●
●●●●●●
●
●
●
●
●●● ●●●●●●
●● ●●
●●
●●
●
●●
● ● ●
●●
●●
●
●●●
●●
●● ● ●
●●● ●
● ●
● ● ●● ●●
●
● ●●
●
● ●
●●● ●●
● ●●
●
●● ●● ●●●●
● ●
●●●●
●
●●●●●●
●
●
●
●
●●● ●●●●●●
●● ●
●●
●●
●●
●
●●
● ● ●
●●
●●
●
●●●
●
●
●● ● ●
●●● ●
● ●
● ● ●● ●●
●
● ●●
●
● ●
●●●
●
●●●
● ●
●●
●
●●●
●●
●●●
● ●●
●● ●
●●
●●
●
●●
●
●
●
●●●●●●
●
●
●
●
●
●●
●●●●
●
●
●
●
●●●
●●●
●
●●
●●●
●●
●●●●●●●
●
●●●●
●
●●
●
●●●
● ●●
●
● ●
●
●●●●
●●
●
●
●
●
●●
●
●●●●●●
●●
●
●
●
●
●
●●
●
●
●●
●●●
● ●
● ●
●●●
● ●
●●
●
●●●
●●
●●●
● ●●
●● ●
●●
●●
●
●●
●
●
●
●●●●●●
●
●
●
●
●
●●
●●●●
●
●
●
●
●●●
●●●
●
●●
●●●
●●
●●●●●●●●
●
●●●
●
●●
●
●●●
● ●●
●
● ●
●
●●●●
●●
●
●
●
●
●●
●
●●●●●●
●●
●
●
●
●
●
●●
●
●
●●
●●●
● ●
●
●
● ●●
●●●
●
●● ●●● ● ●
●
●● ●
●●●
●●●●●●●
●●●●
●●
●
●●● ●
●
●
●
●
●● ●
●
●
●
●
●● ● ●●
● ●
●
●●
●●
●●
●●
●●● ●
● ●
●●●●●
●●●
●● ●●
●
●
●●● ●●
●
● ●
● ● ●
●●● ●
● ●●
●●●
●
●● ●●● ● ●
●
●● ●
●●●
●●●●●●●
●●●●
●●
●
●●● ●
●
●
●
●
●● ●
●
●
●
●
●● ● ●●
● ●
●
●●
●●
●●●
●
●●● ●
● ●
●●●●●
●●●
●● ●●
●
●
●●● ●●
●
● ●
● ● ●
●●●
●●● ● ●●●
● ●
●● ●●●
● ●●●●●●●● ●
● ● ●●● ● ●●● ●●●● ● ●●●
●●●● ● ●● ●●● ● ●●●
● ●● ●●
● ●●●●●●●● ●
● ● ●●● ● ●●● ●●●● ● ●●●
●●●● ● ●●
●●
●●
●
● ●●
●●
●●●●●●●
● ●
●●●●●
●
● ● ●
●
●●
●
●
●
●●
●●
● ●●●
●●●
●●●●●●●
●
●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●●
●
●
●
●
●
●
●
●●
●
●●●
●●●●
●
●●●●
●●
●
●
●● ●●●
●
●
●●
●
●
●
●●
● ●
●
●
●●
●
●
●●
●
●
●
●
●●
●●
● ●
●
●
●
●
●●
●
●
●
●●
●●
●
●
●
●●
●●
●
● ●●
●●
●●●●●●●
● ●
●●●●●
●
● ● ●
●
●●
●
●
●
●
●
●
●●
● ●●●
●
●●●
●●●●●●●
●
●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●●
●
●
●
●
●
●
●
●●
●
● ●
●●
●
●●
●
●●●●
●●
●
●
●● ●●●
●
●
●●
●
●
●
●●
● ●
●
●
●●
●
●
●●
●
●
●
●
●●
●●
● ●
●
●
●
●
●●
●
●
●
●●
●●
●
●
●
●
● ●
●●
●
●
●●● ●
● ●●●● ●
●
●
●●●●
●
● ●
●●●
●●
●● ●
● ●●
● ●●
●● ●●●● ● ●
●● ●
●
●●
●●● ●●
●●
● ●
●●● ● ●
●●●●●●● ●●
●●●●
●●
● ●
●● ●
● ●
●●
●
●
●●● ●
● ●●●● ●
●
●
●●●●
●
● ●
●●●
●●
●● ●
● ●●
● ●●
●● ●●●● ● ●
●● ●
●●
●●
●● ●●
●●
● ●
●●● ● ●
●●●●●●● ●●
●●●●
●●
● ●
●●
●●● ●●● ●● ● ●●●● ●●● ●
●●●●● ●●●●● ●●● ● ● ● ●
●●●●●●
● ●● ● ●●●● ●●● ●● ●● ● ●●●● ●●● ●
●●●●● ●●●●● ●●● ● ● ● ●
●●●●●● ●● ● ●●●●
●●
●●●●●
●● ●
● ● ● ●●
●
●●
●●●●●
●
●●
●
●●●●●● ●●●
●●●
●● ●●● ● ●●●
●●● ●
● ●●
●●●●
●●●●●
●●●●●
●● ●●●●●●●● ●
●
●●
● ●●
●●●●●
●● ●
●
● ● ● ●●
●
●●
●●●●●
●
●●
●
●●●●●● ●●●
●●●
●● ●●● ● ●●●
● ●
● ●
● ●●
●●●●●
●●●●
●●●●●
●
●● ●●●●●●●● ●
●
●●
●
●
● ●●
●
●●●
● ● ●● ●●
●●●
●●● ●
●●
●
●●
●●● ●
● ●
●
●●●
● ●●●
●
●● ●● ●●● ●
●●●
● ●
●●●●
●●
●
●
● ●●●●●
●●
●● ●●●
●●●●●●● ●
●
●
● ●● ● ● ●
● ●●
●
●●●
● ● ●● ●●
●●●
●●● ●
●●
●
●●
●●● ●
● ●
●
●●●
● ●●●
●
●● ●● ●●● ●
●●●
● ●
●●●●
●●
●
●
● ●●●●●
●●
●● ●●●
●●●●●●● ●
●
●
● ●● ● ●
● ● ●● ● ● ● ● ● ● ● ●●● ● ● ●● ● ● ● ● ● ● ● ●●●
0.0
●●
0.0
● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ●●
●
●
●
●
●
●
●●
●
●●
●●●
●● ● ●
●
●
●●●
●
●
●●
●
●
●
●●●
●
●●●
●●●●
●
●●
●
● ●
●●●
●
●
●
●
●● ●
●●
●●
● ●
●●
●● ●●●●
●
●●
●●
● ●●●
● ●
● ●●●
●●
●●●
●
●●
●●●
●
●
●●●
● ●
●
●●● ●
●
●●
●
●
●
●●
●●
●
●●●●
●
●●
●
●
●
●
●
●
●●
●
●●
●●●
●● ● ●
●
●
●●●
●
●
●●
●
●
●
●●●
●
●●●
●●●●
●
●●
●
● ●
●●●
●
●
●
●
●● ●
●●
●●
● ●
●●
●● ●●●●
●
●●
●●
● ●●●
● ●
● ●●●
●●
●●●
●
●●
●●●
●
●
●●●
● ●
●
●●● ●
●
●●
●
●
●
●●
●●
●
●●●●
●
●●
●
●
●
●●
●●●
●
●
●
●●
● ●●●
●
●
●
●●●
●●●●
●
●
●
●●
●
● ● ●●●●
●
●
●●●
●
●●●●●
●●●
●●●
●
●●
●●●
●
●
●●
●●
●
●
●
●
●
●●●●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
● ●
●
●
●
●
●
●
●●
●●●
● ●
●
●
●●●
●●●
●
●
●●
●
●
●
●●
●
●●
●●
●
●●●● ●
●
●
●●
●●●
●
●
●
●●
● ●●●
●
●
●
●●●
●●●●
●
●
●
●●
●
● ● ●●●●
●
●
●●●
●
●●●●●
●●●
●●●
●
●●
●●●
●
●
●●
●●
●
●
●
●
●
●●●●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
● ●
●
●
●
●
●
●
●●
●●●
● ●
●
●
●●●
●●●
●
●
●●
●
●
●
●●
●
●●
●●
●
●●●●
● ●● ●
●● ●●●●● ●
● ●
●
●●●● ●●
●●●●● ●●● ●
●
●●
●
● ● ●●● ●●
●● ●●●● ●
●●●● ●●●● ●
●●● ●
●●● ● ● ●●●●●●●
●● ● ● ●● ●
●● ●●●●● ●
● ●
●
●●●● ●●
●●●●● ●●● ●
●
●●
●
● ● ●●● ●●
●● ●●●● ●
●●●● ●●●● ●
●●● ●
●●● ● ● ●●●●●●●
●● ●
●●●
● ● ●
●● ● ●●
●●
●● ●●
● ●●● ● ●
● ● ● ●●● ●●●●●
●● ●
● ●
●●●● ●●
● ●
● ●●●●
●
● ●
●●●●● ●●
● ● ●● ● ●●●
● ● ●
●● ● ●●
●●
●● ●●
● ●●● ● ●
● ● ● ●●● ●●●●●
●● ●●●
●●●● ●●
● ●
● ●●●●
●
● ●
●●●●● ●●
● ● ●● ●
● ●
●●●● ●●
● ●
●● ● ●● ●
●
●● ●
●
●●●●
● ●
● ●
●●●●
●●●
●● ●●
●●
●●●
●●●●
●●● ● ●
●●●● ●
●●●
●●● ●
●●●● ●
●●
●●●
●●●●●●●● ●
●● ●● ●
●●●●●
● ● ● ●
●●●● ●●
● ●
●● ● ●● ●
●
●● ●
●
●●●●
● ●
● ●
●●●●
●●●
●● ●●
●●
●●●
●●●●
●●● ● ●
●●●● ●
●●●
●●● ●
●●●● ●
●●
●●●
●●●●●●●● ●
●● ●● ●
●●●●●
● ●
●
●● ● ●● ●
●
●●● ●●● ● ● ●●●●
● ●● ●● ●●
●● ●
●● ●●●●
● ● ●●
● ● ●● ● ●
●● ● ●● ●
●
●●● ●●● ● ● ●●●●
● ●● ●● ●●
●● ●
●● ●●●●
● ● ●●
● ● ●● ●
●
●●
●●
●
●
●
●●
● ●
●
●●
●
●
● ●●
●●
●●●●
●●
●●
●
●●
●● ●●●
●
●●
●
●● ●
●●
● ●●
●
●
●●●
● ●
●●●● ●● ●
●●
●●●●●● ●
●
● ●
●● ●
●●
●●
●●● ●
●●●●
●
●
●●
●
●●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●●●●
●●
●
●
●
●
●●
●● ●
●●
●●
●
●
●
●●
● ●
●
●●
●
●
● ●●
●●
●●●●
●●
●●
●
●●
●● ●●●
●
●●
●
●● ●
●●
● ●●
●
●
●●●
● ●
●●●● ●● ●
●●
●●●●●● ●
●
● ●
●● ●
●●
●●
●●● ●
●●●●
●
●
●●
●
●●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●●●●
●●
●
●
●
●
●●
●●
●●●
●
●
●●
●
●●
● ●●●● ● ●●●●
●
●●
●
●●●●
●●● ●
●●●
● ●● ●●●●
●●
● ●●
●●
●
●
●
●●●●
● ●●●●
●●●●●
●●
●●
●●
●●
●● ●●
● ●●●●
●●●
●●●●● ● ●
●●
● ●●●●
●
●
●
●●
●
● ●●●
●
●
●●
●
●●
● ●●●● ● ●●●●
●
●●
●
●●●●
●●● ●
●●●
● ●● ●●●●
●●
● ●●
●●
●
●
●
●●●●
● ●●●●
●●●●●●
●●
●●
● ●●
●● ●●
● ●●●●
●●●
●●●●● ● ●
●●
● ●●●●
●
●
●
●●
●
●
−0.5
● ●● ●●
−0.5
●● ●● ●● ●● ●●●●●●●
● ●● ●
●●●●●●
● ● ● ●●● ●● ●
● ●● ●●●
● ●
●
●● ● ●●
● ●
● ● ●●
●●●●●● ●
●
●●● ●● ●
●● ●● ●● ●●●●●●●
● ●● ●
●●●●●●
● ● ● ●●● ●● ●
● ●● ●●●
● ●
●
●● ● ●●
● ●
● ● ●●
●●●●●● ●
●
●●●
●
●●
●
●●
●
●
●●
●
●●
●
●
●
●
●
● ●●
● ●
●
●
●
●
●●
●● ●
●
●
●
●
●
●
● ●
●
●●
●
●●
●●●●●
●●
●
●
●
●
●
●●
●●●
●●●
●●●●
●●
●
●
●
●
●
●
●
● ●●
●
●●●
●
●
●
●●
●●●
●●●
●●●
●●●●
●
●●
●●●●
●
●
●
●
●●
●
●
● ●●
●
●●
● ●
●
●●●
●
●
●●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●●
●
●●
●
● ●
●
●
●●
●
●●
●
●
●●
●
●●
●
●
●
●
●
● ●●
● ●
●
●
●
●
●●
●● ●
●
●
●
●
●
●
● ●
●
●●
●
●●
●●●●●
●●
●
●
●
●
●
●●
●●●
●●●
●●●●
●●
●
●
●
●
●
●
●
● ●●
●
●●●
●
●
●
●●
●●●
●●●
●●●
●●●●
●
●●
●●●●
●
●
●
●
●●
●
●
● ●●
●
●●
● ●
●
●●●
●
●
●●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●●
●
●●
●
● ●
●
● ●●
●
●●● ●● ● ●
●●●●
●●● ●
●●●●
● ●●
● ● ●●
● ●●●●● ● ● ● ●
●●●
●●
● ●●
●
●● ●
● ●●
●● ●●
●
●● ●●● ●●
●●● ●
●●
●
●●●●●
●● ● ●
●● ● ●●
●
●●● ●● ● ●
●●●●
●●● ●
●●●●
● ●●
● ● ●●
● ●●●●● ● ● ● ●
●●●
●●
● ●●
●
●● ●
● ●●
●● ●●
●
●● ●●● ●●
●●● ●●
●
●●●
● ●●
●● ● ●
●●
●●
●●●
●● ●●●●
●●●● ●
● ●●
●● ●
●●
●
●
●● ●●
● ●
●● ● ●● ●●●
● ●
●●
● ●● ●● ● ●
●● ●●
●● ● ●●
●
●● ●
●
● ●● ●●
●●●
●● ●●●●
●●●● ●
● ●●
●● ●
●●
●
●
●● ●●
● ●
●● ● ●● ●●●
● ●
●●
● ●● ●● ● ●
●● ●●
●● ● ●●
●
●● ●
●
● ●●
●●●
●●●● ●●● ●
●●●● ●● ●
●●● ●
●●
●
●● ●●●●●
●●●●●●
●● ●● ●
● ●
●●●●●
● ●● ●●
●● ● ●●
●
●●●
●
● ● ●●● ●● ●
●●●
●
●●● ●●●
●●●● ●●● ●
●●●● ●● ●
●●● ●
●●
●
●● ●●●●●
●●●●●●
●● ●● ●
● ●
●●●●●
● ●● ●●
●● ● ●●
●
●●●
●
● ● ●●● ●● ●
●●●
●
●●●
●●
●●● ●●●●●●● ●
● ●
●●● ●● ●●●● ●
●● ●● ●●●●● ● ●●●●●●●●●●● ●●●
● ● ●● ●●●●●
● ● ●
● ●●
●●● ●●●●●●● ●
● ●
●●● ●● ●●●● ● ●● ●●●●● ● ●●●● ●
●●●●●● ●●●
● ● ●● ●●●●●
● ● ●
●
●●●●●
●● ●
●●
●
●●
●
●●
●
●
●●●●●
●
● ●●●●●● ●
●●●
●● ● ●● ●
● ●●
●●●●● ●
●●●
●
●
● ●
●
●●
●●
● ●
●● ●
●●●
●●
●●
● ●● ●
●●●●●
●●●●
●
● ●●●●●
●● ●
●●
●
●●
●
●●
●
●
●●●●●
●
● ●●●●●●●
● ●
●●●
●● ● ●● ●
● ●●
●●●●● ●
●●●
●
●
● ●
●
●●
●●
● ●
●● ●
●●●
●●
●●
● ●● ●
●●●●●
●●●●
●
●
●
●
●
●●
●
●
●
●
●
●●●●
●
●●
●● ●
●
●
●●
● ●●
●●
● ●
●●
●
●
●
●
●
●
●
●
● ●●● ●
●●
● ●●●●
●
●
●
●
●
●
●
●●
● ●●
●●●●●●●●
●
●●●●● ●●●●● ●
●
●● ●
●●
●●
●●
●
●
●●●
●●
●●
● ●●●●●
●● ●●●
●●
●●
●
●
●
●●●
● ●
●
●
●●
●
●
●
●
●
●●●●
●
●●
●● ●
●
●
●●
● ●●
●●
● ●
●●
●
●
●
●
●
●
●
●
● ●●● ●
●●
● ●●●●
●
●
●
●
●
●
●
●●
● ●●
●●●●●●●●
●
●●●●● ●●●●● ●
●
●● ●
●●
●●
●●
●
●
●●●
●●
●●
● ●●●●●
●● ●●●
●●
●●
●
●
●
●●●
●
●●●●
●
●
●
●●●
● ●●●
●●●●
● ●
●
●●●
● ●
●●●
●
●
●●
●
●
●●
●●
●●
●●
●●
● ●●●
●●
●● ●●
●●●
●
● ●●
●●
●●●●●
●
●●
●●●
● ●●
●●● ●
●●
●
●
●
●●●
●
●
●
●
●
●●●
●
●
●●●
●●●
●●●
●●
● ●●
●●● ●
●
●
●●
●
●
● ●●●●
●
●
●
●●●
● ●●●
●●●●
● ●
●
●●●
● ●
●●●
●
●
●●
●
●
●●
●●
●●
●●
●●
● ●●●
●●
●● ●●
●●●
●
● ●●
●●
●●●●●
●
●●
●●●
● ●●
●●● ●
●●
●
●
●
●●●
●
●
●
●
●
●●●
●
●
●●●
●●●
●●●
●●
● ●●
●●● ●
●
●
●●
●
●
●
●●●● ● ●●● ● ●
● ●● ●
● ●● ●
● ● ● ●● ● ● ● ●● ● ● ●●●● ● ●●● ● ●
● ●● ●
● ●● ●
● ● ● ●● ● ● ● ●● ● ●
−1.0
−1.0
●
● ●
●●
●●●●
●● ●●● ●●
● ●
●●
●● ●●●
●
●●●● ●●● ●
●
●● ● ●
●
●●
●●● ●
●● ● ●
●●●● ●●●●●●
●
●
●●● ● ●●●
●●
● ●
●●●
●●●●
●●
●
● ●●●●
● ●
● ●
●●
●●●●
●● ●●● ●●
● ●
●●
●● ●●●
●
●●●● ●●● ●
●
●● ● ●
●
●●
●●● ●
●● ● ●
●●●● ●●●●●●
●
●
●●● ● ●●●
●●
● ●
●●●
●●●●
●●
●
● ●●●●
●
●
●● ● ● ● ● ● ● ●●● ● ● ●
●● ● ● ● ● ● ● ●●● ● ●
●●
●
●● ●
●●●●●
●
●●
●●●●● ●●● ●
●
● ●●●●●
●●●●●●●●
●●●●●●●●●
● ●● ●●● ●●●● ●●● ●●●
●● ●●●●
●●●● ●●●●● ●
●●
●
●
●
● ●
●
● ●●
●
●● ●
●●●●●
●
●●
●●●●● ●●● ●
●
● ●●●●●
●●●●●●●●
●●●●●●●●●
● ●● ●●● ●●●● ●●● ●●●
●● ●●●●
●●●● ●●●●● ●
●●
●
●
●
● ●
●
●
−1.0 −0.5 0.0 0.5 1.0 −1.0 −0.5 0.0 0.5 1.0
43 / 102
Results for two performance measures
1 Pm
• Hamming loss: `H (y, h) = m i=1 Jyi 6= hi K ,
Conditional independence
classifier Hamming loss subset 0/1 loss
BR LR 0.4232 0.6723
LP LR 0.4232 0.6725
Conditional dependence
classifier Hamming loss subset 0/1 loss
BR LR 0.3470 0.5499
LP LR 0.3610 0.5146
44 / 102
Linear + XOR synthetic data
1.0
● ●
●● ● ●● ● ●
●● ●● ● ● ●●
●
●● ● ●● ●
●
● ●●
● ● ● ●
● ●
● ●● ● ●●●● ●
● ● ●● ●● ●●● ●●
0.5
● ●●●●
● ● ● ●●●●●●
● ●● ● ● ●
● ●● ●
● ● ● ●
●
● ● ●
● ● ● ●●
● ●
● ●● ● ●●
0.0
●● ● ●
●● ● ● ● ● ●●
●
● ● ●
● ● ●
●●●● ●● ●
● ● ●
●●● ●
● ● ● ●●● ●
−0.5
● ● ● ●●
● ●● ● ● ● ●
● ● ●●●
●●●
● ●●● ●●● ● ●
● ● ●●
● ●● ● ●
● ● ●● ●
● ● ● ● ● ●●
−1.0
●● ● ● ●●●
Figure : Problem with two targets: shapes (4 vs. ◦) and colors ( vs. ).
45 / 102
Linear + XOR synthetic data
Bayes Optimal 0 0
46 / 102
Linear + XOR synthetic data
46 / 102
Linear + XOR synthetic data
1.0
● ●
●● ●● ● ●
cannot handle the label color ( ●● ●●
●
● ● ●●
● ●
● ● ●
vs. ) – the XOR problem. ● ●
● ●●
●
●●
●
●
● ●
● ●● ● ●●●●●● ●
● ●●
● ●● ●●
0.5
● ●●●●●
• LP LR uses four linear classifiers ●
● ●●
● ● ●●●●●●
● ●
● ●
●● ●
● ● ●
to solve 4-class problem (M, N, ●
●
●
● ●
● ● ● ● ●●
●
◦, •): extends the hypothesis ● ●● ● ●●
0.0
●● ● ●
●● ● ● ● ● ●●
space. ●
●
● ●
● ●
●
●●
●●●●● ● ● ●
●●● ●
• BR MLRules uses two non-linear ● ● ● ●●● ●
−0.5
● ● ● ●●
● ●● ● ● ● ●
● ● ● ●●●
classifiers (based on decision ● ● ●●● ● ●
● ●●●
● ● ●●
● ●● ●
rules): XOR problem is not a ●
● ●
●
●
●●
● ●
●
● ●●
−1.0
●● ● ● ●●●
problem.
−1.0 −0.5 0.0 0.5 1.0
• There is no noise in the data.
• Easy to perform unfair
comparison.
47 / 102
Multi-target prediction - probabilistic view
P (Y , X) .
48 / 102
Multi-target prediction - probabilistic view
P (Y , X) .
48 / 102
Multi-target prediction - probabilistic view
P (Y , X) .
48 / 102
Multi-target prediction - probabilistic view
P (Y , X) .
48 / 102
Multi-target prediction - probabilistic view
P (Y , X) .
48 / 102
Multi-target prediction - loss minimization view
49 / 102
Multi-target prediction - loss minimization view
m
1 X
• Hamming loss: `H (y, h) = Jyi 6= hi K
m
i=1
Pm
2 yi hi
i=1 P
• F-measure-based loss: `F (y, h) = 1 − Pm m
i=1 yi + i=1 hi
X 1
• Rank loss: `rnk (y, h) = w(y) Jhi < hj K + Jhi = hj K
yi >yj
2
• ...
50 / 102
Loss minimization view - main issues
I Generalization bounds.
I Computational complexity.
51 / 102
Relations between losses
yi 6= hi .
• In case of deterministic data (no-noise): the optimal prediction
should have the same form for all loss functions and the risk for this
prediction should be 0.
• In case of non-deterministic data (noise): the optimal prediction
and its risk can be different for different losses.
52 / 102
Relations between losses
12
K. Dembczyński, W. Waegeman, W. Cheng, and E. Hüllermeier. On loss minimization and
label dependence in multi-label classification. Machine Learning, 88:5–45, 2012
53 / 102
Risk minimizers
• The risk minimizer for the Hamming loss is the marginal mode:
54 / 102
Consistency of risk minimizers and risk bounds
1
L (h, P | x) ≤ LH (h, P | x) ≤ L0/1 (h, P | x)
m 0/1
55 / 102
Regret analysis
• The previous results may suggest that one of the loss functions can
be used as a proxy (surrogate) for the other:
I For some situations both risk minimizers coincide.
I One can provide mutual bounds for both loss functions.
56 / 102
Regret analysis
• The previous results may suggest that one of the loss functions can
be used as a proxy (surrogate) for the other:
I For some situations both risk minimizers coincide.
I One can provide mutual bounds for both loss functions.
• However, the regret analysis of the worst case shows that
minimization of the subset 0/1 loss may result in a large error
for the Hamming loss and vice versa.
56 / 102
Regret analysis
58 / 102
Regret analysis
m−2
RegH (h∗0/1 , P | x) = LH (h∗0/1 , P | x) − LH (h∗H , P | x) <
m+2
• Moreover, this bound is tight.
• Example:
y P (y)
0 00 0 0.170
0 11 1 0.166
Marginal mode: 1111
1 01 1 0.166 Joint mode: 0000
1 10 1 0.166
1 11 0 0.166
1 11 1 0.166
59 / 102
Relations between losses
• Summary:
I The risk minimizers of Hamming and subset 0/1 loss are different:
60 / 102
Relations between losses
prediction.
• Subset 0/1 loss:
I Very restrictive.
diseases of a patient.
61 / 102
BR vs. LP
62 / 102
BR vs. LP
on label (in)dependence.
I If this would not be true, then we could not optimally solve binary
classification problems!!!
I For other losses, one should probably take additional assumptions:
• For subset 0/1 loss: label independence, high probability of the joint
mode (> 0.5), . . .
I Learning and inference is linear in m (however, faster algorithms exist).
63 / 102
BR vs. LP
space.
I Unfortunately, learning and inference is basically exponential in m
64 / 102
Algorithmic approaches for multivariate losses
• The loss functions, like Hamming loss or subset 0/1 loss, often
referred to as task losses, are usually neither convex nor
differentiable.
• Therefore learning is a hard optimization problem.
• Two approaches try to make this task easier
I Reduction.
65 / 102
Reduction
{(x, y)}ni=1
x Inference ŷ
66 / 102
Structured loss minimization
{(x, y)}ni=1
f (x, y)
x Inference ŷ
67 / 102
Statistical consistency
69 / 102
Conditional random fields
1
P (y | x) = exp(f (x, y))
Z(x)
14
John D. Lafferty, Andrew McCallum, and Fernando C. N. Pereira. Conditional random fields:
Probabilistic models for segmenting and labeling sequence data. In ICML, pages 282–289, 2001
70 / 102
Conditional random fields
71 / 102
Conditional random fields
72 / 102
Conditional random fields
73 / 102
Conditional random fields
exp( m
P
exp(f (x, y)) f (x, yi ))
P (y | x) = P =P Pmi
i=1
y∈Y exp(f (x, y)) y∈Y exp( i=1 fi (x, yi ))
Qm Qm
i=1 exp(fi (x, yi )) exp(fi (x, yi ))
= P Qm = Qm i=1 P
y∈Y i=1 exp(fi (x, yi )) i=1 yi exp(fi (x, yi ))
73 / 102
Conditional random fields
exp( m
P
exp(f (x, y)) f (x, yi ))
P (y | x) = P =P Pmi
i=1
y∈Y exp(f (x, y)) y∈Y exp( i=1 fi (x, yi ))
Qm Qm
i=1 exp(fi (x, yi )) exp(fi (x, yi ))
= P Qm = Qm i=1 P
y∈Y i=1 exp(fi (x, yi )) i=1 yi exp(fi (x, yi ))
m
Y
= P (yi | x)
i=1
73 / 102
Conditional random fields
exp( m
P
exp(f (x, y)) f (x, yi ))
P (y | x) = P =P Pmi
i=1
y∈Y exp(f (x, y)) y∈Y exp( i=1 fi (x, yi ))
Qm Qm
i=1 exp(fi (x, yi )) exp(fi (x, yi ))
= P Qm = Qm i=1 P
y∈Y i=1 exp(fi (x, yi )) i=1 yi exp(fi (x, yi ))
m
Y
= P (yi | x)
i=1
74 / 102
Conditional random fields
74 / 102
Conditional random fields
P P
exp( i fi (yi ) + yk ,yl fk,l (yk , yl ))
P (y) = P P P
y∈Y exp( i fi (yi ) + yk ,yl fk,l (yk , yl ))
74 / 102
Structured loss minimization
15
Y. Tsochantaridis, T. Joachims, T. Hofmann, and Y. Altun. Large margin methods for struc-
tured and interdependent output variables. JMLR, 6:1453–1484, 2005
75 / 102
Structured loss minimization
15
Y. Tsochantaridis, T. Joachims, T. Hofmann, and Y. Altun. Large margin methods for struc-
tured and interdependent output variables. JMLR, 6:1453–1484, 2005
75 / 102
Structured support vector machines
76 / 102
Structured support vector machines
77 / 102
Structured support vector machines
16
B. Hariharan, L. Zelnik-Manor, S.V.N. Vishwanathan, and M. Varma. Large scale max-margin
multi-label classification with priors. In ICML. Omnipress, 2010
78 / 102
Structured support vector machines
16
B. Hariharan, L. Zelnik-Manor, S.V.N. Vishwanathan, and M. Varma. Large scale max-margin
multi-label classification with priors. In ICML. Omnipress, 2010
78 / 102
Structured support vector machines
16
B. Hariharan, L. Zelnik-Manor, S.V.N. Vishwanathan, and M. Varma. Large scale max-margin
multi-label classification with priors. In ICML. Omnipress, 2010
78 / 102
Structured support vector machines
17
W. Gao and Z.-H. Zhou. On the consistency of multi-label learning. Artificial Intelligence,
199-200:22–44, 2013
18
A. Tewari and P.L. Bartlett. On the consistency of multiclass classification methods. JMLR,
8:1007–1025, 2007
D. McAllester. Generalization Bounds and Consistency for Structured Labeling in Predicting
Structured Data. MIT Press, 2007
79 / 102
Structured support vector machines
19
Thomas Finley and Thorsten Joachims. Training structural SVMs when exact inference is
intractable. In ICML. Omnipress, 2008
20
K. Dembczyński, W. Waegeman, W. Cheng, and E. Hüllermeier. An analysis of chaining in
multi-label classification. In ECAI, 2012
80 / 102
SSVMs vs. CRFs
• Their additional advantage is that one can easily sample from the
estimated distribution.
22
J. Read, B. Pfahringer, G. Holmes, and E. Frank. Classifier chains for multi-label classification.
Machine Learning Journal, 85:333–359, 2011
K. Dembczyński, W. Cheng, and E. Hüllermeier. Bayes optimal multilabel classification via
probabilistic classifier chains. In ICML, pages 279–286. Omnipress, 2010
82 / 102
Probabilistic classifier chains
83 / 102
Probabilistic classifier chains
y2 = 0 y2 = 1 y2 = 0 y2 = 1
P (y2=0 | y1=0, x)=0.0 P (y2=1 | y1=0, x)=1.0 P (y2=0 | y1=1, x)=0.4 P (y2=1 | y1=1, x)=0.6
P (y=(0, 0) | x)=0 P (y=(0, 1) | x)=0.4 P (y=(1, 0) | x)=0.24 P (y=(1, 1) | x)=0.36
• For subset 0/1 loss one needs to find h(x) = arg maxy∈Y P (y | x).
• Greedy and approximate search techniques with guarantees exist.23
23
K. Dembczyński, W. Waegeman, W. Cheng, and E. Hüllermeier. An analysis of chaining in
multi-label classification. In ECAI, 2012
A. Kumar, S. Vembu, A.K. Menon, and C. Elkan. Beam search algorithms for multilabel
learning. In Machine Learning, 2013
84 / 102
Probabilistic classifier chains
y2 = 0 y2 = 1 y2 = 0 y2 = 1
P (y2=0 | y1=0, x)=0.0 P (y2=1 | y1=0, x)=1.0 P (y2=0 | y1=1, x)=0.4 P (y2=1 | y1=1, x)=0.6
P (y=(0, 0) | x)=0 P (y=(0, 1) | x)=0.4 P (y=(1, 0) | x)=0.24 P (y=(1, 1) | x)=0.36
84 / 102
Probabilistic classifier chains
Table : PCC vs. SSVMs on Hamming loss and PCC vs. BR on subset 0/1 loss.
85 / 102
Multilabel ranking
Multi-label classification
politics 0
economy 0
business 0
sport 1
tennis 1
soccer 0
show-business 0
celebrities 1
..
.
England 1
USA 1
Poland 1
Lithuania 0
86 / 102
Multilabel ranking
Multilabel ranking
tennis
≺
sport
≺
England
≺
Poland
≺
USA
≺
..
.
≺
politics
86 / 102
Multilabel ranking
• Ranking loss:
X 1
`rnk (y, h) = w(y) Jhi (x) < hj (x)K + Jhi (x) = hj (x)K ,
2
(i,j) : yi >yj
X1 X2 Y1 Y2 ... Ym
x 4.0 2.5 1 0 0
h2 > h1 > ... > hm
87 / 102
Multilabel ranking
• Ranking loss:
X 1
`rnk (y, h) = w(y) Jhi (x) < hj (x)K + Jhi (x) = hj (x)K ,
2
(i,j) : yi >yj
87 / 102
Pairwise surrogate losses
where φ is
I an exponential function (BoosTexter)24 : φ(f ) = e−f ,
I logistic function (LLLR)25 : φ(f ) = log(1 + e−f ) ,
I or hinge function (RankSVM)26 : φ(f ) = max(0, 1 − f ) .
24
R. E. Schapire and Y. Singer. BoosTexter: A Boosting-based System for Text Categorization.
Machine Learning, 39(2/3):135–168, 2000
25
O. Dekel, Ch. Manning, and Y. Singer. Log-linear models for label ranking. In NIPS. MIT
Press, 2004
26
A. Elisseeff and J. Weston. A kernel method for multi-labelled classification. In NIPS, pages
681–687, 2001
88 / 102
Multilabel ranking
• This approach is, however, inconsistent for the most commonly used
convex surrogates.27
• The consistent classifier can be, however, obtained by using
univariate loss functions28 . . .
27
J. Duchi, L. Mackey, and M. Jordan. On the consistency of ranking algorithms. In ICML, pages
327–334, 2010
W. Gao and Z.-H. Zhou. On the consistency of multi-label learning. Artificial Intelligence,
199-200:22–44, 2013
28
K. Dembczynski, W. Kotlowski, and E. Hüllermeier. Consistent multilabel ranking through
univariate losses. In ICML, 2012
89 / 102
Reduction to weighted binary relevance
90 / 102
Reduction to weighted binary relevance
1 ∆1 1 ∆1i
h∗i (x) = log i0 = log ,
c ∆i c W − ∆1i
91 / 102
Reduction to weighted binary relevance
29
R. E. Schapire and Y. Singer. BoosTexter: A Boosting-based System for Text Categorization.
Machine Learning, 39(2/3):135–168, 2000
30
K. Dembczynski, W. Kotlowski, and E. Hüllermeier. Consistent multilabel ranking through
univariate losses. In ICML, 2012
92 / 102
Weighted binary relevance
● ●
WBR−LR ●
WBR−LR
●
0.174
LLLR LLLR
0.186
Bayes risk Bayes risk
rank loss
rank loss
●
●
0.172
●
●
0.184
●
●
●
●
●
●
●
0.170
● ●
●
● ● ●
0.182
● ●
● ●
●
250 500 1000 2000 4000 8000 16000 250 500 1000 2000 4000 8000 16000
# of learning examples # of learning examples
Figure : WBR LR vs. LLLR. Left: independent data. Right: dependent data.
93 / 102
Benchmark data
94 / 102
Maximization of the F-measure
95 / 102
Maximization of the F-measure
96 / 102
SSVMs for Fβ -based loss
31
J. Petterson and T. S. Caetano. Reverse multi-label learning. In NIPS, pages 1912–1920, 2010
J. Petterson and T. S. Caetano. Submodular multi-label learning. In NIPS, pages 1512–1520,
2011
97 / 102
Plug-in rule approach
98 / 102
Plug-in rule approach
• Approximation needed?
32
N. Ye, K. Chai, W. Lee, and H. Chieu. Optimizing F-measures: a tale of two approaches. In
ICML, 2012
33
K. Dembczyński, W. Waegeman, W. Cheng, and E. Hüllermeier. An exact algorithm for F-
measure maximization. In NIPS, volume 25, 2011
34
K. Dembczynski, A. Jachnik, W. Kotlowski, W. Waegeman, and E. Hüllermeier. Optimizing
the F-measure in multi-label classification: Plug-in rule approach versus structured loss mini-
mization. In ICML, 2013
99 / 102
Plug-in rule approach
32
N. Ye, K. Chai, W. Lee, and H. Chieu. Optimizing F-measures: a tale of two approaches. In
ICML, 2012
33
K. Dembczyński, W. Waegeman, W. Cheng, and E. Hüllermeier. An exact algorithm for F-
measure maximization. In NIPS, volume 25, 2011
34
K. Dembczynski, A. Jachnik, W. Kotlowski, W. Waegeman, and E. Hüllermeier. Optimizing
the F-measure in multi-label classification: Plug-in rule approach versus structured loss mini-
mization. In ICML, 2013
99 / 102
Plug-in rule approach
• EFP is consistent.34
32
N. Ye, K. Chai, W. Lee, and H. Chieu. Optimizing F-measures: a tale of two approaches. In
ICML, 2012
33
K. Dembczyński, W. Waegeman, W. Cheng, and E. Hüllermeier. An exact algorithm for F-
measure maximization. In NIPS, volume 25, 2011
34
K. Dembczynski, A. Jachnik, W. Kotlowski, W. Waegeman, and E. Hüllermeier. Optimizing
the F-measure in multi-label classification: Plug-in rule approach versus structured loss mini-
mization. In ICML, 2013
99 / 102
Maximization of the F-measure
F1-‐measure [%]
F1-‐measure
[%]
70
55
60
60
55,73
55
50
43,63
50
45
50
45
40
40
40
35
35
30
30
30
EFP
LFP
RML
SML
BR
EFP
LFP
RML
SML
BR
EFP
LFP
RML
SML
BR
F1-‐measure [%]
F1-‐measure
[%]
54,61
55,49
51,21
67,9
70,19
55
49,35
50,02
70
50
50
60
45
45
50
40
40
40
35
35
30
30
30
EFP
LFP
RML
SML
BR
EFP
LFP
RML
SML
BR
EFP
LFP
RML
SML
BR
100 / 102
Challenges
• Main challenges:
I Learning and inference algorithms for any task losses and output
structures.
I Consistency of the algorithms.
I Large-scale datasets: number of instances, features, and labels.
101 / 102
Conclusions
• Take-away message:
I Two main challenges: loss minimization and target dependence.
I Two views: the individual target and the joint target view.
I The individual target view: joint target regularization
I The joint target view: structured loss minimization and reduction.
Many thanks to Eyke and Willem for collaboration on this tutorial and Arek for a
help in preparing the slides.
This project is partially supported by the Foundation of Polish Science under the
Homing Plus programme, co-financed by the European Regional Development Fund.
102 / 102