Dynamic Classifier Selection Based on Imprecise Probabilities
Meizhu Li Ghent University
Co-work: Jasper De Bock, Gert de Cooman
Dynamic Classifier Selection Based on Imprecise Probabilities - - PowerPoint PPT Presentation
Dynamic Classifier Selection Based on Imprecise Probabilities Meizhu Li Ghent University Co-work: Jasper De Bock, Gert de Cooman Outline Dynamic Strategy of Experiment classifier selection results selection Motivation Normally, one
Meizhu Li Ghent University
Co-work: Jasper De Bock, Gert de Cooman
Dynamic classifier selection Strategy of selection Experiment results
3
4
5
How to select an appropriate classifier for each instance?
Let us denote C as the class variable, taking values c in finite set .
For each , denotes a set of probability mass function P(c).
π¬(c) c β π P(c) = n(c) + 1 N + |π|
Fig.1 Example of a Naive Bayes Classifier
6
7
Let us denote C as the class variable, taking values c in finite set .
For each , denotes a set of probability mass function P(c).
π¬(c) c β π P(c) = n(c) + 1 + st(c) N + |π| + s
, for all .
c β π
where s is a fixed hyperparameter that determines the degree of imprecision, t is any probability mass functions on c.
P(c) = n(c) + 1 N + |π|
Fig.1 Example of a Naive Bayes Classifier
8
Threshold: the largest value of s that does not induce a change of prediction result. Let us denote C as the class variable, taking values c in finite set .
For each , denotes a set of probability mass function P(c).
π¬(c) c β π P(c) = n(c) + 1 + st(c) N + |π| + s
, for all .
c β π
where s is a fixed hyperparameter that determines the degree of imprecision, t is any probability mass functions on c.
P(c) = n(c) + 1 N + |π|
Fig.1 Example of a Naive Bayes Classifier
9
by global sensitivity analysis for MAP inference in graphical models.
[1] De Bock, J., De Campos, C.P., Antonucci, A.: Global sensitivity analysis for MAP inference in graphical models. Advances in Neural Information Processing Systems 27 (Proceedings of NIPS 2014), 2690β2698. (2014)
k training instances whose thresholds are most similar with the testing instance Thresholds of training instances in C1 and C2 Threshold of testing instance in C1 and C2 Local accuracy in C1 (Acc1) and C2 (Acc2) Use C1 for prediction Use C2 for prediction Acc1 Acc2 Acc1<Acc2
β₯
70% Training Set 30% Testing Set
1 0
k training instances whose thresholds are most similar with the testing instance Thresholds of training instances in C1 and C2 Threshold of testing instance in C1 and C2 Local accuracy in C1 (Acc1) and C2 (Acc2) Use C1 for prediction Use C2 for prediction Acc1 Acc2 Acc1<Acc2
β₯
70% Training Set 30% Testing Set
11
h
t
e c i d e k ? h
t
n d i n s t a n c e s w i t h s i m i l a r t h r e s h
d s ?
1 2
Classifier 1 Classifier 2 Testing instance
a1 b1
Training instance 1 Training instance 2 Training instance n
x1 x2 xn y1 y2 yn (a1, b1) (x1, y1) (x2, y2) (xn, yn)
Thresholds
1 3
Threshold in Classifier 1 Threshold in Classifier 2 Chebyshev Distance Euclidean Distance Training Instance Testing Instance
with fifty training points, and for k = 10 and two different distance measures
1 4
Classifier 1 (C1) and Classifier 2 (C2)
[1] UCI Homepage, http://mlr.cs.umass.edu/ml/index.html.
Continuous variables were discretized by their median
1 5
k 2 4 6 8 10 12 14 16 18 20 Accuracy 0.745 0.75 0.755 0.76 0.765 0.77 0.775 0.78 0.785 0.79 Classifier 1 Classifier 2 Euclidean Distance Chebyshev Distance
classifiers: the two original ones (which do not depend on k) and two combined classifiers (one for each of the considered distance measures)
1 6
the individual ones on which they are based.
seems to have very little effect.
1 7
dynamic classifier selection methods that outperform the individual classifiers they select from.
Classifier.
M E I Z H U L I G H E N T U N I V E R S I T Y meizhu.Li@ugent.be