Weak Supervision, noisy labels, and error propagation Marat - - PowerPoint PPT Presentation

weak supervision
SMART_READER_LITE
LIVE PREVIEW

Weak Supervision, noisy labels, and error propagation Marat - - PowerPoint PPT Presentation

Weak Supervision, noisy labels, and error propagation Marat Freytsis hep-ai journal club December 11, 2018 based on Yu et al. [arXiv:1402.5902], Cohen, MF, Osdiek [arXiv:1706.09451] + bits of others Why Weak supervision? Fully supervised


slide-1
SLIDE 1

Weak Supervision,

noisy labels, and error propagation Marat Freytsis

hep-ai journal club — December 11, 2018

based on Yu et al. [arXiv:1402.5902], Cohen, MF, Osdiek [arXiv:1706.09451] + bits of others

slide-2
SLIDE 2

Why Weak supervision?

Fully supervised learning on real data often diffjcult/impossible

  • Individual labels are prohibitively expensive to assign
  • Personalized information legally protected (e.g., medical,

demographic data)

  • For quantum systems, unique labels may be unphysical

Several classes of learning tasks on partially labels well developed

  • semi-supervised: augmenting labeled with unlabeled data
  • multiple instance: presence of signal in bag is marked but not identifjed

One which nicely maps onto many scientifjc data measurements is Learning from Label Proportions

1/ 14

slide-3
SLIDE 3

Plan

  • Learning from Label Proportions
  • Viability and generalization error
  • Proportion uncertainties, stability, and error propagation

1/ 14

slide-4
SLIDE 4

Learning from Label Proportions

general setting

Domain of instance features denoted by X and (discrete) labels by

  • Y. Data consists of bags of events with features ˜

x = (x1, . . . xr) and labels ˜ y = (y1, . . . , yr), drawn iid from a distribution over (X × Y)r. Learner has no access to labels, but instead receives label proportions (˜ x, fi(˜ y)), with fi(˜ y) = r

n=1 Iyn=i/r. From a set of m

bags, the task is to fjnd a classifjer from individual events to labels.

For experimental measurements, fi(˜ y) can be naturally interpreted as, e.g., a rate/cross-section measurement/calculation even if individual events cannot be perfects separated by their features

2/ 14

slide-5
SLIDE 5

Is this even possible?

heuristic argument

Consider binary classifjcation (yi = {0, 1}). Discretize data into bins bm,j. If 2 bags are present, in each bin bA,j = fA,1b1,j + (1 − fA,1)b0,j bB,j = fB,1b1,j + (1 − fB,1)b0,j = ⇒ b0,j = fA,1bB,j−fB,1bA,j

fA,1−fB,1

b1,j = (1−fB,1)bA,j−(1−fA,1)bB,j

fA,1−fB,1

and the distributions can be inverted algebraically. Requirements:

  • Number of bags ≥ number of classes to be distinguished, with

label proportions unique for each bag.

  • The bags need to be drawing from the same underlying

distribution for each class, i.e., however the label proportions were made difgerent should be uncorrelated from the distribution over (X × Y)r.

3/ 14

slide-6
SLIDE 6

Classifjcation in practice

Don’t want to discretize, no guarantee events sample feature space densely enough it even makes sense. How to classify events? Modify loss function!

  • 1. direct attack:

ℓLLP = arg min

h∈H ℓ(h(xi)batch, f (˜

y)batch) typically need re-optimization of hyperparameters

  • 2. clever trick (classifjcation without labels):

ℓCWoLa = arg min

h∈H ℓ(h(xi), f (˜

y))

Metodiev et al. [arXiv:1708.02949]

with your fully-supervised loss function of choice

4/ 14

slide-7
SLIDE 7

Classifjcation without labels

why does the second version work at all?

Theorem

Given mixed samples M1 and M2 defjned in terms of pure samples S and B with signal fractions f1 > f2, an optimal classifjer trained to distinguish M1 from M2 is also optimal for distinguishing S from B.

Proof.

The optimal classifjer to distinguish examples drawn from pM1 and pM2 is the likelihood ratio LM1/M2(x) = pM1(x)/pM2(x). Similarly, the optimal classifjer to distinguish examples drawn from pS and pB is the likelihood ratio LS/B(x) = pS(x)/pB(x). Where pB has support, we can relate these two likelihood ratios algebraically: LM1/M2 = pM1 pM2 = f1pS + (1 − f1)pB f2pS + (1 − f2)pB = f1LS/B + (1 − f1) f2LS/B + (1 − f2) , which is a monotonically increasing rescaling of the likelihood LS/B as long as f1 > f2, since ∂LS/BLM1/M2 = (f1 − f2)/(f2LS/B − f2 + 1)2 > 0. If f1 < f2, then one obtains the reversed

  • classifjer. Therefore, LS/B and LM1/M2 defjne the same classifjer.

Only makes sense for binary classifjcation! Still need to know label proportions to calibrate classifjer.

5/ 14

slide-8
SLIDE 8

Plan

  • Learning from Label Proportions
  • Viability and generalization error
  • Proportion uncertainties, stability, and error propagation

5/ 14

slide-9
SLIDE 9

When is all of this viable?

All of this should clearly work in at least some cases, but can we know when will fails? It turns out the classifjcation without labels results are more general than they seem. Under mild assumptions (more later) a classifjer which can accurately predict bag proportions can be guaranteed to achieve low error on event labels. More precisely, for φr(h) : X r → R, φr(h)(˜ x) = r

n=1 h(xi)/r, the

classifjer selected by arg min

h∈H

  • bags

ℓ(φr(h), f (˜ y)) will also solve the original task with high accuracy.

6/ 14

slide-10
SLIDE 10

Generalization errors for label proportions

For a given empirical bag label proportion error for loss function ℓ, errℓ(h), it is possible to prove a bound on the expected error over the full distribution X × Y, errℓ

G(h) = E(˜ x,˜ y)ℓ(φr(h), f (˜

y)). As a function of the VC dimension of the hypothesis class, with probability 1 − δ, errℓ

G(h) ≤ errℓ(h) + ǫ if the number of bags m is

m ≥ 64 ǫ2

  • 2VC(H) log 12r

ǫ + log 4 δ

  • .

The mild dependence on bag size r means that destabilizing the method by adding more data is not a large concern.

for this proof and following, see arXiv:1402.5902

7/ 14

slide-11
SLIDE 11

Event errors from proportion errors

With some mild assumptions, the above founds can be extended to individual events. If errℓ

G(h) ≤ ǫ with probability 1 − δ, and each bag is at least

(1 − η)-pure 1 − ρ of the time, then h(x) correctly classifjes a fraction (1 − τ)(1 − δ − ρ)(1 − 2η − ǫ) of N events with probability 1 − e− Nτ2

2

(1−δ−ρ)(1−2η−ǫ).

Unfortunately, these bounds are somewhat weak. Guaranteed high performance generically requires extremely pure samples.

8/ 14

slide-12
SLIDE 12

Class distribution independence

The preceding was so weak because no conditional independence

  • f the underlying distributions from the bags was assumed, i.e.,

the assumption that allowed us to invert the class distributions earlier. If all bags are drawn from mixtures of underlying class distributions with difgerent fractions, the probability of event error can be written as a generative model. For binary classifjcation, the probability of getting a classifjer with error ≤ ǫ is then bounded from below by u(ǫ, r).

The general answer becomes quite involved in this case, and I won’t attempt to reproduce it.

0.2 0.4 0.6 0.8 1 0.2 0.4 0.6 0.8 1

ǫ u(ǫ, r)

r = 10 r = 15 r = 20 r = 25 r = 30 r = 35 r = 40 r = 45 r = 50 r = 100

9/ 14

slide-13
SLIDE 13

Plan

  • Learning from Label Proportions
  • Viability and generalization error
  • Proportion uncertainties, stability, and error

propagation

9/ 14

slide-14
SLIDE 14

Label uncertainties

The supervised aspect comes from the provided label proportions. What if these are wrong? Return to the heuristic argument bA,j = fA,1b1,j + (1 − fA,1)b0,j bB,j = fB,1b1,j + (1 − fB,1)b0,j = ⇒ b0,j = fA,1bB,j−fB,1bA,j

fA,1−fB,1

b1,j = (1−fB,1)bA,j−(1−fA,1)bB,j

fA,1−fB,1

A Neyman–Pearson-optimal classifjer is z = b0/(b0 + b1). The dependence on the error from a shift/uncertainty in any label proportion can be worked out analytically.

10/ 14

slide-15
SLIDE 15

Label insensitivity

cartoon version ¯ z′ ¯ z

← − more signal more background − →

¯ z′

good

¯ z′

bad

¯ zcut ¯ z′

cut

all cuts ¯ z′ bad only

As long as the resulting distortion is monotonic, the classifjers are equivalent

11/ 14

slide-16
SLIDE 16

Label insensitivity

concrete example

For a binary classifjer and 2 bags with error fA,1 → fA,1 + δ, ¯ z′ = 1 − fB 1 − 2fB

1−fA−δ 1−fB

− r(x)

1−2fA−2δ 1−2fB

− r(x) = ¯ zi + δ 1−fB

1−2fB − ¯ zi 1−2fB + 2(¯

z2

i − ¯

zi)

fA−fB 1−2fB + 2δ( 1−fB 1−2fB − ¯

zi)

  • ,

where r(x) = bA(x)/bB(x) is the ratio of inferred bag distributions. The classifjer remains equivalent to the optimal one if δ fA − fB 3 − 2 min(fB, 1 − fB)

12/ 14

slide-17
SLIDE 17

A numerical study

impact of mismodelling

10-5 10-4 10-3 10-2 10-1 100

False positve rate

0.4 0.5 0.6 0.7 0.8 0.9 1.0

True positive rate

randomly swap 15% of each class

10-5 10-4 10-3 10-2 10-1 100

False positve rate

0.4 0.5 0.6 0.7 0.8 0.9 1.0

True positive rate

Fully supervised (original) Weakly supervised (original) Fully supervised (mis-modeled) Weakly supervised (mis-modeled)

swap the 10% (15%) most signal-like (background-like)

Using random mutli-gaussian mixture models

13/ 14

slide-18
SLIDE 18

Concluding thoughts

  • Can bounds on generalization errors be made stronger

without assuming distribution independence? (Or assuming something weaker)

  • Understand how optimality arguments change with fjnite

statistics/correlations?

  • Can we propagate input uncertainties through the network?

◮ Where would this be useful?

  • Thank you!

14/ 14