Online Stochastic Matching with Unequal Probabilities Aranyak Mehta - - PowerPoint PPT Presentation

online stochastic matching with unequal probabilities
SMART_READER_LITE
LIVE PREVIEW

Online Stochastic Matching with Unequal Probabilities Aranyak Mehta - - PowerPoint PPT Presentation

Online Stochastic Matching with Unequal Probabilities Aranyak Mehta Bo Waggoner Harvard Morteza Zadimoghaddam SODA 2015 1 Outline Problem and motivation Prior work, our main result Key idea: Adaptivity Ideas behind


slide-1
SLIDE 1

Online Stochastic Matching with Unequal Probabilities

Aranyak Mehta Bo Waggoner Morteza Zadimoghaddam

SODA 2015

1

Harvard

slide-2
SLIDE 2

Outline

  • Problem and motivation
  • Prior work, our main result
  • Key idea: Adaptivity
  • Ideas behind algorithm/analysis

2

slide-3
SLIDE 3

Motivation: Search ads

3

advertisers Time search queries

slide-4
SLIDE 4

Motivation: Search ads

4

advertisers Time search queries

Simplified problem:

  • display one ad per query
  • have estimate of click

probabilities

  • advertisers pay $1 if

click, $0 if no click

  • advertisers have budget

for one click per day How to assign ads?

slide-5
SLIDE 5

Online Stochastic Matching

5

fixed,

  • ffline vertices

Time

  • nline

arrivals

[Mehta and Panigrahi, 2012]

slide-6
SLIDE 6

Online Stochastic Matching

6

fixed,

  • ffline vertices

Time

  • nline

arrivals p11 p31 p41

[Mehta and Panigrahi, 2012]

slide-7
SLIDE 7

Online Stochastic Matching

7

fixed,

  • ffline vertices

Time

  • nline

arrivals p11 p31 p41

[Mehta and Panigrahi, 2012]

Pr[ searcher clicks if we show this ad ]

slide-8
SLIDE 8

Online Stochastic Matching

8

Time

Alg

p31

Assign to vertex 3!

fixed,

  • ffline vertices
  • nline

arrivals

[Mehta and Panigrahi, 2012]

slide-9
SLIDE 9

Online Stochastic Matching

9

Time

Alg

p31 fixed,

  • ffline vertices
  • nline

arrivals

[Mehta and Panigrahi, 2012]

With prob p31: match succeeds With prob 1 - p31: match fails

slide-10
SLIDE 10

Online Stochastic Matching

10

Time

Alg

fixed,

  • ffline vertices
  • nline

arrivals

[Mehta and Panigrahi, 2012]

match succeeded cannot be matched again

slide-11
SLIDE 11

Online Stochastic Matching

11

Time

Alg

fixed,

  • ffline vertices
  • nline

arrivals

[Mehta and Panigrahi, 2012]

match failed may be matched again later disappears (cannot re-try)

slide-12
SLIDE 12

Alg’s performance = # successes

Measuring algorithm performance

12

Alg

fixed,

  • ffline vertices
  • nline

arrivals

slide-13
SLIDE 13

Alg’s performance = E[ # successes ]

13 13

Alg

Measuring algorithm performance

fixed,

  • ffline vertices
  • nline

arrivals

slide-14
SLIDE 14

14 14

Alg’s performance = E[ # successes ] Opt’s performance = size of max weighted assignment, budget 1

Opt Alg

Measuring algorithm performance

fixed,

  • ffline vertices
  • nline

arrivals

slide-15
SLIDE 15

15 15

Alg’s performance = E[ size of matching ] Opt’s performance = size of max weighted assignment, budget 1

Opt Alg

Measuring algorithm performance

fixed,

  • ffline vertices
  • nline

arrivals

Competitive ratio = min Alg Opt

  • ver all input instances.

(Note: Opt is a bit funky … not achievable even with foreknowledge of instance.)

slide-16
SLIDE 16

Prior Work

  • Online Matching with Stochastic Rewards

Mehta, Panigrahi, FOCS 2012. ○ Greedy = 0.5. Opt ○ For case where all p are equal and vanishing: Alg ≥ 0.567. Opt Open: anything better than Greedy for unequal p

16

slide-17
SLIDE 17

This work

17

Opt Alg

≥ 0.534

For unequal, vanishing edge probabilities:

slide-18
SLIDE 18

This work

18

Opt Alg

≥ 0.534

For unequal, vanishing edge probabilities: So what? algorithmic ideas to beat Greedy

slide-19
SLIDE 19

Outline

  • Problem and motivation
  • Prior work, our main result
  • Key idea: Adaptivity
  • Ideas behind algorithm/analysis

19

slide-20
SLIDE 20

Adaptive: sees whether or not assignment succeeds

20

fixed,

  • ffline vertices
  • nline

arrivals

slide-21
SLIDE 21

Our Approach

  • 1. Start with an optimal non-adaptive alg that is

straightforward to analyze

  • 2. Add a small amount of adaptivity

(second choices)

  • 3. Analysis remains tractable by limiting

amount of adaptivity

21

slide-22
SLIDE 22

An optimal non-adaptive algorithm

22

  • MP-2012: nonadaptive algs have upper bound of 0.5
  • How to achieve 0.5? (Previously unknown.) Seems

nonobvious.

slide-23
SLIDE 23

Maximize marginal expected gain

23

  • nline

arrivals

  • ffline

vertices 0.3 0.4 0.2

Assign first arrival to vertex with largest pi1

slide-24
SLIDE 24

Maximize marginal expected gain

24

  • nline

arrivals

  • ffline

vertices

Assign next arrival to max Pr[ i available ] pi2

0.1 0.2 0.3

slide-25
SLIDE 25

Maximize marginal expected gain

25

  • nline

arrivals

  • ffline

vertices

Assign next arrival to max Pr[ i available ] pi2

0.1 0.2 0.3

= (1 - 0.4) * 0.3 = 0.18 = (1) * 0.2 = 0.2

slide-26
SLIDE 26

NonAdaptive

26

Theorem: NonAdaptive has a competitive ratio of 0.5 for the general online stochastic matching problem.

Does not require vanishing probabilities.

slide-27
SLIDE 27

Why do we like NonAdaptive?

  • On a given instance, an arrival has the same

“first choice” every time (regardless of previous realizations)

  • Algorithm tracks/uses competitive ratio

(probabilities of success)

27

slide-28
SLIDE 28

Add Adaptivity (but not too much)

Proposed SemiAdaptive: Assign next arrival to max Pr[ i available ] pij unless already taken, in which case assign to second-highest.

28

slide-29
SLIDE 29

Why do we like SemiAdaptive?

  • On a given instance, an arrival has the same

first and second choices every time (regardless of previous realizations)

  • Algorithm tracks/uses competitive ratio

(probabilities of success) These allow us to analyze SemiAdaptive -- almost...

29

slide-30
SLIDE 30

(Analysis?) Roadblock

  • Want: when first-choice is not available, get measurable

benefit by assigning to second choice → giving improvement over NonAdaptive’s 0.5

30

slide-31
SLIDE 31

(Analysis?) Roadblock

  • Want: when first-choice is not available, get measurable

benefit by assigning to second choice → giving improvement over NonAdaptive’s 0.5

  • Problem: correlation between availability of first and

second choice. Perhaps when first choice is not available, most likely second choice is not available either. → cannot guarantee improvement over NonAdaptive

31

slide-32
SLIDE 32

(Analysis?) Roadblock

  • Want: when first-choice is not available, get measurable

benefit by assigning to second choice → giving improvement over NonAdaptive’s 0.5

  • Problem: correlation between availability of first and

second choice. Perhaps when first choice is not available, most likely second choice is not available either. → cannot guarantee improvement over NonAdaptive

  • Fix: introduce independence / even less adaptivity.

(no time to say more! sorry!)

32

slide-33
SLIDE 33

RECAP

Online stochastic matching problem:

  • edges succeed probabilistically
  • maximize expected number of successes
  • input instance chosen adversarially

New here:

  • edge probabilities

may be unequal

33

p11 p31

slide-34
SLIDE 34

RECAP

Results:

  • optimal 0.5-competitive NonAdaptive
  • 0.534-competitive SemiAdaptive

(with tweak) for vanishing probabilities Key idea:

  • control adaptivity to

control analysis

34

p11 p31

slide-35
SLIDE 35

Future Work

Everything about Online Stochastic Matching:

  • Vanishing probabilities:

○ Equal: 0.567 … ? … 0.62 ○ Unequal: 0.534 … ? … 0.62

  • Large probabilities:

○ Equal: 0.53 … ? … 0.62 ○ Unequal: 0.5 … ? … 0.62

35

slide-36
SLIDE 36

Future Work

Everything about Online Stochastic Matching:

  • Vanishing probabilities:

○ Equal: 0.567 … ? … 0.62 ○ Unequal: 0.534 … ? … 0.62

  • Large probabilities:

○ Equal: 0.53 … ? … 0.62 ○ Unequal: 0.5 … ? … 0.62

Thanks!

36

slide-37
SLIDE 37

Additional slides

37

slide-38
SLIDE 38

Final Algorithm “SemiAdaptive”

38

Assign next arrival to max Pr[ i available ] pij unless already taken, in which case assign to second-highest. * “it would have already been taken by a previous first-choice”

(key point: even less adaptive, more independence)

*

slide-39
SLIDE 39

Ideas behind analysis

39

p12 p42 Pr[ available ] q2 p22 q1 q3 q4 q5

Either first choice is the same as Opt’s...

slide-40
SLIDE 40

Ideas behind analysis

40

...or both first and second choice would give at least as much “gain” as Opt’s. Either first choice is the same as Opt’s...

p42 Pr[ available ] q2 p22 q1 q3 q4 q5 p12

slide-41
SLIDE 41

Ideas behind analysis

41

...or both first and second choice would give at least as much “gain” as Opt’s. Either first choice is the same as Opt’s...

p42 Pr[ available ] q2 p22 q1 q3 q4 q5 p12

Very good because gains “compound”. Good because we get “second-choice gains”.

slide-42
SLIDE 42

Note: Can only get 1 - 1/e ≈ 0.632 even with knowledge of instance

42

  • nline

arrivals

42

Opt Alg 1/n 1/n 1/n 1/n 1/n 1/n

Weighted matching: 1 E[ # of matches ] = 1 - Pr[ all fail ] = 1 - (1 - 1/n)n → 1 - 1/e

slide-43
SLIDE 43

43 43

Alg’s performance = E[ size of matching ] Opt’s performance = size of max weighted assignment, budget 1

Opt Alg

Example of defining Opt

fixed,

  • ffline vertices
  • nline

arrivals

1/2 2/3 1/4 1/4 Opt gets 1 Opt gets 1/2