Correlated bandits or: How to minimize mean-squared error online 1 - - PowerPoint PPT Presentation

correlated bandits or how to minimize mean squared error
SMART_READER_LITE
LIVE PREVIEW

Correlated bandits or: How to minimize mean-squared error online 1 - - PowerPoint PPT Presentation

Correlated bandits or: How to minimize mean-squared error online 1 LinkedIn Corp. 2 Indian Institute of Technology Madras. A portion of this work was done while the authors were at University of Maryland, College Park 1 V. Praneeth Boda 1 and


slide-1
SLIDE 1

Correlated bandits or: How to minimize mean-squared error online

  • V. Praneeth Boda1 and Prashanth L. A.2

1LinkedIn Corp. 2Indian Institute of Technology Madras. A portion of this work was done while the authors were at University of Maryland, College Park

1

slide-2
SLIDE 2

Centrality among Bandits

▶ Placement of sensors used

for measuring temperature in a region.

▶ Best set of towers which

approximate the whole network. Aim: Find arm with highest information about other arms

2

slide-3
SLIDE 3

Minimum Mean Squared Error Estimation

▶ Jointly Gaussian arms XM = (X1, . . . , XK), with zero mean and

covariance matrix Σ ≜ E[XT

MXM].

MMSE Ei ≜ min

g

E [( XM − g(Xi) )T( XM − g(Xi) )] =

K

j=1

E [( Xj − E[Xj|Xi] )2] = ∑

j̸=i

σ2

j (1 − ρ2 ij)

The optimal g∗(Xi) = E[XM|Xi] = [E[X1|Xi] . . . E[XK|Xi]]T , with E[Xj|Xi] = E[XjXi] E[X2

i ] Xi = ρijσj

σi Xi.

3

slide-4
SLIDE 4

Correlated Bandits

Input: set of arm-pairs S ≜ {(i, j) | i, j = 1, . . . , K, i < j}, number of rounds n For t = 1, 2, . . . , n do Select a pair (it, jt) ∈ S Observe a sample from the bivariate distribution corresponding to the arms it, jt endfor Output an arm ˆ An based on sample-based MSE-value estimates necessary for estimating correlation structure so that P (An ̸= i∗) is minimized. Here i∗ = arg min

i∈M

Ei. 4

slide-5
SLIDE 5

MSE Estimation and Concentration

Based on samples of the Gaussian arms: MSE of arm i ˆ Ei ≜ ∑

j̸=i

ˆ σ2

j

( 1 − ˆ ρ2

ij

) . Sample variance Sample correlation MSE Concentration: Assume σ2

i ≤ 1, i = 1, . . . , K. Then, for any

i = 1, . . . , K, and for any ϵ ∈ [0, 2K], we have P (

  • ˆ

Ei − Ei

  • > ϵ

) ≤ 14K exp ( −nl2ϵ2 cK5 ) ,

where c is a universal constant, and 0 < l = min

i

σ2

i .

5

slide-6
SLIDE 6

SR algorithm: Illustration of arm-pair elimination

Maintain active arms and arm-pairs

(1,2) (1,3) (1,4) (1,5) (2,3) (2,4) (2,5) (3,4) (3,5) (4,5)

Active arm-pairs after arms 4, 5 are eliminated

(1,2) (1,3) (1,4) (1,5) (2,3) (2,4) (2,5) (3,4) (3,5) (4,5)

Active arm-pairs after arms 3, 4, 5 are eliminated

6

slide-7
SLIDE 7

Successive Rejects: An algorithm to find the best arm

Initial- ization A1 = all arm pairs, B1 = {1, . . . , K},

nk = ⌈ n − (K

2

) C(K) (K + 1 − k) ⌉ , C(K) ≈ K log K.

Phase 1 Pull each pair in A1, n1 times; Set Bk+1 = Bk\ arm with lowest MSE Phase 2 Play each arm pair in A2, n2 − n1 times; Eliminate . . . . . . . . . Phase K − 1 Play the remaining two arm pairs nK−1 − nK−2 times ▶ One arm pair played n1

times, . . ., another two played n2 times

▶ k arms played nk+1 times ▶

K−1

k=1

(k − 1)nk + (K − 1)nK−1 < n, ▶ nk increases with k ▶ Adaptive exploration:

better than uniform (= play each arm-pair n/ (K 2 ) times)

7

slide-8
SLIDE 8
  • Thanks. Questions?

8