Structure of optimal strategies for remote estimation over - - PowerPoint PPT Presentation

structure of optimal strategies for remote estimation
SMART_READER_LITE
LIVE PREVIEW

Structure of optimal strategies for remote estimation over - - PowerPoint PPT Presentation

Structure of optimal strategies for remote estimation over Gilbert-Elliott channel with feedback Jhelum Chakravorty Joint work with Aditya Mahajan McGill University ISIT June 27, 2017 1 / 18 Motivation Sequential transmission of data Zero


slide-1
SLIDE 1

Structure of optimal strategies for remote estimation over Gilbert-Elliott channel with feedback

Jhelum Chakravorty Joint work with Aditya Mahajan

McGill University

ISIT June 27, 2017

1 / 18

slide-2
SLIDE 2

Motivation

Sequential transmission of data Zero delay in reconstruction

2 / 18

slide-3
SLIDE 3

Motivation

Applications? Smart grids

2 / 18

slide-4
SLIDE 4

Motivation

Applications? Environmental monitoring, sensor network

2 / 18

slide-5
SLIDE 5

Motivation

Applications? Internet of things

2 / 18

slide-6
SLIDE 6

Motivation

Applications? Smart grids Environmental monitoring, sensor network Internet of things Salient features Sensing is cheap Transmission is expensive Size of data-packet is not critical

2 / 18

slide-7
SLIDE 7

Motivation

We study the structure of optimal strategies for a fundamental trade-off between estimation accuracy and transmission cost!

2 / 18

slide-8
SLIDE 8

The model

3 / 18

slide-9
SLIDE 9

The remote-state estimation setup

Markov Process Transmitter Erasure Channel Receiver 𝑌𝑢 𝑉𝑢 𝑍𝑢 ˆ 𝑌𝑢

ACK/NACK

Source model Generic: Xt ∈ X, X: finite or Borel-measurable; Stylized: Xt+1 = aXt + Wt; Xt ∈ X, Wt i.i.d.

4 / 18

slide-10
SLIDE 10

The remote-state estimation setup

Markov Process Transmitter Erasure Channel Receiver 𝑌𝑢 𝑉𝑢 𝑍𝑢 ˆ 𝑌𝑢

ACK/NACK

Transmitter Ut = ft(X0:t, S0:t−1, Y0:t−1) ∈ {0, 1}

4 / 18

slide-11
SLIDE 11

The remote-state estimation setup

Markov Process Transmitter Erasure Channel Receiver 𝑌𝑢 𝑉𝑢 𝑍𝑢 ˆ 𝑌𝑢

ACK/NACK

Channel model St Markovian; St = 1: channel ON, St = 0: channel OFF State transition matrix Q. Yt =      Xt, if Ut = 1 and St = 1 E1, if Ut = 0 and St = 1 E0, if St = 0.

4 / 18

slide-12
SLIDE 12

The remote-state estimation setup

Markov Process Transmitter Erasure Channel Receiver 𝑌𝑢 𝑉𝑢 𝑍𝑢 ˆ 𝑌𝑢

ACK/NACK

Receiver ˆ Xt = gt(Y0:t) Per-step distortion: d(Xt − ˆ Xt). d(·): even and quasi-convex. Communication Transmission strategy f = {ft}∞

t=0

strategies Estimation strategy g = {gt}∞

t=0

4 / 18

slide-13
SLIDE 13

The infinite horizon optimization problem

Discounted setup: β ∈ (0, 1)

Dβ(f , g) := (1 − β)E(f ,g) ∞

  • t=0

βtd(Xt − ˆ Xt)

  • X0 = 0
  • Nβ(f , g) := (1 − β)E(f ,g) ∞
  • t=0

βtUt

  • X0 = 0
  • Long-term average setup: β = 1

D1(f , g) := lim sup

T→∞

1 T E(f ,g) T−1

  • t=0

d(Xt − ˆ Xt)

  • X0 = 0
  • N1(f , g) := lim sup

T→∞

1 T E(f ,g) T−1

  • t=0

Ut

  • X0 = 0
  • 5 / 18
slide-14
SLIDE 14

The infinite horizon optimization problem

Problem C ∗

β(λ) := inf (f ,g) Dβ(f , g) + λNβ(f , g), β ∈ (0, 1]

5 / 18

slide-15
SLIDE 15

The infinite horizon optimization problem

Problem C ∗

β(λ) := inf (f ,g) Dβ(f , g) + λNβ(f , g), β ∈ (0, 1]

Salient features Multiple decision makers — Transmitter and Estimator: decentralized control system Cooperative set-up — minimization of a common objective function Modeled as a Team problem; Team: Multiple decision makers to achieve a common goal

5 / 18

slide-16
SLIDE 16

Decentralized control systems

Pioneers: Theory of teams Economics: Marschak, 1955; Radner, 1962 Systems and control: Witsenhausen, 1971; Ho, Chu, 1972

6 / 18

slide-17
SLIDE 17

Decentralized control systems

Pioneers: Theory of teams Economics: Marschak, 1955; Radner, 1962 Systems and control: Witsenhausen, 1971; Ho, Chu, 1972 Remote-state estimation as Team problem No packet drop - Marshak, 1954; Kushner, 1964; Åstrom, Bernhardsson, 2002; Xu and Hespanha, 2004; Imer and Başar, 2005; Lipsa and Martins, 2011; Molin and Hirche, 2012; Nayyar, Başar, Teneketzis and Veeravalli, 2013; D. Shi, L. Shi and Chen, 2015 With packet drop - Ren, Wu, Johansson, G. Shi and L. Shi, 2016; Chen, Wang, D. Shi and L. Shi, 2017; With noise - Gao, Akyol and Başar, 2015–2017

6 / 18

slide-18
SLIDE 18

Structural results

7 / 18

slide-19
SLIDE 19

Structure of optimal strategies

Generic model: X is finite or Borel-measurable. Belief states based on common information π1

t (x) := Pf (Xt = x | S0:t−1 = s0:t−1, Y0:t−1 = y0:t−1),

π2

t (x) := Pf (Xt = x | S0:t = s0:t, Y0:t = y0:t).

Theorem 1: structure of optimal strategies Ut = f ∗

t (Xt, St−1, Π1 t ),

ˆ Xt = g∗

t (Π2 t ).

POMDP-like dynamic programming formulation.

8 / 18

slide-20
SLIDE 20

Structure of optimal strategies

Stylized model: Xt+1 = aXt + Wt; Wt: Unimodal and symmetric. Theorem 2: Optimal estimator Time homogeneous! ˆ Xt =

  • Yt,

if Yt ∈ {E0, E1}; a ˆ Xt−1, if Yt ∈ {E0, E1}.

8 / 18

slide-21
SLIDE 21

Structure of optimal strategies

Stylized model: Xt+1 = aXt + Wt; Wt: Unimodal and symmetric. Theorem 2: Optimal estimator Time homogeneous! ˆ Xt =

  • Yt,

if Yt ∈ {E0, E1}; a ˆ Xt−1, if Yt ∈ {E0, E1}. Theorem 2: Optimal transmitter Xt ∈ R; Ut is threshold based action: Ut =

  • 1,

if |Xt − a ˆ Xt−1| ≥ k(St−1) 0, if |Xt − a ˆ Xt−1| < k(St−1)

8 / 18

slide-22
SLIDE 22

Proof sketch

Theorem 1 Use notion of Irrelevant Information to show that (Xt, S0:t−1, Y0:t−1) is sufficient information at the transmitter Identify the common information (S0:t−1, Y0:t−1) at the transmitter and (S0:t, Y0:t) at the receiver Local information at the transmitter: Xt and at the receiver: ∅ Belief states: at the transmitter π1

t := P(Xt | S0:t−1, Y0:t−1),

at the receiver π2

t := P(Xt | S0:t, Y0:t)

Common information approach - Nayyar, Mahajan, Teneketzis TAC’13: show that (Xt, St−1, π1

t ) is sufficient

statistic at the transmitter and π2

t is sufficient statistic at the

receiver

9 / 18

slide-23
SLIDE 23

Proof sketch

Theorem 2 Change of variables: Et, E +

t , ˆ

Et Zt =

  • aZt−1,

if Yt ∈ {E0, E1} Yt, if Yt ∈ {E0, E1} Et := Xt − aZt−1, E +

t := Xt − Zt,

ˆ Et := ˆ Xt − Zt Step 1: Forward induction method utilizing majorization properties to show optimal ˆ Et = 0 — leads to the structure of

  • ptimal estimator

Step 2: Fix the optimal estimator. Show by constructing a threshold based prescription that such a transmission strategy is optimal

9 / 18

slide-24
SLIDE 24

Computation of optimal performances: autoregressive model

10 / 18

slide-25
SLIDE 25

Step 1: computation of the performance of a threshold based strategy

f (k)(Et, St−1) =

  • 1,

if St−1 = 0 & |Et| ≥ k(St−1) 0, if St−1 = 0 & |Et| < k(St−1). τ (k): the time a packet was last received successfully.

11 / 18

slide-26
SLIDE 26

Step 1: computation of the performance of a threshold based strategy

τ (k): the time a packet was last received successfully. Till first successful reception L(k)

β

:= E τ (k)−1

  • t=0

βtd(Et)

  • E0 = 0, S0 = 1
  • M(k)

β

:= E τ (k)−1

  • t=0

βt

  • E0 = 0, S0 = 1
  • K (k)

β

:= E τ (k)

  • t=0

βtUt

  • E0 = 0, S0 = 1
  • 11 / 18
slide-27
SLIDE 27

Step 1: computation of the performance of a threshold based strategy

Et is regenerative process Renewal relationships D(k)

β

:= Dβ(f (k), g∗) = L(k)

β

M(k)

β

N(k)

β

:= Nβ(f (k), g∗) = K (k)

β

M(k)

β

11 / 18

slide-28
SLIDE 28

Step 2: Optimality condition (JC & AM: TAC’17, NecSys ’16)

D(k)

β , N(k) β , C (k) β

  • differentiable in k.

Theorem If (k, λ) satisfies ∇kD(k)

β

+ λ∇kN(k)

β

= 0, then, (f (k), g∗) optimal for costly comm. with cost λ.

12 / 18

slide-29
SLIDE 29

Step 2: Optimality condition (JC & AM: TAC’17, NecSys ’16)

D(k)

β , N(k) β , C (k) β

  • differentiable in k.

Theorem If (k, λ) satisfies ∇kD(k)

β

+ λ∇kN(k)

β

= 0, then, (f (k), g∗) optimal for costly comm. with cost λ. C ∗

β(λ) := Cβ(f (k), g∗; λ) is continuous, increasing and concave in λ.

12 / 18

slide-30
SLIDE 30

Step 2: Computation of optimal thresholds

Numerically compute L(k)

β , M(k) β

and K (k)

β ; Renewal relationship

to compute C (k)

β .

Analytical formulae are difficult to obtain.

13 / 18

slide-31
SLIDE 31

Step 2: Computation of optimal thresholds

Numerically compute L(k)

β , M(k) β

and K (k)

β ; Renewal relationship

to compute C (k)

β .

Analytical formulae are difficult to obtain. Simulation based approach - JC, JS & AM ACC’17 Two DP based approaches - Monte Carlo (MC) and Temporal Difference (TD)

MC - High variance due to one sample path; low bias TD - Low variance due to bootstrapping; high bias

13 / 18

slide-32
SLIDE 32

Step 2: Computation of optimal thresholds

Numerically compute L(k)

β , M(k) β

and K (k)

β ; Renewal relationship

to compute C (k)

β .

Analytical formulae are difficult to obtain. Simulation based approach - JC, JS & AM ACC’17 Two DP based approaches - Monte Carlo (MC) and Temporal Difference (TD)

MC - High variance due to one sample path; low bias TD - Low variance due to bootstrapping; high bias

Exploit regenerative property of the underlying state (error) process Renewal Monte Carlo (RMC) - low variance (independent sample paths from renewal) and low bias (since MC)

13 / 18

slide-33
SLIDE 33

Step 2: Computation of optimal thresholds

Numerically compute L(k)

β , M(k) β

and K (k)

β ; Renewal relationship

to compute C (k)

β .

Analytical formulae are difficult to obtain. Key idea Renewal Monte Carlo

Pick a k, compute sample values L, M, K till first successful reception Sample average to compute L(k)

β , M(k) β , K (k) β .

Stochastic approximation techniques to compute optimal k.

13 / 18

slide-34
SLIDE 34

Step 2: Computation of optimal thresholds

Key steps of the algorithms Noisy policy evaluation - MC till successful reception: constitutes one episode; sample average over few episodes to find ˆ L, ˆ M, ˆ K and hence ˆ C. Policy improvement - Smoothed Functional ˆ ki+1 = ˆ ki − γi η 2˜ β ˆ C(ˆ ki + ˜ βη) − ˆ C(ˆ ki − ˜ βη)

  • k = [k(0), k(1)]⊺; η : 2 × 1 Gaussian perturbation vector, ˜

β : tuning parameter

13 / 18

slide-35
SLIDE 35

Smoothed Functional algo.- Katkovnik & Kulchitsky ’72

A Simultaneous Perturbation variant to estimate the gradient: ∇kC (k)

β

Interpretation - Cost function is convolved with a particular smooth kernel (e.g. Gaussian, Cauchy), effectively making the cost function more convex-esque Efficient scalability to higher dimensions

14 / 18

slide-36
SLIDE 36

Simulation results to find optimal thresholds

10,000 20,000 30,000 40,000 50,000 5 10 15 20

k∗ k∗

1

Iterations Threshold k∗

0; k∗ 1 vs iterations; λ: 100

Figure: k∗

0 , k∗ 1 plots for λ = 100: β = 0.9, q00 = 0.3, q10 = 0.1.

15 / 18

slide-37
SLIDE 37

Simulation results to find optimal thresholds

10,000 20,000 30,000 40,000 50,000 5 10 15 20

k∗ k∗

1

Iterations Threshold k∗

0; k∗ 1 vs iterations; λ: 500

Figure: k∗

0 , k∗ 1 plots for λ = 500: β = 0.9, q00 = 0.3, q10 = 0.1.

15 / 18

slide-38
SLIDE 38

Optimal performance from simulation

100 200 300 400 500 600 700 800 900 2 4 6 8 10 λ C∗

0.9(λ)

q00 = 0.3, q10 = 0.1

Figure: C ∗

0.9(λ) vs λ: q00 = 0.3, q10 = 0.1.

16 / 18

slide-39
SLIDE 39

Future work

Computation of the optimal constrained performance using stochastic approximation based method Extension of the results to vector valued source processes.

17 / 18

slide-40
SLIDE 40

Thank you

18 / 18