Anytime Capacity of Stabilization of a Linear System over Noisy - - PowerPoint PPT Presentation

anytime capacity of stabilization of a linear system over
SMART_READER_LITE
LIVE PREVIEW

Anytime Capacity of Stabilization of a Linear System over Noisy - - PowerPoint PPT Presentation

Anytime Capacity of Stabilization of a Linear System over Noisy Channel Graduate Seminar in Area I (6.454) October 26, 2011 1 / 38 Outline 1 Introduction 2 A Counter Example 3 Necessity of Anytime Capacity 4 Conclusions 2 / 38 Outline 1


slide-1
SLIDE 1

Anytime Capacity of Stabilization of a Linear System over Noisy Channel

Graduate Seminar in Area I (6.454) October 26, 2011

1 / 38

slide-2
SLIDE 2

Outline

1 Introduction 2 A Counter Example 3 Necessity of Anytime Capacity 4 Conclusions

2 / 38

slide-3
SLIDE 3

Outline

1 Introduction 2 A Counter Example 3 Necessity of Anytime Capacity 4 Conclusions

3 / 38

slide-4
SLIDE 4

Control and Communications

General Problem: Stabilizing an unstable plant with noisy feedback. How much “information” do we need? What is the correct measure of “information”?

4 / 38

slide-5
SLIDE 5

Control and Communications

General Problem: Stabilizing an unstable plant with noisy feedback. How much “information” do we need? What is the correct measure of “information”?

4 / 38

slide-6
SLIDE 6

Control and Communications

General Problem: Stabilizing an unstable plant with noisy feedback. How much “information” do we need? What is the correct measure of “information”?

4 / 38

slide-7
SLIDE 7

Control and Communications

General Problem: Stabilizing an unstable plant with noisy feedback. How much “information” do we need? What is the correct measure of “information”?

4 / 38

slide-8
SLIDE 8

Control and Communications

Main insights How much “information” do we need?

◮ No single answer. It depends on the degree of “stability” desirable.

What is the correct measure of “information”?

◮ Shannon capacity may not be adequate for stronger notions of

  • stability. Need anytime capacity.

5 / 38

slide-9
SLIDE 9

Control and Communications

Main insights How much “information” do we need?

◮ No single answer. It depends on the degree of “stability” desirable.

What is the correct measure of “information”?

◮ Shannon capacity may not be adequate for stronger notions of

  • stability. Need anytime capacity.

5 / 38

slide-10
SLIDE 10

Plan of This talk

A simple example to illustrate that Shannon capacity is not strong enough for control applications.

◮ In particular, a plant can be unstable even if the Shannon capacity

  • f the channel is infinite.

A necessary condition for stability in terms of anytime capacity.

6 / 38

slide-11
SLIDE 11

Plan of This talk

A simple example to illustrate that Shannon capacity is not strong enough for control applications.

◮ In particular, a plant can be unstable even if the Shannon capacity

  • f the channel is infinite.

A necessary condition for stability in terms of anytime capacity.

6 / 38

slide-12
SLIDE 12

Main Reference

  • A. Sahai, S. K. Mitter, “The Necessity and Sufficiency of Anytime

Capacity for Stabilization of a Linear System Over a Noisy Communication Link. Part I: Scalar Systems,” IEEE Trans. Inform. Th., vol. 52, no. 8, pp. 3369-3395, Aug. 2006.

7 / 38

slide-13
SLIDE 13

Outline

1 Introduction 2 A Counter Example 3 Necessity of Anytime Capacity 4 Conclusions

8 / 38

slide-14
SLIDE 14

The Control Problem

Xt+1 = λXt + Ut + Wt, t ∈ Z+. Time (discrete): t ∈ Z+. State: Xt ∈ R. control: Ut ∈ R. Bounded disturbance: |Wt| < Ω

2 , with probability 1.

To make things interesting: unstable gain: λ > 1.

9 / 38

slide-15
SLIDE 15

The Control Problem

Xt+1 = λXt + Ut + Wt, t ∈ Z+. Goal: choose good Ut to keep Xt “small”. If feedback is perfect, simply set Ut = −λXt. What if feedback is sent through a noisy channel?

10 / 38

slide-16
SLIDE 16

The Control Problem

Xt+1 = λXt + Ut + Wt, t ∈ Z+. Goal: choose good Ut to keep Xt “small”. If feedback is perfect, simply set Ut = −λXt. What if feedback is sent through a noisy channel?

10 / 38

slide-17
SLIDE 17

The Control Problem

Xt+1 = λXt + Ut + Wt, t ∈ Z+. Goal: choose good Ut to keep Xt “small”. If feedback is perfect, simply set Ut = −λXt. What if feedback is sent through a noisy channel?

10 / 38

slide-18
SLIDE 18

Definition of Stability

Observer O: sees Xt and generates channel input at. Controller C: observes channel output Bt and generates control signal Ut.

11 / 38

slide-19
SLIDE 19

The Control Problem

Definition: η-stability

A closed-loop system is η-stable if there exists K < ∞, such that E [|Xt|η] < K for all t ≥ 0. (More general notions of stability can be defined, but we will focus on η-stability for now.)

12 / 38

slide-20
SLIDE 20

Counter Example in Real-Erasure Channel

When is Shannon capacity not sufficient in describing communications in control systems?

Real Erasure Channel (REC)

The real packet erasure channel has Input alphabet: A = R. Output alphabet: B = R. Transition probabilities p(x|x) = 1 − δ, p(0|x) = δ. I.e., a symbol is either received perfectly, or received as zero.

13 / 38

slide-21
SLIDE 21

Counter Example in Real-Erasure Channel

What is the Shannon capacity of the channel? It is infinite, because a real number can carry as many bits as we want.

14 / 38

slide-22
SLIDE 22

Counter Example in Real-Erasure Channel

What is the Shannon capacity of the channel? It is infinite, because a real number can carry as many bits as we want.

14 / 38

slide-23
SLIDE 23

Counter Example in Real-Erasure Channel

What is the optimal communication / control policy? Communication: set at = Xt. Control: set Ut = −λBt. Resulting dynamics: Xt is reset to 0 every Geo(δ) steps.

15 / 38

slide-24
SLIDE 24

Counter Example in Real-Erasure Channel

Is the system η-stable under optimal control? It is 1-stable, E[|Xt|] = 3 2 1 2

  • < 1,

for all t. However, it is not η-stable, for η ≥ 2, E[|Xt|2] > 4σ2 5

t

  • i=0

9 8 i+1 − 1 2 i+1 which diverges as t → ∞.

16 / 38

slide-25
SLIDE 25

Counter Example in Real-Erasure Channel

Lesson learned: notion of information depends on the strength of stability required (e.g., values of η). Why was Shannon capacity insufficient? Need good information about the system state at all times, not just the end of a large block. Fix: define a stronger notion of capacity to guarantee good estimation of system state at any point in time (“anytime capacity”).

17 / 38

slide-26
SLIDE 26

Counter Example in Real-Erasure Channel

Lesson learned: notion of information depends on the strength of stability required (e.g., values of η). Why was Shannon capacity insufficient? Need good information about the system state at all times, not just the end of a large block. Fix: define a stronger notion of capacity to guarantee good estimation of system state at any point in time (“anytime capacity”).

17 / 38

slide-27
SLIDE 27

Counter Example in Real-Erasure Channel

Lesson learned: notion of information depends on the strength of stability required (e.g., values of η). Why was Shannon capacity insufficient? Need good information about the system state at all times, not just the end of a large block. Fix: define a stronger notion of capacity to guarantee good estimation of system state at any point in time (“anytime capacity”).

17 / 38

slide-28
SLIDE 28

Outline

1 Introduction 2 A Counter Example 3 Necessity of Anytime Capacity 4 Conclusions

18 / 38

slide-29
SLIDE 29

Anytime Reliability and Capacity

Communication System

A rate R communication system is Encoder receives R-bit message Mt in slot t. (details on whiteboard) Encoder produces channel input based on all past messages and possible feedback Bt−1−θ

1

(with delay 1 + θ). Decoder updates estimates of all past messages, ˆ Mi(t), for all i ≤ t, based on all channel outputs till time t.

19 / 38

slide-30
SLIDE 30

Anytime Reliability and Capacity

20 / 38

slide-31
SLIDE 31

Anytime Reliability and Capacity

Anytime Reliability

A rate R communication system achieves anytime reliability α if there exists constant K such that P

  • ˆ

Mi

1(t) = Mi 1

  • ≤ K2−α(t−i).

The system is uniformly anytime reliable if the above holds for all messages M. Comparing to Shannon reliability? Block versus sequential? Exercise: fix t or i and vary the other.

21 / 38

slide-32
SLIDE 32

Anytime Reliability and Capacity

α-anytime Capacity

Cany(α) of a channel is the highest rate R, at which the channel can achieve uniform anytime reliability α. More stringent than Shannon capacity, C: Cany(α) ≤ C, for any α > 0.

22 / 38

slide-33
SLIDE 33

Necessity of Anytime Capacity

Theorem: Necessity of Anytime Capacity

If there exists an observer / controller pair that achieves η-stability under bounded disturbance, then the channel’s feedback anytime capacity satisfies Cany(η log2 λ) ≥ log2 λ,

23 / 38

slide-34
SLIDE 34

Necessity of Anytime Capacity: Proof

Use the control system as a black box to construct a communication system with good anytime reliability. (sketch on white board)

1 Encoder sits with the plant; decoder with the controller. 2 Encode messages in the disturbance, Wt. 3 Controller must somehow know the disturbances, otherwise there

is no way to stabilize the plant.

4 Decoder then reads off the control actions chosen by the controller

to decode message.

24 / 38

slide-35
SLIDE 35

Necessity of Anytime Capacity: Proof

Use the control system as a black box to construct a communication system with good anytime reliability. (sketch on white board)

1 Encoder sits with the plant; decoder with the controller. 2 Encode messages in the disturbance, Wt. 3 Controller must somehow know the disturbances, otherwise there

is no way to stabilize the plant.

4 Decoder then reads off the control actions chosen by the controller

to decode message.

24 / 38

slide-36
SLIDE 36

Necessity of Anytime Capacity: Proof

Use the control system as a black box to construct a communication system with good anytime reliability. (sketch on white board)

1 Encoder sits with the plant; decoder with the controller. 2 Encode messages in the disturbance, Wt. 3 Controller must somehow know the disturbances, otherwise there

is no way to stabilize the plant.

4 Decoder then reads off the control actions chosen by the controller

to decode message.

24 / 38

slide-37
SLIDE 37

Necessity of Anytime Capacity: Proof

Use the control system as a black box to construct a communication system with good anytime reliability. (sketch on white board)

1 Encoder sits with the plant; decoder with the controller. 2 Encode messages in the disturbance, Wt. 3 Controller must somehow know the disturbances, otherwise there

is no way to stabilize the plant.

4 Decoder then reads off the control actions chosen by the controller

to decode message.

24 / 38

slide-38
SLIDE 38

Necessity of Anytime Capacity: Proof

25 / 38

slide-39
SLIDE 39

Necessity of Anytime Capacity: Proof

But what do you mean by “knowing the disturbances”? Alright, let’s be more concrete here. Write Xt = Yt + Zt, such that X0 = Y0 = Z0 = 0.

26 / 38

slide-40
SLIDE 40

Necessity of Anytime Capacity: Proof

But what do you mean by “knowing the disturbances”? Alright, let’s be more concrete here. Write Xt = Yt + Zt, such that X0 = Y0 = Z0 = 0.

26 / 38

slide-41
SLIDE 41

Necessity of Anytime Capacity: Proof

Yt is the control branch Yt+1 = λYt + Wt. Zt is the disturbance branch Zt+1 = λZt + Ut. Can easily verify by recursion Xt = Yt + Zt.

27 / 38

slide-42
SLIDE 42

Necessity of Anytime Capacity: Proof

Key idea: encoder can control Wt (hence Yt), while decoder knows Zt perfectly. The plant is η-stable, so |Xt| must be small at all times. Therefore, we must have Yt ≈ −Zt. Voila! Decoder should be able to extract good information of Wt by looking at Zt.

28 / 38

slide-43
SLIDE 43

Necessity of Anytime Capacity: Proof

Key idea: encoder can control Wt (hence Yt), while decoder knows Zt perfectly. The plant is η-stable, so |Xt| must be small at all times. Therefore, we must have Yt ≈ −Zt. Voila! Decoder should be able to extract good information of Wt by looking at Zt.

28 / 38

slide-44
SLIDE 44

Necessity of Anytime Capacity: Proof

Key idea: encoder can control Wt (hence Yt), while decoder knows Zt perfectly. The plant is η-stable, so |Xt| must be small at all times. Therefore, we must have Yt ≈ −Zt. Voila! Decoder should be able to extract good information of Wt by looking at Zt.

28 / 38

slide-45
SLIDE 45

Necessity of Anytime Capacity: Proof

Key idea: encoder can control Wt (hence Yt), while decoder knows Zt perfectly. The plant is η-stable, so |Xt| must be small at all times. Therefore, we must have Yt ≈ −Zt. Voila! Decoder should be able to extract good information of Wt by looking at Zt.

28 / 38

slide-46
SLIDE 46

Part 1: Encoding

Now, down to business: step 1, encoding. Let each message Mt be a collection of R bits. Let Si ∈ {−1, 1} be the ith bit in the system. Write Yt = λYt−1 + Wt−1 = λt−1

t−1

  • j=0

λ−jWj.

29 / 38

slide-47
SLIDE 47

Part 1: Encoding

Now, down to business: step 1, encoding. Let each message Mt be a collection of R bits. Let Si ∈ {−1, 1} be the ith bit in the system. Write Yt = λYt−1 + Wt−1 = λt−1

t−1

  • j=0

λ−jWj.

29 / 38

slide-48
SLIDE 48

Part 1: Encoding

Now, down to business: step 1, encoding. Let each message Mt be a collection of R bits. Let Si ∈ {−1, 1} be the ith bit in the system. Write Yt = λYt−1 + Wt−1 = λt−1

t−1

  • j=0

λ−jWj.

29 / 38

slide-49
SLIDE 49

Part 1: Encoding

Now, down to business: step 1, encoding. Let each message Mt be a collection of R bits. Let Si ∈ {−1, 1} be the ith bit in the system. Write Yt = λYt−1 + Wt−1 = λt−1

t−1

  • j=0

λ−jWj.

29 / 38

slide-50
SLIDE 50

Part 1: Encoding

Encoding: choose Wt to be the value of the fractional representation of {Si}, ⌊Rt⌋ + 1 ≤ i ≤ ⌊R(t + 1)⌋, In particular, set Wt = γλt+1

⌊R(t+1)⌋

  • k=⌊Rt⌋+1

(2 + ǫ1)−k Sk. Need the right constants to make things work ǫ1 = 2

log2 λ R

− 2, γ = Ω 2λ1+ 1

R

.

30 / 38

slide-51
SLIDE 51

Part 1: Encoding

Encoding: choose Wt to be the value of the fractional representation of {Si}, ⌊Rt⌋ + 1 ≤ i ≤ ⌊R(t + 1)⌋, In particular, set Wt = γλt+1

⌊R(t+1)⌋

  • k=⌊Rt⌋+1

(2 + ǫ1)−k Sk. Need the right constants to make things work ǫ1 = 2

log2 λ R

− 2, γ = Ω 2λ1+ 1

R

.

30 / 38

slide-52
SLIDE 52

Part 1: Encoding

Encoding: choose Wt to be the value of the fractional representation of {Si}, ⌊Rt⌋ + 1 ≤ i ≤ ⌊R(t + 1)⌋, In particular, set Wt = γλt+1

⌊R(t+1)⌋

  • k=⌊Rt⌋+1

(2 + ǫ1)−k Sk. Need the right constants to make things work ǫ1 = 2

log2 λ R

− 2, γ = Ω 2λ1+ 1

R

.

30 / 38

slide-53
SLIDE 53

Part 2: Decoding

Sent...fingers crossed...how much separation did we get? Main technical lemma:

Technical Lemma

Let ˆ Si(t) be the estimate of bit Si at time t. For all 0 ≤ j ≤ t,

  • ω
  • ∃i ≤ j, ˆ

Si(t) = ˆ Si(t)

  • ω
  • |Xt| ≥ λt− j

R

γǫ1 1 + ǫ1

  • 31 / 38
slide-54
SLIDE 54

Part 2: Decoding

Sent...fingers crossed...how much separation did we get? Main technical lemma:

Technical Lemma

Let ˆ Si(t) be the estimate of bit Si at time t. For all 0 ≤ j ≤ t,

  • ω
  • ∃i ≤ j, ˆ

Si(t) = ˆ Si(t)

  • ω
  • |Xt| ≥ λt− j

R

γǫ1 1 + ǫ1

  • 31 / 38
slide-55
SLIDE 55

Part 2: Decoding

Technical Lemma

Let ˆ Si(t) be the estimate of bit Si at time t. For all 0 ≤ j ≤ t,

  • ω
  • ∃i ≤ j, ˆ

Si(t) = ˆ Si(t)

  • ω
  • |Xt| ≥ λt− j

R

γǫ1 1 + ǫ1

  • Intuition: if early disturbances (St) were guessed incorrectly,

control will blow up exponentially fast! Hence small |Xt| must imply good estimates of early disturbances.

32 / 38

slide-56
SLIDE 56

Part 2: Decoding

Proof: If two message differ in the first bit, S1, how much will they differ

  • n resulting Yt?

inf

¯ S: ¯ S1=S1

  • Y1(S) − Y1( ¯

S)

γλt    1 −

⌊Rt⌋

  • k=1

(2 + ǫ1)−k   −  −1 −

⌊Rt⌋

  • k=1

(2 + ǫ1)−k     > γλt 2  1 −

⌊Rt⌋

  • k=1

(2 + ǫ1)−k   = λt 2ǫ1γ 1 + ǫ1

  • Note the exponential dependence on λ (will force controller to

report good estimates).

33 / 38

slide-57
SLIDE 57

Part 2: Decoding

More generally, inf

¯ S: ¯ Si=Si

  • Yt(S) − Yt( ¯

S)

  • > λt− i

R

2ǫ1γ 1 + ǫ1

  • ,

if i ≤ ⌊Rt⌋. We will decode to get the ˆ Si by pretending that −Zt is Yt. Complete the proof of Lemma by noting that

  • Zt(S) − Zt( ¯

S)

  • Yt(S) − Yt( ¯

S)

  • Xt(S) − Xt( ¯

S)

  • .

In other words, smallness of Xt guarantees the closeness of −Zt and Yt.

34 / 38

slide-58
SLIDE 58

Part 3: Probability of Error

P (|Xt| > m) = P (|Xt|η > mη) ≤ E (|Xt|η) m−η < Km−η (definition of η-stability). Combine this with the Technical Lemma P

  • ˆ

Si

1(t) = Si 1(t)

P

  • |Xt| > λt− j

R

γǫ1 1 + ǫ1

  • <
  • K

1 γ + 1 γǫ1 η 2−(η log2 λ)(t− i

R).

This proves the theorem.

35 / 38

slide-59
SLIDE 59

Review: we just proved...

Theorem: Necessity of Anytime Capacity

If there exists an observer / controller pair that achieves η-stability under bounded disturbance, then the channel’s feedback anytime capacity satisfies Cany(η log2 λ) ≥ log2 λ. Sufficient conditions for anytime reliability in stabilizing a plant is in the paper, but will not be covered here. Is the necessary condition tight? Are there simpler ways to interpret / proof this result?

36 / 38

slide-60
SLIDE 60

Outline

1 Introduction 2 A Counter Example 3 Necessity of Anytime Capacity 4 Conclusions

37 / 38

slide-61
SLIDE 61

Conclusions

38 / 38

slide-62
SLIDE 62

Concluding Remarks

Thinking about the required information rate for particular application: may need different (or stronger) notion of capacity / reliability. Exponentially unstable nature of linear control system underlies the higher information barrier. Put in an adversarial way, instability and noise are both our enemies in communications. Any other major adversaries that we should consider?

38 / 38