On the Cost of CSI Acquisition in Large MIMO Systems Giuseppe - - PowerPoint PPT Presentation

on the cost of csi acquisition in large mimo systems
SMART_READER_LITE
LIVE PREVIEW

On the Cost of CSI Acquisition in Large MIMO Systems Giuseppe - - PowerPoint PPT Presentation

On the Cost of CSI Acquisition in Large MIMO Systems Giuseppe Durisi Chalmers, Sweden June, 2013 Joint work with Wei Yang , G unther Koliander , Erwin Riegler , Franz Hlawatsch , Tobias Koch , Yury Polyanskiy Many thanks to Ericsson Research


slide-1
SLIDE 1

On the Cost of CSI Acquisition in Large MIMO Systems

Giuseppe Durisi Chalmers, Sweden June, 2013

Joint work with Wei Yang, G¨ unther Koliander, Erwin Riegler, Franz Hlawatsch, Tobias Koch, Yury Polyanskiy Many thanks to Ericsson Research Foundation!

slide-2
SLIDE 2

CSI acquisition limits large-MIMO gains

TX RX . . . TX

?

Pilot symbols

  • G. Durisi

2 / 19

slide-3
SLIDE 3

CSI acquisition limits large-MIMO gains

TX RX . . . TX

?

Pilot symbols

Capacity in the absence of a priori channel knowledge is the ultimate limit on the rate of reliable communication

  • G. Durisi

2 / 19

slide-4
SLIDE 4

Outline

1

Beyond the pre-log

2

Generic block-fading models

3

From asymptotics to finite-blocklength bounds

  • G. Durisi

3 / 19

slide-5
SLIDE 5

A simple channel model

n |hn| L

Constant block-memoryless Rayleigh-fading channel

  • G. Durisi

4 / 19

slide-6
SLIDE 6

Coherence time is the bottleneck

MIMO input-output relation

= × L MT MR X Y S W +

  • G. Durisi

5 / 19

slide-7
SLIDE 7

Coherence time is the bottleneck

MIMO input-output relation

= × L MT MR X Y S W +

No closed-form expression available for C(ρ)

  • G. Durisi

5 / 19

slide-8
SLIDE 8

Coherence time is the bottleneck

MIMO input-output relation

= × L MT MR X Y S W +

No closed-form expression available for C(ρ) Pre-log [Zheng & Tse, 2002] χ = lim

ρ→∞

C(ρ) log ρ = M∗

  • 1 − M∗

L

  • where M∗ = min{MT , MR, L/2}
  • G. Durisi

5 / 19

slide-9
SLIDE 9

The underlying geometry: MT = MR = M

= × L MT MR X Y S

χ = M

  • 1 − M

L

  • G. Durisi

6 / 19

slide-10
SLIDE 10

The underlying geometry: MT = MR = M

= × L MT MR X Y S

χ = M

  • 1 − M

L

  • G. Durisi

6 / 19

slide-11
SLIDE 11

The underlying geometry: MT = MR = M

= × L MT MR X Y S

χ = M

  • 1 − M

L

  • G. Durisi

6 / 19

slide-12
SLIDE 12

The underlying geometry: MT = MR = M

= × L MT MR X Y S

χ = M

  • 1 − M

L

  • Communications on the

Grassmannian manifold

  • G. Durisi

6 / 19

slide-13
SLIDE 13

Geometry suggests a signaling scheme

Uniform distribution on the Grassmannian X =

  • Lρ U

U : (truncated) unitary and isotropically distributed Unitary space-time modulation (USTM)

  • G. Durisi

7 / 19

slide-14
SLIDE 14

A conjecture

Case L ≥ MT + MR (“small MIMO”) [Zheng & Tse (IT 2002)]: C(ρ) = RUSTM(ρ) + o(1)

  • G. Durisi

8 / 19

slide-15
SLIDE 15

A conjecture

Case L ≥ MT + MR (“small MIMO”) [Zheng & Tse (IT 2002)]: C(ρ) = RUSTM(ρ) + o(1) Conjecture for L < MT + MR (“large MIMO”) [Zheng & Tse (IT 2002)]:

USTM not o(1)-optimal

  • G. Durisi

8 / 19

slide-16
SLIDE 16

BSTM is the optimal distribution

[Yang, Durisi, Riegler (JSAC 2013)] BSTM is o(1)-optimal when L < MT + MR (large-MIMO) X = DU with U i.d. and unitary D2 diagonal; contains the eigenvalues of a complex matrix-variate beta distributed matrix

  • G. Durisi

9 / 19

slide-17
SLIDE 17

Why is BSTM optimal? The SIMO case

= × L MR x Y s + W

Large MIMO ⇒ L < 1 + MR

  • G. Durisi

10 / 19

slide-18
SLIDE 18

Why is BSTM optimal? The SIMO case

= × L MR x Y s + W

Large MIMO ⇒ L < 1 + MR USTM ⇒ x i.d., x2 = Lρ

  • G. Durisi

10 / 19

slide-19
SLIDE 19

Why is BSTM optimal? The SIMO case

= × L MR x Y s + W

Large MIMO ⇒ L < 1 + MR USTM ⇒ x i.d., x2 = Lρ BSTM ⇒ x i.d.,

L−1 ρLMR x2 ∼ Beta(L − 1, MR + 1 − L)

  • G. Durisi

10 / 19

slide-20
SLIDE 20

Why is BSTM optimal? The SIMO case

= × L MR x Y s + W

Large MIMO ⇒ L < 1 + MR USTM ⇒ x i.d., x2 = Lρ BSTM ⇒ x i.d.,

L−1 ρLMR x2 ∼ Beta(L − 1, MR + 1 − L)

I(x; Y) = h(Y) − h(Y | x)

  • G. Durisi

10 / 19

slide-21
SLIDE 21

Why is BSTM optimal? The SIMO case

= × L MR x Y s + W

Large MIMO ⇒ L < 1 + MR USTM ⇒ x i.d., x2 = Lρ BSTM ⇒ x i.d.,

L−1 ρLMR x2 ∼ Beta(L − 1, MR + 1 − L)

I(x; Y) = h(Y) − h(Y | x) ≈ h(sx) + 2(L − 1 − MR) E[log x] + const

  • G. Durisi

10 / 19

slide-22
SLIDE 22

Outline

1

Beyond the pre-log

2

Generic block-fading models

3

From asymptotics to finite-blocklength bounds

  • G. Durisi

11 / 19

slide-23
SLIDE 23

The “generic” block-fading model

Constant block-fading model for subchannel (r, t) hr,t = 1L · sr,t, sr,t ∼ CN(0, 1)

  • G. Durisi

12 / 19

slide-24
SLIDE 24

The “generic” block-fading model

Constant block-fading model for subchannel (r, t) hr,t = 1L · sr,t, sr,t ∼ CN(0, 1) A more accurate model for MIMO CP-OFDM systems hr,t = zr,t · sr,t, sr,t ∼ CN(0, 1) zr,t ∈ CL ⇒ Fourier transf. of power-delay profile

  • G. Durisi

12 / 19

slide-25
SLIDE 25

The “generic” block-fading model

Constant block-fading model for subchannel (r, t) hr,t = 1L · sr,t, sr,t ∼ CN(0, 1) A more accurate model for MIMO CP-OFDM systems hr,t = zr,t · sr,t, sr,t ∼ CN(0, 1) zr,t ∈ CL ⇒ Fourier transf. of power-delay profile We assume that {zr,t} are generic

  • G. Durisi

12 / 19

slide-26
SLIDE 26

Generic {zr,t} yield larger pre-log

[Riegler, Koliander, Durisi, Hlawatsch (ISIT 2013)] {zr,t} generic and MR > MT (L−1)

L−T

with MT < L/2

  • G. Durisi

13 / 19

slide-27
SLIDE 27

Generic {zr,t} yield larger pre-log

[Riegler, Koliander, Durisi, Hlawatsch (ISIT 2013)] {zr,t} generic and MR > MT (L−1)

L−T

with MT < L/2 Then χgen = MT

  • 1 − 1

L

  • G. Durisi

13 / 19

slide-28
SLIDE 28

Generic {zr,t} yield larger pre-log

[Riegler, Koliander, Durisi, Hlawatsch (ISIT 2013)] {zr,t} generic and MR > MT (L−1)

L−T

with MT < L/2 Then χgen = MT

  • 1 − 1

L

  • Compare with constant block-fading model

χconst = MT

  • 1 − MT

L

  • G. Durisi

13 / 19

slide-29
SLIDE 29

Intuition behind pre-log increase: MR = 3, MT = 2, L = 4

Constant block-fading: χconst = MT

  • 1 − MT

L

  • = 1

= × L MT MR X Y S

  • G. Durisi

14 / 19

slide-30
SLIDE 30

Intuition behind pre-log increase: MR = 3, MT = 2, L = 4

Constant block-fading: χconst = MT

  • 1 − MT

L

  • = 1

= × L MT MR X Y S

Generic block-fading: χgen = MT

  • 1 − 1

L

  • = 3

2 = + yr diag{zr,1} diag{zr,2} x1 x2 s1,r s2,r

  • G. Durisi

14 / 19

slide-31
SLIDE 31

Outline

1

Beyond the pre-log

2

Generic block-fading models

3

From asymptotics to finite-blocklength bounds

  • G. Durisi

15 / 19

slide-32
SLIDE 32

Lost in “asymptotia”?

  • G. Durisi

16 / 19

slide-33
SLIDE 33

Lost in “asymptotia”?

capacity characterizations up to o(1) yield tight bounds pre-log sensitive to small changes in the channel model

  • G. Durisi

16 / 19

slide-34
SLIDE 34

From asymptotia to tight bounds

[Yang, Durisi, Koch, Polyanskiy (ITW 2012)]

2 4 6 8 10 12 14 16 18 20 0.5 0.55 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 1

SNR [dB] Capacity bounds / Capacity with channel knowledge Lower bound Upper bound L = 20 χ = 1 − 1 20 = 0.95

  • G. Durisi

17 / 19

slide-35
SLIDE 35

From asymptotia to tight bounds

[Yang, Durisi, Koch, Polyanskiy (ITW 2012)]

10 10

1

10

2

10

3

10

4

1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3

P{error} ≤ 10−3 blocklength = 4 × 104 SNR = 10 dB Coherence time L Rate [bits/channel use] Perfect channel knowledge Lower bound Upper bound

  • G. Durisi

17 / 19

slide-36
SLIDE 36

From asymptotia to tight bounds

[Yang, Durisi, Koch, Polyanskiy (ISIT 2013)]

Outage capacity (Cǫ) LTE-Advanced codes Converse Achievability Normal Approximation Blocklength, n Rate, bits/ch. use 1000 100 200 300 400 500 600 700 800 900 0.2 0.4 0.6 0.8 1

  • G. Durisi

17 / 19

slide-37
SLIDE 37

Zero dispersion

AWGN channel [Polyanskiy, Poor, Verd´ u (IT 2010)] R∗

awgn(n, ǫ) = Cawgn −

  • V

n Q−1(ǫ) − O log n n

  • G. Durisi

18 / 19

slide-38
SLIDE 38

Zero dispersion

AWGN channel [Polyanskiy, Poor, Verd´ u (IT 2010)] R∗

awgn(n, ǫ) = Cawgn −

  • V

n Q−1(ǫ) − O log n n

  • SISO quasi static [Yang, Durisi, Koch, Polyanskiy (ISIT 2013)]

{R∗

csirt(n, ǫ), R∗ no(n, ǫ)} = Cǫ −

❅ ❅

  • 1

n − O log n n

  • G. Durisi

18 / 19

slide-39
SLIDE 39

Summary

Capacity without a-priori CSI

TX RX . . . TX

?

Pilot symbols

  • G. Durisi

19 / 19

slide-40
SLIDE 40

Summary

Capacity without a-priori CSI

TX RX . . . TX

?

Pilot symbols

Too conservative estimates? USTM ⇒ BSTM M (1 − M/L) ⇒ M (1 − 1/L)

  • G. Durisi

19 / 19

slide-41
SLIDE 41

Summary

Capacity without a-priori CSI

TX RX . . . TX

?

Pilot symbols

Too conservative estimates? USTM ⇒ BSTM M (1 − M/L) ⇒ M (1 − 1/L) From asymptotia to finite blocklength

10 10 1 10 2 10 3 10 4 1 1.2 1.4 1.6 1.8 2 2.2 2.4 2.6 2.8 3 P{error} ≤ 10−3 blocklength = 4 × 104 SNR = 10 dB Coherence time L Rate [bits/channel use] Perfect channel knowledge Lower bound Upper bound
  • G. Durisi

19 / 19

slide-42
SLIDE 42

Backup Slides

  • G. Durisi

20 / 19

slide-43
SLIDE 43

Gain of BSTM over USTM for large-MIMO systems

10 20 30 40 50 60 70 80 90 100 0.02 0.04 0.06 0.08 0.1 0.12 0.14 RBSTM − RUSTM RUSTM MR L = 100 L = 50 L = 20 L = 10 ρ = 30 dB MT = min{MR, L/2}

  • G. Durisi

21 / 19

slide-44
SLIDE 44

Achievability for finite blocklength

  • G. Durisi

22 / 19