Fast Polarization for Processes with Memory Boaz Shuval and Ido Tal - - PowerPoint PPT Presentation

fast polarization for processes with memory
SMART_READER_LITE
LIVE PREVIEW

Fast Polarization for Processes with Memory Boaz Shuval and Ido Tal - - PowerPoint PPT Presentation

Fast Polarization for Processes with Memory Boaz Shuval and Ido Tal Andrew and Erna Viterbi Department of Electrical Engineering Technion Israel Institute of Technology Haifa, 32000, Israel June 2018 1/21 In this talk Setting :


slide-1
SLIDE 1

Fast Polarization for Processes with Memory

Boaz Shuval and Ido Tal

Andrew and Erna Viterbi Department of Electrical Engineering Technion — Israel Institute of Technology Haifa, 32000, Israel

June 2018

1/21

slide-2
SLIDE 2

In this talk

Setting: binary-input, symmetric, memoryless channel

2/21

slide-3
SLIDE 3

In this talk

Setting: binary-input, symmetric, memoryless channel

2/21

slide-4
SLIDE 4

In this talk

Setting: binary-input, symmetric, memoryless channel

2/21

slide-5
SLIDE 5

In this talk

Setting: binary-input, symmetric, memoryless channel

2/21

slide-6
SLIDE 6

In this talk

Setting: binary-input, symmetric, ✭✭✭✭✭✭

✭ ❤❤❤❤❤❤ ❤

memoryless channel

2/21

slide-7
SLIDE 7

Polar codes: [Arıkan:09], [ArıkanTelatar:09], [S

¸as ¸o˘ glu+:09], [KoradaUrbanke:10], [HondaYamamoto:13]

◮ Setting: Memoryless i.i.d. process (Xi, Yi)N

i=1

◮ For simplicity: Assume Xi binary ◮ Polar transform: UN

1 = XN 1 · GN

◮ Index sets: Low entropy: ΛN =

  • i : H(Ui|Ui−1

1

, YN

1 ) < 2−Nβ

High entropy: ΩN =

  • i : H(Ui|Ui−1

1

, YN

1 ) > 1 − 2−Nβ

◮ Polarization: lim

N→∞

1 N |ΛN| = 1 − H(X1|Y1) lim

N→∞

1 N |ΩN| = H(X1|Y1)

3/21

slide-8
SLIDE 8

Polar codes:

Optimal rate for:

◮ Coding for non-symmetric memoryless channels ◮ Coding for memoryless channels with non-binary inputs ◮ (Lossy) compression of memoryless sources

Question

◮ How to handle memory?

4/21

slide-9
SLIDE 9

A framework for memory

◮ Process: (Xi, Yi, Si)N

i=1

◮ Finite number of states: Si ∈ S, where |S| < ∞ ◮ Hidden state: Si is unknown to encoder and decoder

◮ Probability distribution: P(xi, yi, si|si−1)

◮ Stationary: same for all i ◮ Markov:

P(xi, yi, si|si−1) = P(xi, yi, si|{xj, yj, sj}j<i)

◮ State sequence: aperiodic and irreducible Markov chain

5/21

slide-10
SLIDE 10

Example 1

◮ Model: Finite state channel Ps(y|x) , s ∈ S ◮ Input distribution: Xi i.i.d. and independent of state ◮ State transition: π(si|si−1) ◮ Distribution: P(xi, yi, si|si−1) = P(xi)π(si|si−1)Psi(yi|xi)

6/21

slide-11
SLIDE 11

Example 2

◮ Model: ISI + noise Yi = h0Xi + h1Xi−1 + · · · + hmXi−m + noise ◮ Input: Xi has memory P(xi|xi−1, xi−2, . . . , xi−m, xi−m−1) ◮ State: Si =

  • Xi

Xi−1 · · · Xi−m

  • ◮ Distribution: For xi, si, si−1 compatible,

P(xi, yi, si|si−1) = Pnoise(yi|hTsi) · P(xi|si−1)

7/21

slide-12
SLIDE 12

Example 3

◮ Model: (d, k)-RLL constrained system with noise 1

(1, ∞)-RLL Constraint

8/21

slide-13
SLIDE 13

Example 3

◮ Model: (d, k)-RLL constrained system with noise 1

(1, ∞)-RLL Constraint

BSC(p) XN

1

YN

1

8/21

slide-14
SLIDE 14

Example 3

◮ Model: (d, k)-RLL constrained system with noise 1

(1, ∞)-RLL Constraint

1 − α 1 α

state Markov chain

BSC(p) XN

1

YN

1

1/1

( 1 − α ) ( 1 − p )

1/0

(1 − α)(p)

0/0

1 ( 1 − p )

1/0

1(p)

0/0

α(1 − p)

0/1

α(p)

P(xi, yi, si|si−1)

8/21

slide-15
SLIDE 15

Polar codes: [S

¸as ¸o˘ glu:11], [S ¸as ¸o˘ gluTal:16], [ShuvalTal:17]

◮ Setting: Process (Xi, Yi, Si)N

i=1 with memory, as above

◮ Hidden state: State unknown to encoder and decoder ◮ Polar transform: UN

1 = XN 1 · GN

UN

1 are neither independent, nor identically distributed

◮ Index sets: Low entropy: ΛN =

  • i : H(Ui|Ui−1

1

, YN

1 ) < 2−Nβ

High entropy: ΩN =

  • i : H(Ui|Ui−1

1

, YN

1 ) > 1 − 2−Nβ

◮ Polarization: lim

N→∞

1 N |ΛN| = 1 − H⋆(X|Y) lim

N→∞

1 N |ΩN| = H⋆(X|Y) limN→∞ 1

NH(XN 1|YN 1 )

9/21

slide-16
SLIDE 16

Achievable rate

◮ Achievable rate: In all examples, R approaches I⋆(X; Y) = lim

N→∞

1 NI(XN

1; YN 1 )

◮ Also lossy compression of a source with memory ◮ Successive cancellation: [Wang+:15] ◮ Without state estimation!

10/21

slide-17
SLIDE 17

Three parameters

◮ Joint distribution P(x, y) ◮ For simplicity: X ∈ {0, 1} ◮ Parameters: ◮ Connections: H ≈ 0 ⇐ ⇒ Z ≈ 0 ⇐ ⇒ K ≈ 1 H ≈ 1 ⇐ ⇒ Z ≈ 1 ⇐ ⇒ K ≈ 0 H(X|Y) = −

x,y P(x, y) log P(x|y)

Z(X|Y) = 2

y

  • P(0, y)P(1, y)

K(X|Y) =

y |P(0, y) − P(1, y)|

Entropy Bhattacharyya T.V. distance

11/21

slide-18
SLIDE 18

Three processes

For n = 1, 2, . . . ◮ N = 2n ◮ UN

1 = XN 1GN

◮ Pick Bn ∈ {0, 1} uniform, i.i.d. ◮ Random index from {1, 2, . . . , N} i = 1 + B1 B2 · · · Bn2 ◮ Processes: {Xi, Yi, Si} {Xi, Yi} Hn = H(Ui|Ui−1

1

, YN

1 )

Zn = Z(Ui|Ui−1

1

, YN

1 )

Kn = K(Ui|Ui−1

1

, YN

1 )

Entropy Bhattacharyya T.V. distance

12/21

slide-19
SLIDE 19

Proof — memoryless case

Slow polarization

Hn ∈ (ǫ, 1 − ǫ) |Hn+1 − Hn| > 0

Fast polarization

Zn+1 ≤

  • 2Zn

Bn+1 = 0 Z2

n

Bn+1 = 1 1 N|ΛN| − − − →

n→∞ 1 − H(X1|Y1)

Kn+1 ≤

  • K2

n

Bn+1 = 0 2Kn Bn+1 = 1 1 N|ΩN| − − − →

n→∞ H(X1|Y1)

New

Low entropy set High entropy set 13/21

slide-20
SLIDE 20

Proof — memory✟✟

✟ ❍❍ ❍

less case [S

¸as ¸o˘ gluTal:16], [ShuvalTal:17]

Slow polarization

Hn ∈ (ǫ, 1 − ǫ) |Hn+1 − Hn| > 0

Fast polarization

Zn+1 ≤

  • 2ψZn

Bn+1 = 0 ψZ2

n

Bn+1 = 1 1 N|ΛN| − − − →

n→∞ 1 − H⋆(X|Y)

ˆ Kn+1 ≤

  • ψˆ

K2

n

Bn+1 = 0 2ˆ Kn Bn+1 = 1 1 N|ΩN| − − − →

n→∞ H⋆(X|Y)

{Xi, Yi, Si} {Xi, Yi}

ψ = ψ(0) = max

s

1 π(s) π: stationary state distribution

Low entropy set High entropy set

New

14/21

slide-21
SLIDE 21

Fast polarization to high entropy set ΩN

◮ Memoryless case:

◮ Parameter evolution inequality hinges on independence:

P(x2N

1 , y2N 1 ) = P(xN 1, yN 1) · P(x2N N+1, y2N N+1)

◮ Memory case:

◮ Force independence: condition on middle state SN

P(x2N

1 , y2N 1 |sN) = P(xN 1, yN 1|sN) · P(x2N N+1, y2N N+1|sN)

◮ New processes: ˆ Hn = H(Ui|Ui−1

1

, YN

1 , S0, SN)

ˆ Kn = K(Ui|Ui−1

1

, YN

1 , S0, SN)

15/21

slide-22
SLIDE 22

Polarization of Kn (memoryless case)

◮ Memoryless assumption: P(ui, vi, qi, ri) = P(ui, qi) · P(vi, ri) ◮ Notation: Ti = Ui + Vi ◮ One step polarization: Kn+1 =

  • K(Ti|Qi, Ri)

Bn+1 = 0 ‘−’ transform K(Vi|Ti, Qi, Ri) Bn+1 = 1 ‘+’ transform ◮ Recall: K(X|Y) =

  • y

|P(0, y) − P(1, y)|

UN

1 = XN 1 · GN

VN

1 = X2N N+1 · GN

Qi = (Ui−1

1

, YN

1 )

Ri = (Vi−1

1

, Y2N

N+1) 16/21

slide-23
SLIDE 23

Polarization of Kn (memoryless case), ‘−’ transform

Kn+1 =

  • q,r

|PTi,Qi,Ri(0, q, r) − PTi,Qi,Ri(1, q, r)| =

  • q,r
  • 1
  • v=0

P(v, r)(P(v, q) − P(v + 1, q))

  • =
  • q,r
  • P(0, q) − P(1, q)
  • P(0, r) − P(1, r)
  • =
  • q,r

|P(0, q) − P(1, q)| · |P(0, r) − P(1, r)| =

  • q

|P(0, q) − P(1, q)| ·

  • r

|P(0, r) − P(1, r)| = K2

n,

17/21

slide-24
SLIDE 24

Polarization of Kn (memoryless case), ‘+’ transform

Kn+1 =

  • t,q,r
  • PTi,Vi,Qi,Ri(t, 0, q, r) − PTi,Vi,Qi,Ri(t, 1, q, r)
  • =
  • t,q,r

|P(t, q)P(0, r) − P(t + 1, q)P(1, r)|

(∗)

≤ 1 2

  • t,q,r

P(q) |P(0, r) − P(1, r)| + P(r) |P(t, q) − P(t + 1, q)| = 1 2

  • t,r

|P(0, r) − P(1, r)| + 1 2

  • t,q

|P(t, q) − P(t + 1, q)| = 2Kn, Identity for (∗): For any a, b, c, d: ab − cd = (a + c)(b − d) + (b + d)(a − c) 2

18/21

slide-25
SLIDE 25

Polarization of ˆ Kn (memory)

◮ Follows steps of memoryless case ◮ Requires additional inequalities

◮ Inequality I: For states s0, sN, s2N ∈ S,

P(s0, sN, s2N) = P(s0, sN) · P(sN, s2N) P(sN) ≤ ψ · P(s0, sN) · P(sN, s2N) where ψ = max

s

1 π(s)

◮ Inequality II: For f, g ≥ 0,

  • sN

f(sN)g(sN) ≤

  • sN

f(sN)

  • s′

N

g(s′

N)

19/21

slide-26
SLIDE 26

Connections

Extreme Values

H ≈ 0 ⇔ Z ≈ 0 ⇔ K ≈ 1 H ≈ 1 ⇔ Z ≈ 1 ⇔ K ≈ 0 also for ˆ ( · ) processes

Ordering

ˆ Hn ≤ Hn ˆ Zn ≤ Zn ˆ Kn ≥ Kn All six processes (Hn, ˆ Hn, Zn, ˆ Zn, Kn, ˆ Kn) polarize fast both to 0 and 1 with any β < 1/2

20/21

slide-27
SLIDE 27

Summary

◮ A general framework for memory: P(xi, yi, si|si−1)

◮ Memory allowed in both source and channel ◮ State sequence Si

◮ Hidden ◮ Stationary ◮ Finite state Markov ◮ Aperiodic and irreducible

◮ Achieve rate I⋆(X; Y) through polar codes ◮ No change to polarization exponent (β < 1/2)

21/21