112
Common Information Abbas El Gamal Stanford University Viterbi Lecture, USC, April 2014

CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Embed Size (px)

Citation preview

Page 1: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Common Information

Abbas El Gamal

Stanford University

Viterbi Lecture, USC, April 2014

Page 2: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Andrew Viterbi’s Fabulous Formula, IEEE Spectrum, 2010

El Gamal (Stanford University) Disclaimer Viterbi Lecture 2 / 36

Page 3: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Andrew Viterbi’s Fabulous Formula, IEEE Spectrum, 2010

“A couple of professors at Stanford were being quoted in thepress saying that CDMA violated the laws of physics.”

El Gamal (Stanford University) Disclaimer Viterbi Lecture 2 / 36

Page 4: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

About this lecture

∙ Common information between correlated information sources

El Gamal (Stanford University) Outline Viterbi Lecture 3 / 36

Page 5: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

About this lecture

∙ Common information between correlated information sources

∙ The lecture is tutorial in nature with many examples

El Gamal (Stanford University) Outline Viterbi Lecture 3 / 36

Page 6: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

About this lecture

∙ Common information between correlated information sources

∙ The lecture is tutorial in nature with many examples

∙ No proofs, but a potentially interesting framework and new result

El Gamal (Stanford University) Outline Viterbi Lecture 3 / 36

Page 7: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

About this lecture

∙ Common information between correlated information sources

∙ The lecture is tutorial in nature with many examples

∙ No proofs, but a potentially interesting framework and new result

∙ Outline:

é Brief introduction to information theory

é Information sources and measuring information

El Gamal (Stanford University) Outline Viterbi Lecture 3 / 36

Page 8: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

About this lecture

∙ Common information between correlated information sources

∙ The lecture is tutorial in nature with many examples

∙ No proofs, but a potentially interesting framework and new result

∙ Outline:

é Brief introduction to information theory

é Information sources and measuring information

é Brief introduction to network information theory

é Correlated information sources and measuring common information

El Gamal (Stanford University) Outline Viterbi Lecture 3 / 36

Page 9: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

About this lecture

∙ Common information between correlated information sources

∙ The lecture is tutorial in nature with many examples

∙ No proofs, but a potentially interesting framework and new result

∙ Outline:

é Brief introduction to information theory

é Information sources and measuring information

é Brief introduction to network information theory

é Correlated information sources and measuring common information

∙ Some of the basic models, ideas, and results of information theory

El Gamal (Stanford University) Outline Viterbi Lecture 3 / 36

Page 10: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Information theory

The fundamental problem of communication is that of reproducing atone point either exactly or approximately a message selected at anotherpoint. Frequently the messages have meaning . . . . These semanticaspects of communication are irrelevant to the engineering problem.

– A Mathematical Theory of Communication, Shannon (1948)

El Gamal (Stanford University) Information theory Viterbi Lecture 4 / 36

Page 11: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Information theory

∙ Mathematical models for the source and the channel

El Gamal (Stanford University) Information theory Viterbi Lecture 5 / 36

Page 12: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Information theory

∙ Mathematical models for the source and the channel

∙ Measures of information (entropy and mutual information)

∙ Limits on compression and communication

El Gamal (Stanford University) Information theory Viterbi Lecture 5 / 36

Page 13: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Information theory

∙ Mathematical models for the source and the channel

∙ Measures of information (entropy and mutual information)

∙ Limits on compression and communication

∙ Bits as a universal interface between the source and the channel

El Gamal (Stanford University) Information theory Viterbi Lecture 5 / 36

Page 14: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Information theory

∙ Coding schemes that achieve the Shannon limits

é Viterbi decoding algorithm (Viterbi 1967)

é Lempel–Ziv compression algorithm (Ziv–Lempel 1977, 1978)

é Trellis coded modulation (Ungerboeck 1982)

é Turbo codes (Berrou–Glavieux 1996)

é LDPC codes (Gallager 1963, Richardson–Urbanke 2008)

é Polar codes (Arıkan 2009)

El Gamal (Stanford University) Information theory Viterbi Lecture 5 / 36

Page 15: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Information source

An information source which produces a message or sequence of mes-sages to be communicated to the receiving terminal. The message maybe of various types: e.g. (a) A sequence of letters as in a telegraphof teletype system; (b) A single function of time f (t) as in radio ortelephony; . . .

– A Mathematical Theory of Communication, Shannon (1948)

El Gamal (Stanford University) Information sources Viterbi Lecture 6 / 36

Page 16: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Modeling the information source

We can think of a discrete source as generating the message, symbolby symbol. . . . A physical system, or a mathematical model of a sys-tem which produces such a sequence of symbols governed by a set ofprobabilities, is known as a stochastic process.

El Gamal (Stanford University) Information sources Viterbi Lecture 7 / 36

Page 17: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring information of the source

Can we define a quantity which will measure, in some sense, how muchinformation is “produced” by such a process, or better, at what rate infor-mation is produced?

El Gamal (Stanford University) Information sources Viterbi Lecture 8 / 36

Page 18: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring information of the source

Suppose we have a set of possible events whose probabilities of occurrenceare p1, p2, . . . , pn. . . . If there is such a measure, say H(p1 , p2 , . . . , pn), itis reasonable to require of it the following properties:1. H should be continuous in the pi .2. If all the pi are equal, H should be monotonic increasing function of n.3. If a choice be broken down into two successive choices, the original Hshould be the weighted sum of the individual values of H .

El Gamal (Stanford University) Information sources Viterbi Lecture 8 / 36

Page 19: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring information of the source

Theorem 2: The only H satisfying the three above assumptions is of theform:

H = −K nHi=1

pi log pi

El Gamal (Stanford University) Information sources Viterbi Lecture 8 / 36

Page 20: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring information of the source

Theorem 2: The only H satisfying the three above assumptions is of theform:

H = −K nHi=1

pi log pi

This theorem, and the assumptions required for its proof, are in no waynecessary for the present theory. . . . The real justification of these defini-tions, however, will reside in their implications.

– A Mathematical Theory of Communication, Shannon (1948)

El Gamal (Stanford University) Information sources Viterbi Lecture 8 / 36

Page 21: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Setups for measuring information

∙ Entropy arises naturally as a measure of information in many settings

El Gamal (Stanford University) Measuring information Viterbi Lecture 9 / 36

Page 22: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Setups for measuring information

∙ Entropy arises naturally as a measure of information in many settings

∙ Will describe three of them

é Zero-error compression

é Randomness extraction

é Source simulation

El Gamal (Stanford University) Measuring information Viterbi Lecture 9 / 36

Page 23: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Setups for measuring information

∙ Entropy arises naturally as a measure of information in many settings

∙ Will describe three of them

é Zero-error compression

é Randomness extraction

é Source simulation

∙ Assume a discrete memoryless source (X , p(x)) (DMS X)

X generates independent identically distributed (i.i.d.) X1 , X2 , . . . ∼ p(x)

El Gamal (Stanford University) Measuring information Viterbi Lecture 9 / 36

Page 24: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Setups for measuring information

∙ Entropy arises naturally as a measure of information in many settings

∙ Will describe three of them

é Zero-error compression

é Randomness extraction

é Source simulation

∙ Assume a discrete memoryless source (X , p(x)) (DMS X)

X generates independent identically distributed (i.i.d.) X1 , X2 , . . . ∼ p(x)∙ Example: Bernoulli source with parameter p ∈ [0, 1] (Bern(p) source)

X = {0, 1}, pX(1) = p; generates i.i.d X1 , X2 , . . . ∼ Bern(p)

El Gamal (Stanford University) Measuring information Viterbi Lecture 9 / 36

Page 25: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Setups for measuring information

∙ Entropy arises naturally as a measure of information in many settings

∙ Will describe three of them

é Zero-error compression

é Randomness extraction

é Source simulation

∙ Assume a discrete memoryless source (X , p(x)) (DMS X)

X generates independent identically distributed (i.i.d.) X1 , X2 , . . . ∼ p(x)∙ Example: Bernoulli source with parameter p ∈ [0, 1] (Bern(p) source)

X = {0, 1}, pX(1) = p; generates i.i.d X1 , X2 , . . . ∼ Bern(p)∙ Example: “DNA source,” X = {A,G,C,T}; p(x) = [p1 p2 p3 p4]

El Gamal (Stanford University) Measuring information Viterbi Lecture 9 / 36

Page 26: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Zero-error compression

Xn c(Xn) ∈ {0, 1}∗ Xn

Encoder DecoderDMS X

∙ Xn = (X1 , X2 , . . . , Xn) i.i.d. ∼ p(x)∙ Variable-length prefix-free code: c(xn) ∈ {0, 1}∗ = {0, 1, 00, 01, 10, 11, . . .}

El Gamal (Stanford University) Measuring information Viterbi Lecture 10 / 36

Page 27: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Zero-error compression

Xn c(Xn) ∈ {0, 1}∗ Xn

Encoder DecoderDMS X

∙ Xn = (X1 , X2 , . . . , Xn) i.i.d. ∼ p(x)∙ Variable-length prefix-free code: c(xn) ∈ {0, 1}∗ = {0, 1, 00, 01, 10, 11, . . .}∙ Example: X = {A,G,C,T}, p(x) = [5/8 1/4 1/16 1/16], n = 1

0

0

0

1

1

1

0 ← A

10 ← G

110 ← C

111 ← T

El Gamal (Stanford University) Measuring information Viterbi Lecture 10 / 36

Page 28: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Zero-error compression

Xn c(Xn) ∈ {0, 1}∗ Xn

Encoder DecoderDMS X

∙ Xn = (X1 , X2 , . . . , Xn) i.i.d. ∼ p(x)∙ Variable-length prefix-free code: c(xn) ∈ {0, 1}∗ = {0, 1, 00, 01, 10, 11, . . .}∙ Example: X = {A,G,C,T}, p(x) = [5/8 1/4 1/16 1/16], n = 1

0

0

0

1

1

1

0 ← A

10 ← G

110 ← C

111 ← T∙ Let Ln be the codeword length and Rn = (1/n) E(Ln) bits/symbol

E(L1) = 1 ⋅ 58+ 2 ⋅ 1

4+ 3 ⋅ 1

8= 1.5 bits

El Gamal (Stanford University) Measuring information Viterbi Lecture 10 / 36

Page 29: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Zero-error compression

Xn c(Xn) ∈ {0, 1}∗ Xn

Encoder DecoderDMS X

∙ Xn = (X1 , X2 , . . . , Xn) i.i.d. ∼ p(x)∙ Variable-length prefix-free code: c(xn) ∈ {0, 1}∗ = {0, 1, 00, 01, 10, 11, . . .}∙ Let Ln be the codeword length and Rn = (1/n) E(Ln) bits/symbol∙ Measure the information rate of X by: R∗ = infn mincodes Rn

El Gamal (Stanford University) Measuring information Viterbi Lecture 10 / 36

Page 30: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Zero-error compression

Xn c(Xn) ∈ {0, 1}∗ Xn

Encoder DecoderDMS X

∙ Xn = (X1 , X2 , . . . , Xn) i.i.d. ∼ p(x)∙ Variable-length prefix-free code: c(xn) ∈ {0, 1}∗ = {0, 1, 00, 01, 10, 11, . . .}∙ Let Ln be the codeword length and Rn = (1/n) E(Ln) bits/symbol∙ Measure the information rate of X by: R∗ = infn mincodes Rn

Information rate is the entropy (Shannon 1948)

R∗ = H(X) = Hx

−p(x) log p(x) bits/symbol

El Gamal (Stanford University) Measuring information Viterbi Lecture 10 / 36

Page 31: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Entropy examples

∙ X ∼ Bern(p): H(X) = H(p) = −p log p − (1 − p) log(1 − p)

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1H(X)

p

El Gamal (Stanford University) Measuring information Viterbi Lecture 11 / 36

Page 32: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Entropy examples

∙ X ∼ Bern(p): H(X) = H(p) = −p log p − (1 − p) log(1 − p)

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1H(X)

p∙ For DNA source: H(X) = 1.424 bits (E(L1) = 1.5)

El Gamal (Stanford University) Measuring information Viterbi Lecture 11 / 36

Page 33: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Randomness extraction

. . . , X2 , X1 Bn

Extractor

∙ X is a DMS; Bn is an i.i.d. ∼ Bern(1/2) sequence∙ Let Ln be length of X sequence and Rn = n/ E(Ln) bits/symbol

El Gamal (Stanford University) Measuring information Viterbi Lecture 12 / 36

Page 34: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Randomness extraction

. . . , X2 , X1 Bn

Extractor

∙ X is a DMS; Bn is an i.i.d. ∼ Bern(1/2) sequence∙ Let Ln be length of X sequence and Rn = n/ E(Ln) bits/symbol∙ Example: Let X be Bern(1/3) source, n = 1 (B1 ∼ Bern(1/2))

0

1

X1 = 0

X2 = 0

X1 = 1

X2 = 1

01 → 0

10 → 1

0001 → 0

0010 → 1

1101 → 0

1110 → 1

El Gamal (Stanford University) Measuring information Viterbi Lecture 12 / 36

Page 35: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Randomness extraction

. . . , X2 , X1 Bn

Extractor

∙ X is a DMS; Bn is an i.i.d. ∼ Bern(1/2) sequence∙ Let Ln be length of X sequence and Rn = n/ E(Ln) bits/symbol∙ Example: Let X be Bern(1/3) source, n = 1 (B1 ∼ Bern(1/2))

0

1

X1 = 0

X2 = 0

X1 = 1

X2 = 1

01 → 0

10 → 1

0001 → 0

0010 → 1

1101 → 0

1110 → 1

E(L1) = 4

9⋅ 2 + 5

9⋅ (2 + E(L1)) ⇒ R1 = 1/E(L1) = 2

9

El Gamal (Stanford University) Measuring information Viterbi Lecture 12 / 36

Page 36: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Randomness extraction

. . . , X2 , X1 Bn

Extractor

∙ X is a DMS; Bn is an i.i.d. ∼ Bern(1/2) sequence∙ Let Ln be length of X sequence and Rn = n/ E(Ln) bits/symbol∙ The information rate of X: R∗ = supn maxextractor Rn

El Gamal (Stanford University) Measuring information Viterbi Lecture 12 / 36

Page 37: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Randomness extraction

. . . , X2 , X1 Bn

Extractor

∙ X is a DMS; Bn is an i.i.d. ∼ Bern(1/2) sequence∙ Let Ln be length of X sequence and Rn = n/ E(Ln) bits/symbol∙ The information rate of X: R∗ = supn maxextractor Rn

Information rate is the entropy (Elias 1972)

R∗ = H(X) bits/symbol

El Gamal (Stanford University) Measuring information Viterbi Lecture 12 / 36

Page 38: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Source simulation

. . . , B2 , B1 Xn

Generator

∙ B is a Bern(1/2) source; Xn i.i.d. ∼ p(x) sequence∙ Let Ln be length of the B sequence and Rn = (1/n) E(Ln) bits/symbol

El Gamal (Stanford University) Measuring information Viterbi Lecture 13 / 36

Page 39: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Source simulation

. . . , B2 , B1 Xn

Generator

∙ B is a Bern(1/2) source; Xn i.i.d. ∼ p(x) sequence∙ Let Ln be length of the B sequence and Rn = (1/n) E(Ln) bits/symbol∙ Example: Let X be Bern(1/3) source, n = 1

B1 = 0

B1 = 1

B2 = 0

B2 = 1

0 → 0

10 → 1

110 → 0

1110 → 1

P{X1 = 1} = 1

4+ 1

42+ 1

43+ ⋅ ⋅ ⋅ = 1

3

E(L1) = 1

2⋅ 1 + 1

4⋅ 2 + 1

4⋅ (2 + E(L1)) ⇒ E(L1) = 2

El Gamal (Stanford University) Measuring information Viterbi Lecture 13 / 36

Page 40: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Source simulation

. . . , B2 , B1 Xn

Generator

∙ B is a Bern(1/2) source; Xn i.i.d. ∼ p(x) sequence∙ Let Ln be length of the B sequence and Rn = (1/n) E(Ln) bits/symbol∙ The information rate of X: R∗ = infn mingenerator Rn

El Gamal (Stanford University) Measuring information Viterbi Lecture 13 / 36

Page 41: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Source simulation

. . . , B2 , B1 Xn

Generator

∙ B is a Bern(1/2) source; Xn i.i.d. ∼ p(x) sequence∙ Let Ln be length of the B sequence and Rn = (1/n) E(Ln) bits/symbol∙ The information rate of X: R∗ = infn mingenerator Rn

Information rate is the entropy (Knuth–Yao 1976)

R∗ = H(X) bits/symbol

El Gamal (Stanford University) Measuring information Viterbi Lecture 13 / 36

Page 42: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ Entropy arises naturally as a measure of information rate:

é The minimum average description length in bits/symbol of X

é The maximum number of bits/symbol that can be extracted from X

é The minimum number of bits/symbol needed to simulate X

El Gamal (Stanford University) Measuring information Viterbi Lecture 14 / 36

Page 43: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Correlated sources

∙ Sensor network

Los Angeles

San Diego

San Francisco

El Gamal (Stanford University) Correlated sources Viterbi Lecture 15 / 36

Page 44: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Correlated sources

∙ Wireless multipath

X1

X2

X3

El Gamal (Stanford University) Correlated sources Viterbi Lecture 15 / 36

Page 45: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Correlated sources

∙ Distributed secret key generation

X Y

Key Key

El Gamal (Stanford University) Correlated sources Viterbi Lecture 15 / 36

Page 46: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Correlated sources

∙ Distributed simulation

W

X1

X2

X3

X4

El Gamal (Stanford University) Correlated sources Viterbi Lecture 15 / 36

Page 47: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Correlated sources

How to measure common information between correlated sources

El Gamal (Stanford University) Measuring common information Viterbi Lecture 16 / 36

Page 48: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Network information theory

Communication networkSource

Node

∙ Establishes limits on communication/distributed processing in networks

El Gamal (Stanford University) Measuring common information Viterbi Lecture 16 / 36

Page 49: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Network information theory

Communication networkSource

Node

∙ Establishes limits on communication/distributed processing in networks∙ First paper (Shannon 1961): “Two-way communication channels”∙ Significant progress in 1970s (Cover 1972, Slepian–Wolf 1973)∙ Internet and wireless communication revived interest (EG–Kim 2011)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 16 / 36

Page 50: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring common information

∙ Setups for measuring common information:

é Distributed compression

é Distributed key generation

é Distributed simulation

El Gamal (Stanford University) Measuring common information Viterbi Lecture 17 / 36

Page 51: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring common information

∙ Setups for measuring common information:

é Distributed compression

é Distributed key generation

é Distributed simulation∙ Assume a 2-discrete memoryless source (X × Y , p(x, y)) (2-DMS (X , Y ))(X ,Y ) generates i.i.d. sequence (X1 ,Y1), (X2 ,Y2), . . . ∼ p(x, y)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 17 / 36

Page 52: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring common information

∙ Setups for measuring common information:

é Distributed compression

é Distributed key generation

é Distributed simulation∙ Assume a 2-discrete memoryless source (X × Y , p(x, y)) (2-DMS (X , Y ))(X ,Y ) generates i.i.d. sequence (X1 ,Y1), (X2 ,Y2), . . . ∼ p(x, y)∙ Examples:

0

1

0

11 − p

1 − p

X ∼ Bern(1/2) Y

Doubly symmetric binary source

DSBS

El Gamal (Stanford University) Measuring common information Viterbi Lecture 17 / 36

Page 53: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Measuring common information

∙ Setups for measuring common information:

é Distributed compression

é Distributed key generation

é Distributed simulation∙ Assume a 2-discrete memoryless source (X × Y , p(x, y)) (2-DMS (X , Y ))(X ,Y ) generates i.i.d. sequence (X1 ,Y1), (X2 ,Y2), . . . ∼ p(x, y)∙ Examples:

0

1

0

11 − p

1 − p

X ∼ Bern(1/2) Y

0

1

0

1

e

1 − p

1 − p

X ∼ Bern(1/2) Y

Doubly symmetric binary source Symmetric binary erasure source

DSBS SBES

El Gamal (Stanford University) Measuring common information Viterbi Lecture 17 / 36

Page 54: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed zero-error compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn ,Yn

c1(Xn)c2(Yn)

∙ Let (X ,Y ) be a 2-DMS; prefix-free codes c1(xn), c2(yn)∙ Let L jn be length of c j and R jn = (1/n) E(L jn), j = 1, 2

El Gamal (Stanford University) Measuring common information Viterbi Lecture 18 / 36

Page 55: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed zero-error compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn ,Yn

c1(Xn)c2(Yn)

∙ Let (X ,Y ) be a 2-DMS; prefix-free codes c1(xn), c2(yn)∙ Let L jn be length of c j and R jn = (1/n) E(L jn), j = 1, 2∙ Measure the common information rate between X and Y by

R∗C = H(X) + H(Y ) − inf

nmincodes

(R1n + R2n) bits/symbol-pair

El Gamal (Stanford University) Measuring common information Viterbi Lecture 18 / 36

Page 56: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed zero-error compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn ,Yn

c1(Xn)c2(Yn)

∙ Let (X ,Y ) be a 2-DMS; prefix-free codes c1(xn), c2(yn)∙ Let L jn be length of c j and R jn = (1/n) E(L jn), j = 1, 2∙ Measure the common information rate between X and Y by

R∗C = H(X) + H(Y ) − inf

nmincodes

(R1n + R2n) bits/symbol-pair

∙ Example: X = (U ,W), Y = (V ,W), where U ,V ,W are independent:

R∗C = (H(U) + H(W)) + (H(V ) + H(W)) − (H(U) + H(V ) + H(W))

El Gamal (Stanford University) Measuring common information Viterbi Lecture 18 / 36

Page 57: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed zero-error compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn ,Yn

c1(Xn)c2(Yn)

∙ Measure the common information rate between X and Y by

R∗C = H(X) + H(Y ) − inf

nmincodes

(R1n + R2n) bits/symbol-pair

∙ Example: SBES (H(X) = 1, H(Y ) = (1 − p) + H(p))

0

1

0

1

e

1 − p

1 − p

X ∼ Bern(1/2) Y

El Gamal (Stanford University) Measuring common information Viterbi Lecture 18 / 36

Page 58: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed zero-error compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn ,Yn

c1(Xn)c2(Yn)

∙ Example: SBES (H(X) = 1, H(Y ) = (1 − p) + H(p))Can show that R∗

C = 1 − p:

é Encoder 1 sends Xn: R1n = 1 bit/symbol

é Encoder 2 sends erasure location sequence: infn R2n = H(p) bit/symbol

⇒ R∗C ≤ H(X) +H(Y) − (R1n + infn R2n) = 1 − p

El Gamal (Stanford University) Measuring common information Viterbi Lecture 18 / 36

Page 59: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed zero-error compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn ,Yn

c1(Xn)c2(Yn)

∙ Example: For SBES, R∗C = 1 − p∙ Example: For DSBS, R∗C = 0 for 0 < p < 1

0

1

0

11 − p

1 − p

X ∼ Bern(1/2) Y

El Gamal (Stanford University) Measuring common information Viterbi Lecture 18 / 36

Page 60: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed zero-error compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn ,Yn

c1(Xn)c2(Yn)

∙ Example: For SBES, R∗C = 1 − p∙ Example: For DSBS, R∗C = 0 for 0 < p < 1∙ No computable expression for R∗

C is known (Korner–Orlitsky 1998)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 18 / 36

Page 61: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed secret key generation

Xn Yn

Kn Kn

Alice Bob

∙ (X ,Y ) is a 2-DMS∙ Define key rate as Rn = (1/n)H(Kn) bits/symbol-pair

El Gamal (Stanford University) Measuring common information Viterbi Lecture 19 / 36

Page 62: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed secret key generation

Xn Yn

Kn Kn

Alice Bob

∙ (X ,Y ) is a 2-DMS∙ Define key rate as Rn = (1/n)H(Kn) bits/symbol-pair∙ The common information rate: R∗ = supn maxextractors Rn

El Gamal (Stanford University) Measuring common information Viterbi Lecture 19 / 36

Page 63: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed secret key generation

Xn Yn

Kn Kn

Alice Bob

∙ (X ,Y ) is a 2-DMS∙ Define key rate as Rn = (1/n)H(Kn) bits/symbol-pair∙ The common information rate: R∗ = supn maxextractors Rn∙ Example: X = (U ,W), Y = (V ,W), where U ,V ,W are independent

Then again R∗ = H(W)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 19 / 36

Page 64: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed secret key generation

Xn Yn

Kn Kn

Alice Bob

∙ Example: Consider (X , Y ) with pmf:

XY

W = 1 W = 2

W = 1

W = 2 0

0

0

00

0

0

0

1

1

2

2

3

3

4

4

0.10.1

0.10.1

0.1

0.1

0.2

0.2

El Gamal (Stanford University) Measuring common information Viterbi Lecture 19 / 36

Page 65: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed secret key generation

Xn Yn

Kn Kn

Alice Bob

∙ Example: Consider (X , Y ) with pmf:

XY

W = 1 W = 2

W = 1

W = 2 0

0

0

00

0

0

0

1

1

2

2

3

3

4

4

0.10.1

0.10.1

0.1

0.1

0.2

0.2

∙ Alice and Bob can agree on i.i.d. key Wn ⇒ Rn = H(W)El Gamal (Stanford University) Measuring common information Viterbi Lecture 19 / 36

Page 66: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed secret key generation

Xn Yn

Kn Kn

Alice Bob

∙ Example: Consider (X , Y ) with pmf:

XY

W = 1 W = 2

W = 1

W = 2 0

0

0

00

0

0

0

1

1

2

2

3

3

4

4

0.10.1

0.10.1

0.1

0.1

0.2

0.2

∙ Alice and Bob can agree on i.i.d. key Wn ⇒ Rn = H(W)∙ Turns out: R∗ = H(W) = 0.881 bits/symbol-pair

El Gamal (Stanford University) Measuring common information Viterbi Lecture 19 / 36

Page 67: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Gacs–Korner (1973), Witsenhausen (1975) common information

∙ Arrange p(x, y) in block diagonal form with largest # of blocks:

XY

W = 1

W = 2

W = k

0

00

0

0

0

∙ W is called the common part between X and Y

It is the highest entropy r.v. that X and Y can agree on

K(X ; Y ) = H(W) is called GKW common information

El Gamal (Stanford University) Measuring common information Viterbi Lecture 20 / 36

Page 68: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Gacs–Korner (1973), Witsenhausen (1975) common information

∙ Arrange p(x, y) in block diagonal form with largest # of blocks:

XY

W = 1

W = 2

W = k

0

00

0

0

0

∙ W is called the common part between X and Y∙ The common part between Xn and Yn is Wn

El Gamal (Stanford University) Measuring common information Viterbi Lecture 20 / 36

Page 69: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Gacs–Korner (1973), Witsenhausen (1975) common information

∙ Arrange p(x, y) in block diagonal form with largest # of blocks:

XY

W = 1

W = 2

W = k

0

00

0

0

0

Common information rate is the GKW common information

R∗ = K(X ; Y ) = H(W)El Gamal (Stanford University) Measuring common information Viterbi Lecture 20 / 36

Page 70: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed simulation

Wn

( Alice

Bob

Xn

Yn

∙ Wn is a common information random variable∙ Wish to generate (Xn ,Yn) i.i.d. ∼ p(x, y)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 21 / 36

Page 71: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed simulation

Wn

( Alice

Bob

Xn

Yn

∙ Wn is a common information random variable∙ Wish to generate (Xn ,Yn) i.i.d. ∼ p(x, y)∙ Define the distributed simulation rate Rn = (1/n)H(Wn)∙ The common information rate: R∗S = infn minWn , generators

Rn

El Gamal (Stanford University) Measuring common information Viterbi Lecture 21 / 36

Page 72: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed simulation

Wn

( Alice

Bob

Xn

Yn

∙ Wn is a common information random variable∙ Wish to generate (Xn ,Yn) i.i.d. ∼ p(x, y)∙ Define the distributed simulation rate Rn = (1/n)H(Wn)∙ The common information rate: R∗S = infn minWn , generators

Rn∙ Example: X = (U ,W), Y = (V ,W), where U ,V ,W are independent

Then again R∗S = H(W) (set Wn =Wn)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 21 / 36

Page 73: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed simulation

Wn

( Alice

Bob

Xn

Yn

∙ Wn is a common information random variable∙ Wish to generate (Xn ,Yn) i.i.d. ∼ p(x, y)∙ Define the distributed simulation rate Rn = (1/n)H(Wn)∙ The common information rate: R∗S = infn minWn , generators

Rn∙ No computable expression for R∗S is known

El Gamal (Stanford University) Measuring common information Viterbi Lecture 21 / 36

Page 74: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed simulation

∙ Example: SBES (Kumar–Li–EG, ISIT 2014)

0

1

0

1

e

1 − p

1 − p

X ∼ Bern(1/2) Y

El Gamal (Stanford University) Measuring common information Viterbi Lecture 22 / 36

Page 75: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed simulation

∙ Example: SBES (Kumar–Li–EG, ISIT 2014)

R∗S = ®1 if p ≤ 1/2

H(p) if p > 1/2

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1

R∗ S

p

El Gamal (Stanford University) Measuring common information Viterbi Lecture 22 / 36

Page 76: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ Entropy is a “universal” measure of information

El Gamal (Stanford University) Measuring common information Viterbi Lecture 23 / 36

Page 77: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ Entropy is a “universal” measure of information

∙ For X = (U ,W), Y = (V ,W), where U ,V ,W are independent

é Common information is measured by H(W)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 23 / 36

Page 78: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ Entropy is a “universal” measure of information

∙ In general, common information has several measures:

é For zero-error distributed compression: R∗C (no computable expression)

é For distributed key generation: K(X ;Y) (GKW common information)

é For distributed simulation: R∗S (no computable expression)

El Gamal (Stanford University) Measuring common information Viterbi Lecture 23 / 36

Page 79: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ Entropy is a “universal” measure of information

∙ In general, common information has several measures:

é For zero-error distributed compression: R∗C (no computable expression)

é For distributed key generation: K(X ;Y) (GKW common information)

é For distributed simulation: R∗S (no computable expression)

∙ We can say more by considering relaxed versions of these setups

é Distributed lossless compression

é Distributed approximate simulation

∙ Common information for approximate key generation same as for exact

El Gamal (Stanford University) Measuring common information Viterbi Lecture 23 / 36

Page 80: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Lossless compression

Xn Mn ∈ [1 : 2nR] Xn

Encoder Decoder

∙ Let X be a DMS∙ Use fixed-length (block) codes (Mn ∈ {1, . . . , 2nR} is an nR-bit sequence)∙ Probability of error: P(n)e = P{Xn = Xn}

El Gamal (Stanford University) Approximate common information Viterbi Lecture 24 / 36

Page 81: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Lossless compression

Xn Mn ∈ [1 : 2nR] Xn

Encoder Decoder

∙ Let X be a DMS∙ Use fixed-length (block) codes (Mn ∈ {1, . . . , 2nR} is an nR-bit sequence)∙ Probability of error: P(n)e = P{Xn = Xn}∙ R is achievable if ∃ a sequence of codes such that limn→∞ P(n)

e = 0∙ The information rate of X: R∗ = inf{R : R is achievable}

El Gamal (Stanford University) Approximate common information Viterbi Lecture 24 / 36

Page 82: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Lossless compression

Xn Mn ∈ [1 : 2nR] Xn

Encoder Decoder

∙ Let X be a DMS∙ Use fixed-length (block) codes (Mn ∈ {1, . . . , 2nR} is an nR-bit sequence)∙ Probability of error: P(n)e = P{Xn = Xn}∙ R is achievable if ∃ a sequence of codes such that limn→∞ P(n)

e = 0∙ The information rate of X: R∗ = inf{R : R is achievable}Information rate is the entropy (Shannon 1948)

R∗ = H(X)∙ Same as in zero-error!

El Gamal (Stanford University) Approximate common information Viterbi Lecture 24 / 36

Page 83: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed lossless compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn , Yn

M1n ∈ [1 : 2nR1]M2n ∈ [1 : 2nR2]

∙ Again assume fixed length codes

∙ Probability of error: P(n)e = P{(Xn , Yn) = (Xn ,Yn)}

∙ (R1 , R2) achievable if ∃ a sequence of codes such that limn→∞ P(n)e = 0

El Gamal (Stanford University) Approximate common information Viterbi Lecture 25 / 36

Page 84: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed lossless compression

Xn

Yn

Encoder 1

Encoder 2

DecoderXn , Yn

M1n ∈ [1 : 2nR1]M2n ∈ [1 : 2nR2]

∙ Again assume fixed length codes

∙ Probability of error: P(n)e = P{(Xn , Yn) = (Xn ,Yn)}

∙ (R1 , R2) achievable if ∃ a sequence of codes such that limn→∞ P(n)e = 0

∙ The common information rate

R∗ = H(X) + H(Y ) − inf{R1 + R2 : (R1 , R2) is achievable}

El Gamal (Stanford University) Approximate common information Viterbi Lecture 25 / 36

Page 85: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed lossless compression

∙ The common information rate

R∗ = H(X) + H(Y ) − inf{R1 + R2 : (R1 , R2) is achievable}Common information rate is mutual information (Slepian–Wolf 1973)

R∗ = I(X ; Y ) = H(X) + H(Y ) −H(X , Y )∙ The same compression rate as if X and Y are compressed together!

El Gamal (Stanford University) Approximate common information Viterbi Lecture 26 / 36

Page 86: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed lossless compression

∙ The common information rate

R∗ = H(X) + H(Y ) − inf{R1 + R2 : (R1 , R2) is achievable}Common information rate is mutual information (Slepian–Wolf 1973)

R∗ = I(X ; Y ) = H(X) + H(Y ) −H(X , Y )∙ Example: (X ,Y ) is SBES, I(X ; Y ) = 1 − p = R∗

C (zero-error)

0

1

0

1

e

1 − p

1 − p

X ∼ Bern(1/2) Y

El Gamal (Stanford University) Approximate common information Viterbi Lecture 26 / 36

Page 87: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed lossless compression

∙ The common information rate

R∗ = H(X) + H(Y ) − inf{R1 + R2 : (R1 , R2) is achievable}Common information rate is mutual information (Slepian–Wolf 1973)

R∗ = I(X ; Y ) = H(X) + H(Y ) −H(X , Y )∙ Example: (X ,Y ) is SBES, I(X ; Y ) = 1 − p = R∗

C (zero-error)

∙ Example: (X ,Y ) is DSBS, I(X ; Y ) = 1 − H(p), R∗C = 0, 0 < p < 1

0

1

0

11 − p

1 − p

X ∼ Bern(1/2) Y

El Gamal (Stanford University) Approximate common information Viterbi Lecture 26 / 36

Page 88: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed lossless compression

∙ The common information rate

R∗ = H(X) + H(Y ) − inf{R1 + R2 : (R1 , R2) is achievable}Common information rate is mutual information (Slepian–Wolf 1973)

R∗ = I(X ; Y ) = H(X) + H(Y ) −H(X , Y )∙ Example: (X ,Y ) is SBES, I(X ; Y ) = 1 − p = R∗

C (zero-error)

∙ Example: (X ,Y ) is DSBS, I(X ; Y ) = 1 − H(p), R∗C = 0, 0 < p < 1∙ Also, I(X ; Y ) = 0 only if X and Y are independent∙ Hence, captures dependence better than previous measures

El Gamal (Stanford University) Approximate common information Viterbi Lecture 26 / 36

Page 89: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed approximate simulation

Wn ∈ [1 : 2nR] Alice

Bob

Xn

Yn

∙ Wn randomly distributed over [1 : 2nR] (random nR-bit sequence)∙ Wish to simulate (Xn ,Yn) i.i.d. ∼ p(x, y) approximately

El Gamal (Stanford University) Approximate common information Viterbi Lecture 27 / 36

Page 90: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed approximate simulation

Wn ∈ [1 : 2nR] Alice

Bob

Xn

Yn

∙ Wn randomly distributed over [1 : 2nR] (random nR-bit sequence)∙ Wish to simulate (Xn ,Yn) i.i.d. ∼ p(x, y) approximately∙ Total variation distance between (Xn , Yn) and (Xn ,Yn):dn = H

(xn ,yn)

!!!! pXn ,Yn(xn , yn) − nIi=1

pX ,Y (xi , yi) !!!!

El Gamal (Stanford University) Approximate common information Viterbi Lecture 27 / 36

Page 91: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed approximate simulation

Wn ∈ [1 : 2nR] Alice

Bob

Xn

Yn

∙ Wn randomly distributed over [1 : 2nR] (random nR-bit sequence)∙ Wish to simulate (Xn ,Yn) i.i.d. ∼ p(x, y) approximately∙ Total variation distance between (Xn , Yn) and (Xn ,Yn):dn = H

(xn ,yn)

!!!! pXn ,Yn(xn , yn) − nIi=1

pX ,Y (xi , yi) !!!!∙ R is achievable if ∃ a sequence of generators with limn→∞ dn = 0∙ The common information rate: R∗ = inf{R : R is achievable}

El Gamal (Stanford University) Approximate common information Viterbi Lecture 27 / 36

Page 92: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed approximate simulation

Common information rate is Wyner (1975) common information

R∗ = J(X ;Y ) = minX→W→Y

I(X ,Y ;W)

El Gamal (Stanford University) Approximate common information Viterbi Lecture 28 / 36

Page 93: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed approximate simulation

Common information rate is Wyner (1975) common information

R∗ = J(X ;Y ) = minX→W→Y

I(X ,Y ;W)∙ Example: SBES

0

1

0

1

e

1 − p

1 − p

X ∼ Bern(1/2) Y

El Gamal (Stanford University) Approximate common information Viterbi Lecture 28 / 36

Page 94: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed approximate simulation

Common information rate is Wyner (1975) common information

R∗ = J(X ;Y ) = minX→W→Y

I(X ,Y ;W)∙ Example: SBES

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1

J(X;Y

)

p

El Gamal (Stanford University) Approximate common information Viterbi Lecture 28 / 36

Page 95: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Distributed approximate simulation

Common information rate is Wyner (1975) common information

R∗ = J(X ;Y ) = minX→W→Y

I(X ,Y ;W)∙ Example: SBES, J(X ; Y ) = R∗

S (exact common information)

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1

J(X;Y

)

p

El Gamal (Stanford University) Approximate common information Viterbi Lecture 28 / 36

Page 96: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ There are several well-motivated measures of common information

El Gamal (Stanford University) Summary Viterbi Lecture 29 / 36

Page 97: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ There are several well-motivated measures of common information

∙ For distributed compression:

é Zero-error: R∗C (no computable expression)

é Lossless: I(X ;Y) (mutual information)

El Gamal (Stanford University) Summary Viterbi Lecture 29 / 36

Page 98: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ There are several well-motivated measures of common information

∙ For distributed compression:

é Zero-error: R∗C (no computable expression)

é Lossless: I(X ;Y) (mutual information)

∙ For distributed key generation: K(X ; Y ) (GKW common information)

é Same for approximate key agreement

El Gamal (Stanford University) Summary Viterbi Lecture 29 / 36

Page 99: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Summary

∙ There are several well-motivated measures of common information

∙ For distributed compression:

é Zero-error: R∗C (no computable expression)

é Lossless: I(X ;Y) (mutual information)

∙ For distributed key generation: K(X ; Y ) (GKW common information)

é Same for approximate key agreement

∙ For distributed simulation:

é Exact simulation: R∗S (no computable expression)

é Approximate simulation: J(X ;Y) (Wyner common information)

El Gamal (Stanford University) Summary Viterbi Lecture 29 / 36

Page 100: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Comparing common information measures

0 ≤ K(X ; Y ) ≤ R∗C ≤ I(X ; Y ) ≤ J(X ;Y ) ≤ R∗

S ≤ min{H(X),H(Y )}

El Gamal (Stanford University) Comparing common information measures Viterbi Lecture 30 / 36

Page 101: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Comparing common information measures

0 ≤ K(X ; Y ) ≤ R∗C ≤ I(X ; Y ) ≤ J(X ;Y ) ≤ R∗

S ≤ min{H(X),H(Y )}

∙ Common information between X and Y ≤ information of each∙ Strict inequality for SBES:

é H(X) = 1, H(Y) = (1 − p) + H(p)é R∗

S = H(p) if p > 1/2

El Gamal (Stanford University) Comparing common information measures Viterbi Lecture 30 / 36

Page 102: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Comparing common information measures

0 ≤ K(X ; Y ) ≤ R∗C ≤ I(X ; Y ) ≤ J(X ;Y ) ≤ R∗

S ≤ min{H(X),H(Y )}

∙ Common information for approximate distributed simulation ≤ for exact∙ Equal for SBES (ISIT 2014)

∙ Open problem: Are they equal in general?

El Gamal (Stanford University) Comparing common information measures Viterbi Lecture 30 / 36

Page 103: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Comparing common information measures

0 ≤ K(X ; Y ) ≤ R∗C ≤ I(X ; Y ) ≤ J(X ;Y ) ≤ R∗

S ≤ min{H(X),H(Y )}

∙ Strict inequality for SBES:

é I(X ;Y) = 1 − p

é J(X ;Y) = 1 if p ≤ 1/2

El Gamal (Stanford University) Comparing common information measures Viterbi Lecture 30 / 36

Page 104: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Comparing common information measures

0 ≤ K(X ; Y ) ≤ R∗C ≤ I(X ; Y ) ≤ J(X ;Y ) ≤ R∗

S ≤ min{H(X),H(Y )}

∙ Common information for distributed zero-error compression ≤ for lossless∙ Strict inequality for DSBS:

é R∗C = 0

é I(X ;Y) = 1 −H(p)

El Gamal (Stanford University) Comparing common information measures Viterbi Lecture 30 / 36

Page 105: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Comparing common information measures

0 ≤ K(X ; Y ) ≤ R∗C ≤ I(X ; Y ) ≤ J(X ;Y ) ≤ R∗

S ≤ min{H(X),H(Y )}

∙ Common part needs to be sent only once in distributed compression∙ Strict inequality for SBEC:

é K(X ;Y) = 0

é R∗C = 1 − p

El Gamal (Stanford University) Comparing common information measures Viterbi Lecture 30 / 36

Page 106: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Conclusion

∙ Entropy is a universal measure of information

é For both zero-error and lossless compression

é For randomness generation and source simulation

El Gamal (Stanford University) Conclusion Viterbi Lecture 31 / 36

Page 107: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Conclusion

∙ Entropy is a universal measure of information

é For both zero-error and lossless compression

é For randomness generation and source simulation

∙ The story for common information is much richer:

é Five well-motivated common information measures

K ≤ R∗C ≤ I ≤ J ≤ R∗

S

é They are not always equal!

é Some don’t seem to be computable in general

é Different for zero-error distributed compression from lossless

é Are they also different for exact distributed simulation from approximate?

El Gamal (Stanford University) Conclusion Viterbi Lecture 31 / 36

Page 108: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Acknowledgments

∙ Work on exact distributed simulation is joint with:

é Gowtham Kumar

é Cheuk Ting Li

∙ Young-Han Kim provided many valuable suggestions

El Gamal (Stanford University) Acknowledgments Viterbi Lecture 32 / 36

Page 109: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

Thank You!

Page 110: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

References

Arıkan, E. (2009). Channel polarization: A method for constructing capacity-achievingcodes for symmetric binary-input memoryless channels. IEEE Trans. Inf. Theory,55(7), 3051–3073.

Berrou, C. and Glavieux, A. (1996). Near optimum error correcting coding anddecoding: Turbo-codes. IEEE Trans. Comm., 44(10), 1261–1271.

Cover, T. M. (1972). Broadcast channels. IEEE Trans. Inf. Theory, 18(1), 2–14.

El Gamal, A. and Kim, Y.-H. (2011). Network Information Theory. Cambridge,Cambridge.

Elias, P. (1972). The efficient construction of an unbiased random sequence. TheAnnals of Mathematical Statistics, 43(3), 865–870.

Gacs, P. and Korner, J. (1973). Common information is far less than mutualinformation. Probl. Control Inf. Theory, 2(2), 149–162.

Gallager, R. G. (1963). Low-Density Parity-Check Codes. MIT Press, Cambridge, MA.

Knuth, D. E. and Yao, A. C. (1976). The complexity of random number generation. InAlgorithms and Complexity (Proc. Symp., Carnegie Mellon Univ., Pittsburgh, Pa.,1976), pp. 357–428. Academic Press, New York.

El Gamal (Stanford University) Acknowledgments Viterbi Lecture 34 / 36

Page 111: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

References (cont.)

Korner, J. and Orlitsky, A. (1998). Zero-error information theory. IEEE Trans. Inf.Theory, 44(6), 2207–2229.

Orlitsky, A. and Roche, J. R. (2001). Coding for computing. IEEE Trans. Inf. Theory,47(3), 903–917.

Richardson, T. and Urbanke, R. (2008). Modern Coding Theory. Cambridge UniversityPress, Cambridge.

Shannon, C. E. (1948). A mathematical theory of communication. Bell Syst. Tech. J.,27(3), 379–423, 27(4), 623–656.

Shannon, C. E. (1961). Two-way communication channels. In Proc. 4th Berkeley Symp.Math. Statist. Probab., vol. I, pp. 611–644. University of California Press, Berkeley.

Slepian, D. and Wolf, J. K. (1973). Noiseless coding of correlated information sources.IEEE Trans. Inf. Theory, 19(4), 471–480.

Ungerboeck, G. (1982). Channel coding with multilevel/phase signals. IEEE Trans. Info.Theory, 28(1), 55–67.

Viterbi, A. J. (1967). Error bounds for convolutional codes and an asymptoticallyoptimum decoding algorithm. IEEE Trans. Inf. Theory, 13(2), 260–269.

El Gamal (Stanford University) Acknowledgments Viterbi Lecture 35 / 36

Page 112: CommonInformation - Stanford Universityisl.stanford.edu/~abbas/presentations/lect00-viterbi.pdf · CommonInformation Abbas El Gamal ... ∙ The lecture is tutorial in nature with

References (cont.)

Witsenhausen, H. S. (1975). On sequences of pairs of dependent random variables.SIAM J. Appl. Math., 28(1), 100–113.

Wyner, A. D. (1975). The common information of two dependent random variables.IEEE Trans. Inf. Theory, 21(2), 163–179.

Ziv, J. and Lempel, A. (1977). A universal algorithm for sequential data compression.IEEE Trans. Inf. Theory, 23(3), 337–343.

Ziv, J. and Lempel, A. (1978). Compression of individual sequences via variable-ratecoding. IEEE Trans. Inf. Theory, 24(5), 530–536.

El Gamal (Stanford University) Acknowledgments Viterbi Lecture 36 / 36