54
1 Information and interactive computation January 16, 2012 Mark Braverman Computer Science, Princeton University

Information and interactive computation

Embed Size (px)

DESCRIPTION

Information and interactive computation. Mark Braverman Computer Science, Princeton University. January 16, 2012. Prelude: one-way communication. Basic goal : send a message from Alice to Bob over a channel. communication channel. Bob. Alice. One-way communication. Encode; Send; - PowerPoint PPT Presentation

Citation preview

Page 1: Information and interactive computation

1

Information and interactive computation

January 16, 2012

Mark BravermanComputer Science, Princeton University

Page 2: Information and interactive computation

Prelude: one-way communication

• Basic goal: send a message from Alice to Bob over a channel.

2

communication channel

Alice Bob

Page 3: Information and interactive computation

One-way communication1) Encode;2) Send;3) Decode.

3

communication channel

Alice Bob

Page 4: Information and interactive computation

Coding for one-way communication• There are two main problems a good

encoding needs to address:– Efficiency: use the least amount of the

channel/storage necessary.– Error-correction: recover from (reasonable)

errors;

4

Page 5: Information and interactive computation

Interactive computation

Today’s themeExtending information and coding theory to interactive computation.

5

I will talk about interactive information theory and Anup Rao will talk about

interactive error correction.

Page 6: Information and interactive computation

Efficient encoding• Can measure the cost of storing a random

variable X very precisely. • Entropy: H(X) = ∑Pr[X=x] log(1/Pr[X=x]).• H(X) measures the average amount of

information a sample from X reveals. • A uniformly random string of 1,000 bits has

1,000 bits of entropy.

6

Page 7: Information and interactive computation

Efficient encoding

7

• H(X) = ∑Pr[X=x] log(1/Pr[X=x]).• The ZIP algorithm works because

H(X=typical 1MB file) < 8Mbits.• P[“Hello, my name is Bob”] >>

P[“h)2cjCv9]dsnC1=Ns{da3”].• For one-way encoding, Shannon’s source

coding theorem states that

Communication ≈ Information.

Page 8: Information and interactive computation

Efficient encoding

8

• The problem of sending many samples of X can be implemented in H(X) communication on average.

• The problem of sending a single sample of X can be implemented in <H(X)+1 communication in expectation.

Page 9: Information and interactive computation

Communication complexity [Yao]

• Focus on the two party setting.

9

A B

X YA & B implement a functionality F(X,Y).

F(X,Y)

e.g. F(X,Y) = “X=Y?”

Page 10: Information and interactive computation

Communication complexity

10

A B

X Y

Goal: implement a functionality F(X,Y).A protocol π(X,Y) computing F(X,Y):

F(X,Y)

m1(X)m2(Y,m1)

m3(X,m1,m2)

Communication cost = #of bits exchanged.

Page 11: Information and interactive computation

Distributional communication complexity

• The input pair (X,Y) is drawn according to some distribution μ.

• Goal: make a mistake on at most an ε fraction of inputs.

• The communication cost: C(F,μ,ε):C(F,μ,ε) := minπ computes F with error≤ε C(π, μ).

11

Page 12: Information and interactive computation

Example

12

μ is a distribution of pairs of files. F is “X=Y?”:

MD5(X) (128b)

X=Y? (1b)

Communication cost = 129 bits. ε ≈ 2-128.

A B

X Y

Page 13: Information and interactive computation

Randomized communication complexity

• Goal: make a mistake of at most ε on every input.

• The communication cost: R(F,ε).• Clearly: C(F,μ,ε)≤R(F,ε) for all μ.• What about the converse?• A minimax(!) argument [Yao]:

R(F,ε)=maxμ C(F,μ,ε).

13

Page 14: Information and interactive computation

A note about the model

• We assume a shared public source of randomness.

14A B

X YR

Page 15: Information and interactive computation

The communication complexity of EQ(X,Y)

• The communication complexity of equality:R(EQ,ε) ≈ log 1/ε

• Send log 1/ε random hash functions applied to the inputs. Accept if all of them agree.

• What if ε=0?R(EQ,0) ≈ n,

where X,Y in {0,1}n.15

Page 16: Information and interactive computation

Information in a two-way channel

• H(X) is the “inherent information cost” of sending a message distributed according to X over the channel.

16

communication channel

Alice BobXWhat is the two-way

analogue of H(X)?

Page 17: Information and interactive computation

Entropy of interactive computation

A B

X YR

• “Inherent information cost” of interactive two-party tasks.

Page 18: Information and interactive computation

One more definition: Mutual Information

• The mutual information of two random variables is the amount of information knowing one reveals about the other:

I(A;B) = H(A)+H(B)-H(AB)• If A,B are independent, I(A;B)=0.• I(A;A)=H(A).

18

H(A) H(B)I(A,B)

Page 19: Information and interactive computation

Information cost of a protocol

• [Chakrabarti-Shi-Wirth-Yao-01, Bar-Yossef-Jayram-Kumar-Sivakumar-04, Barak-B-Chen-Rao-10].

• Caution: different papers use “information cost” to denote different things!

• Today, we have a better understanding of the relationship between those different things.

19

Page 20: Information and interactive computation

Information cost of a protocol • Prior distribution: (X,Y) ~ μ.

A B

X Y

Protocol πProtocol transcript π

I(π, μ) = I(π;Y|X) + I(π;X|Y) what Alice learns about Y + what Bob learns about X

Page 21: Information and interactive computation

External information cost• (X,Y) ~ μ.

A B

X Y

Protocol πProtocol transcript π

Iext(π, μ) = I(π;XY) what Charlie learns about (X,Y)

C

Page 22: Information and interactive computation

Another view on I and Iext

• It is always the case that C(π, μ) ≥ Iext(π, μ) ≥ I(π, μ).

• Iext measures the ability of Alice and Bob to compute F(X,Y) in an information theoretically secure way if they are afraid of an eavesdropper.

• I measures the ability of the parties to compute F(X,Y) if they are afraid of each other.

Page 23: Information and interactive computation

Example

23

• F is “X=Y?”.• μ is a distribution where w.p. ½ X=Y and w.p. ½

(X,Y) are random.

MD5(X) [128b]

X=Y?A B

X Y

Iext(π, μ) = I(π;XY) = 129 bits what Charlie learns about (X,Y)

Page 24: Information and interactive computation

Example• F is “X=Y?”.• μ is a distribution where w.p. ½ X=Y and w.p. ½

(X,Y) are random.

MD5(X) [128b]

X=Y?A B

X Y

I(π, μ) = I(π;Y|X)+I(π;X|Y) ≈what Alice learns about Y + what Bob learns about X

1 + 64.5 = 65.5 bits

Page 25: Information and interactive computation

The (distributional) information cost of a problem F

• Recall:C(F,μ,ε) := minπ computes F with error≤ε C(π, μ).

• By analogy:I(F, μ, ε) := infπ computes F with error≤ε I(π, μ).

Iext (F, μ, ε) := infπ computes F with error≤ε Iext (π, μ).

25

Page 26: Information and interactive computation

I(F,μ,ε) vs. C(F,μ,ε): compressing interactive computation

Source Coding Theorem: the problem of sending a sample of X can be

implemented in expected cost <H(X)+1 communication – the information

content of X.

Is the same compression true for interactive protocols?

Can F be solved in I(F,μ,ε) communication?

Or in Iext(F,μ,ε) communication?

Page 27: Information and interactive computation

The big question

• Can interactive communication be compressed?

• Can π be simulated by π’ such that C(π’, μ) ≈ I(π, μ)?

Does I(F,μ,ε) ≈ C(F,μ,ε)?

27

Page 28: Information and interactive computation

Compression results we know• Let ε, ρ be constants; let π be a protocol

that computes F with error ε.• π’s costs: C, Iext, I.• Then π can be simulated using:

– (I·C)½·polylog(C) communication; [Barak-B-Chen-Rao’10]

– Iext·polylog(C) communication; [Barak-B-Chen-Rao’10]

– 2O(I) communication; [B’11]

while introducing an extra error of ρ.28

Page 29: Information and interactive computation

The amortized cost of interactive computation

Source Coding Theorem: the amortized cost of sending many independent

samples of X is =H(X).

What is the amortized cost of computing many independent

copies of F(X,Y)?

Page 30: Information and interactive computation

Information = amortized communication

• Theorem[B-Rao’11]: for ε>0I(F,μ,ε) = limn→∞ C(Fn,μn,ε)/n.

• I(F,μ,ε) is the interactive analogue of H(X).

30

Page 31: Information and interactive computation

Information = amortized communication

• Theorem[B-Rao’11]: for ε>0I(F,μ,ε) = limn→∞ C(Fn,μn,ε)/n.

• I(F,μ,ε) is the interactive analogue of H(X).• Can we get rid of μ? I.e. make I(F,ε) a

property of the task F?

C(F,μ,ε) I(F,μ,ε) R(F,ε) ?

Page 32: Information and interactive computation

Prior-free information cost

• Define:I(F,ε) := infπ computes F with error≤ε maxμ I(π, μ)

• Want a protocol that reveals little information against all priors μ!

• Definitions are cheap!• What is the connection between the

“syntactic” I(F,ε) and the “meaningful” I(F,μ,ε)?

• I(F,μ,ε) ≤ I(F,ε)…32

Page 33: Information and interactive computation

33

Prior-free information cost

• I(F,ε) := infπ computes F with error ≤ε maxμ I(π, μ).• I(F,μ,ε) ≤ I(F,ε) for all μ.• Recall: R(F,ε)=maxμ C(F,μ,ε).• Theorem[B’11]:

I(F,ε) ≤ 2·maxμ I(F,μ,ε/2).

I(F,0) = maxμ I(F,μ,0).

Page 34: Information and interactive computation

34

Prior-free information cost

• Recall: I(F,μ,ε) = limn→∞ C(Fn,μn,ε)/n.• Theorem: for ε>0

I(F,ε) = limn→∞ R(Fn,ε)/n.

Page 35: Information and interactive computation

Example

• R(EQ,0) ≈ n.• What is I(EQ,0)?

35

Page 36: Information and interactive computation

The information cost of Equality

• What is I(EQ,0)?• Consider the following protocol.

36A B

X in {0,1}n Y in {0,1}n

A non-singular in nn2Z

A1·XA1·Y

A2·XA2·Y

Continue for n steps, or until a disagreement is

discovered.

Page 37: Information and interactive computation

Analysis (sketch)

• If X≠Y, the protocol will terminate in O(1) rounds on average, and thus reveal O(1) information.

• If X=Y… the players only learn the fact that X=Y (≤1 bit of information).

• Thus the protocol has O(1) information complexity.

37

Page 38: Information and interactive computation

Direct sum theorems

• I(F,ε) = limn→∞ R(Fn,ε)/n.• Questions:

– Does R(Fn,ε)=Ω(n·R(F,ε))?– Does R(Fn,ε)=ω(R(F,ε))?

38

Page 39: Information and interactive computation

Direct sum strategy

• The strategy for proving direct sum results. • Take a protocol for Fn that costs Cn=R(Fn,ε),

and make a protocol for F that costs ≈Cn/n.• This would mean that C<Cn/n, i.e. Cn>n C.∙

39

~ ~

A protocol for n

copies of F

1 copy of FCnCn/n?

Page 40: Information and interactive computation

Direct sum strategy

• If life were so simple…

40

1 copy of FCnCn/nEasy!

Copy 1Copy 2

Copy n

Page 41: Information and interactive computation

Direct sum strategy

• Theorem: I(F,ε) = I(Fn,ε)/n ≤ Cn = R(Fn,ε)/n.• Compression → direct sum!

41

Page 42: Information and interactive computation

The information cost angle

• There is a protocol of communication cost Cn, but information cost ≤Cn/n.

1 copy of F

Cn

Cn/n

Restriction

1 bit

Copy

1

Copy

2

Copy

n

C n/n

info

Compression?

Page 43: Information and interactive computation

Direct sum theorems

Best known general simulation [BBCR’10]:• A protocol with C communication and I

information cost can be simulated using (I·C)½·polylog(C) communication.

• Implies: R(Fn,ε) = Ω(n1/2 R(F,∙ ε)).

43

~

Page 44: Information and interactive computation

Compression vs. direct sum

• We saw that compression → direct sum. • A form of the converse is also true. • Recall: I(F,ε) = limn→∞ R(Fn,ε)/n.• If there is a problem such that I(F,ε)=o(R(F,ε)),

then R(Fn,ε)=o(n·R(F,ε)).

44

Page 45: Information and interactive computation

A complete problem

• Can define a problem called Correlated Pointer Jumping – CPJ(C,I).

• The problem has communication cost C and information cost I.

• CPJ(C,I) is the “least compressible problem”. • If R(CPJ(C,I),1/3)=O(I), then R(F,1/3)=O(I(F,1/3))

for all F.

45

Page 46: Information and interactive computation

The big picture

R(F, ε) R(Fn,ε)/n

I(F, ε) I(Fn,ε)/n

direct sum for information

information = amortized

communicationdirect sum for

communication?

interactive compression?

Page 47: Information and interactive computation

Partial progress

• Can compress bounded-round interactive protocols.

• The main primitive is a one-shot version of Slepian-Wolf theorem.

• Alice gets a distribution PX. • Bob gets a prior distribution PY. • Goal: both must sample from PX.

47

Page 48: Information and interactive computation

Correlated sampling

48

A B

PX PY

M ~ PXM ~ PX

• The best we can hope for is D(PX||PY).

Uu Y

XXYX uP

uPuPPPD

)(

)(log)()||(

Page 49: Information and interactive computation

49

Proof Idea• Sample using D(PX||PY)+O(log 1/ε+D(PX||PY)½)

communication with statistical error ε.

PXPY

u1 u1

u2 u2

u3 u3

u4 u4

u4

~|U| samplesPublic randomness:

q1 q2 q3 q4 q5 q6 q7 ….u1 u2 u3 u4 u5 u6 u7

PX PY

1 1

0 0

Page 50: Information and interactive computation

Proof Idea• Sample using D(PX||PY)+O(log 1/ε+D(PX||PY)½)

communication with statistical error ε.

u4u2

h1(u4) h2(u4)

50PX

PYu4

PX PY

1 1

0 0u2

Page 51: Information and interactive computation

5151

Proof Idea• Sample using D(PX||PY)+O(log 1/ε+D(PX||PY)½)

communication with statistical error ε.

u4u2

h4(u4)… hlog 1/ ε(u4)

u4

h3(u4)

PX 2PY

PXPYu4

u4

h1(u4), h2(u4)

1 1

0 0

Page 52: Information and interactive computation

52

Analysis

• If PX(u4)≈2k PY(u4), then the protocol will reach round k of doubling.

• There will be ≈2k candidates.• About k+log 1/ε hashes.• The contribution of u4 to cost:

– PX(u4) (log PX (u4)/PY (u4) + log 1/ε).

Uu Y

XXYX uP

uPuPPPD

)(

)(log)()||(

Done!

Page 53: Information and interactive computation

Directions• Can interactive communication be fully

compressed? R(F, ε) = I(F, ε)?• What is the relationship between I(F, ε),

Iext(F, ε) and R(F, ε)?• Many other questions on interactive coding

theory!

53

Page 54: Information and interactive computation

54

Thank You