information complexity lower bounds

48
Information Complexity Lower Bounds Rotem Oshman, Princeton CCI Based on: Bar-Yossef,Jayram,Kumar,Srinivasan’04 Braverman,Barak,Chen,Rao’10

Upload: jael

Post on 19-Feb-2016

66 views

Category:

Documents


2 download

DESCRIPTION

Information Complexity Lower Bounds. Rotem Oshman, Princeton CCI Based on: Bar-Yossef,Jayram,Kumar,Srinivasan’04 Braverman,Barak,Chen,Rao’10. Communication Complexity. = ?. Yao ‘79, “Some complexity questions related to distributive computing ”. Communication Complexity. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Information  Complexity Lower Bounds

Information Complexity Lower Bounds

Rotem Oshman, Princeton CCIBased on:

Bar-Yossef,Jayram,Kumar,Srinivasan’04 Braverman,Barak,Chen,Rao’10

Page 2: Information  Complexity Lower Bounds

Communication Complexity

𝑋 𝑌

= ?

Yao ‘79, “Some complexity questions related to distributive computing”

Page 3: Information  Complexity Lower Bounds

• Applications:– Circuit complexity– Streaming algorithms– Data structures– Distributed computing– Property testing– …

Communication Complexity

Page 4: Information  Complexity Lower Bounds

Deterministic Protocols

• A protocol specifies, at each point:– Which player speaks next– What should the player say– When to halt and what to output

• Formally,

what we’ve said so far

who speaks next: Alice, Bob, = halt

what to say/output

Page 5: Information  Complexity Lower Bounds

Randomized Protocols

• Can use randomness to decide what to say– Private randomness: each player has a separate

source of random bits– Public randomness: both players can use the same

random bits• Goal: for any compute correctly with

probability • Communication complexity: worst-case length

of transcript in any execution

Page 6: Information  Complexity Lower Bounds

Randomness Can Help a Lot

• Example: EQUALITY– Input: – Output: is ?

• Trivial protocol: Alice sends to Bob• For deterministic protocols, this is optimal!

Page 7: Information  Complexity Lower Bounds

EQUALITY Lower Bound

0𝑛 1𝑛…0𝑛

1𝑛

111111111

#rectangles

Page 8: Information  Complexity Lower Bounds

Randomized Protocol

• Protocol with public randomness:– Select random – Alice sends – Bob accepts iff

• If : always accept• If :

• Reject with probability

non-zero vector

Page 9: Information  Complexity Lower Bounds

Set Disjointness

• Input: • Output: ?• Theorem [Kalyanasundaran, Schnitger ‘92,

Razborov ‘92]: randomized CC = – Easy to see for deterministic protocols

• Today we’ll see a proof by Bar-Yossef, Jayram, Kumar, Srinivasan ‘04

Page 10: Information  Complexity Lower Bounds

Application: Streaming Lower Bounds

• Streaming algorithm:

• Example: how many distinct items in the data?• Reduction from Disjointness [Alon, Matias,

Szegedy ’99]

data

algorithmHow much spaceis required to approximate f(data)?

Page 11: Information  Complexity Lower Bounds

Reduction from Disjointness:

• Fix a streaming algorithm for Distinct Elements with space , universe size

• Construct a protocol for Disj. with elements:

𝑋={𝑥1 ,…,𝑥𝑘 } 𝑌={𝑦 1 ,…, 𝑦 ℓ }

algorithmState of the algorithm and (#bits = )

⇔ #distinct elements in is

Page 12: Information  Complexity Lower Bounds

Application 2: KW Games

• Circuit depth lower bounds:

• How deep does the circuit need to be?

∧∨ ∧

∨∨∧ ∨𝑥1 … 𝑥𝑛

𝑓 (𝑥¿¿1 ,…,𝑥𝑛)¿

Page 13: Information  Complexity Lower Bounds

Application 2: KW Games

• Karchmer-Wigderson’93,Karchmer-Raz-Wigderson’94:

𝑋 : 𝑓 ( 𝑋 )=0 𝑌 : 𝑓 (𝑌 )=1

find such that

Page 14: Information  Complexity Lower Bounds

Application 2: KW Games

• Claim: if has deterministic CC , then requires circuit depth .

• Circuit with depth protocol with length

∧∨ ∧

∨∨∧ ∨𝑥1 … 𝑥𝑛𝑋 : 𝑓 ( 𝑋 )=0 𝑌 : 𝑓 (𝑌 )=1

0 1

1 10 1

Page 15: Information  Complexity Lower Bounds

Information-Theoretic Lower Bound on Set Disjointness

Page 16: Information  Complexity Lower Bounds

Some Basic Concepts from Info Theory

• Entropy of a random variable:

• Important properties:

– is deterministic– = expected # bits needed to encode

Page 17: Information  Complexity Lower Bounds

Some Basic Concepts from Info Theory

• Conditional entropy: • Important properties:

– are independent • Example:–– If then , if 1 then

Page 18: Information  Complexity Lower Bounds

Some Basic Concepts from Info Theory

• Mutual information:

• Conditional mutual information:

• Important properties:

– are independent

Page 19: Information  Complexity Lower Bounds

Some Basic Concepts from Info Theory

• Chain rule for mutual information:

• More generally,

Page 20: Information  Complexity Lower Bounds

Information Cost of Protocols

• Fix an input distribution on • Given a protocol , let also denote the

distribution of ’s transcript• Information cost of :

• Information cost of a function :

Page 21: Information  Complexity Lower Bounds

Information Cost of Protocols

• Important property: • Proof: by induction. Let .• : what we know after r rounds

what we knew after r-1 rounds

what we learn in round r, given what we already know

Page 22: Information  Complexity Lower Bounds

Information vs. Communication

• Want: • Suppose is sent by Alice.• What does Alice learn?– is a function of and so

• What does Bob learn?

Page 23: Information  Complexity Lower Bounds

Information vs. Communication

• Important property: • Lower bound on information cost ⇒ lower

bound on communication complexity• In fact, IC lower bounds are the most powerful

technique we know

Page 24: Information  Complexity Lower Bounds

Information Complexity of Disj.

• Disjointness: is ?• Strategy: for some “hard distribution” ,

1. Direct sum: 2. Prove that .

Page 25: Information  Complexity Lower Bounds

Hard Distribution for Disjointness

• For each coordinate :

𝑋 𝑖=0 𝑋 𝑖=1

𝑌 𝑖=0

𝑌 𝑖=1

1/3

1/3

1/3

0

Page 26: Information  Complexity Lower Bounds

𝑋

𝑌

𝐼 𝐶𝜇𝑛 (Disj )≥𝑛⋅ 𝐼𝐶𝜇 (¿)• Let be a protocol for on • Construct for as follows:– Alice and Bob get inputs – Choose a random coordinate , set – Sample and run – For each ,

𝑈

𝑉

Page 27: Information  Complexity Lower Bounds

𝐼 𝐶𝜇𝑛 (Disj )≥𝑛⋅ 𝐼𝐶𝜇 (¿)• Let be a protocol for on • Construct for as follows:– Alice and Bob get inputs – Choose a random coordinate , set – Bad idea: publicly sample

𝑈

𝑉

𝑋

𝑌Suppose in , Alice sends .

In , Bob learns one bit in he should learn bitBut if is public Bob learns 1 bit about !

Page 28: Information  Complexity Lower Bounds

𝐼 𝐶𝜇𝑛 (Disj )≥𝑛⋅ 𝐼𝐶𝜇 (¿)• Let be a protocol for on • Construct for as follows:– Alice and Bob get inputs – Choose a random coordinate , set – Another bad idea: publicly sample , Bob privately samples

given – But the players can’t sample , independently…

Page 29: Information  Complexity Lower Bounds

𝐼 𝐶𝜇𝑛 (Disj )≥𝑛⋅ 𝐼𝐶𝜇 (¿)• Let be a protocol for on • Construct for as follows:– Alice and Bob get inputs – Choose a random coordinate , set

𝑈

𝑉

Publicly sample

Publicly sample

Privately sample

Privately sample

𝑋

𝑌

Page 30: Information  Complexity Lower Bounds

Direct Sum Theorem• Transcript of • Need to show:

Page 31: Information  Complexity Lower Bounds

Information Complexity of Disj.

• Disjointness: is ?• Strategy: for some “hard distribution” ,

1. Direct sum: 2. Prove that .

Page 32: Information  Complexity Lower Bounds

Hardness of AND

1101

0 0 10

1/3

1/3

0

1/3

¿0

¿0

transcript on should be“very different”

Page 33: Information  Complexity Lower Bounds

Hellinger Distance

• Examples:

– If have disjoint support,

Page 34: Information  Complexity Lower Bounds

Hellinger Distance

• Hellinger distance is a metric– , with equality iff

– Triangle inequality:

𝑃𝑄

𝑅

Page 35: Information  Complexity Lower Bounds

Hellinger Distance

• If for some we have then

1101

0 0 10

h≥ 23√2

Page 36: Information  Complexity Lower Bounds

Hellinger Distance vs. Mutual Info

• Let be two distributions• Select by choosing , then drawing • Then

𝐼 (Π ;𝑌|𝑋=0 )≥h2 (Π 00 , Π 01 )

1101

0 0 10

1/3

1/3

0

1/3𝐼 (Π ; 𝑋|𝑌=0 )≥h2 (Π 00 , Π 10 )

Page 37: Information  Complexity Lower Bounds

Hardness of AND

1101

0 0 10

1/3

1/3

0

1/3

h≥ 23√2

Same for Alice untilBob acts differently

Same for Bob untilAlice acts differently

Page 38: Information  Complexity Lower Bounds

“Cut-n-Paste Lemma”

• Recall: • Enough to show: we can write

Page 39: Information  Complexity Lower Bounds

“Cut-n-Paste Lemma”

• We can write

• Proof:– induces a distribution on “partial transcripts” of

each length : probability that first bits are – By induction:

• Base case: – Set

Page 40: Information  Complexity Lower Bounds

“Cut-n-Paste Lemma”

• Step: • Suppose after it is Alice’s turn to speak• What Alice says depends on:– Her input– Her private randomness– The transcript so far,

• So • Set

Page 41: Information  Complexity Lower Bounds

Hardness of AND

1101

0 0 10

1/3

1/3

0

1/3

h≥ 23√2

Page 42: Information  Complexity Lower Bounds

Multi-Player Communication Complexity

Page 43: Information  Complexity Lower Bounds

The Coordinator Model

sites

𝑓 (𝑋 1,… ,𝑋𝑘 )=?

bits

𝑋 1 𝑋 2 𝑋𝑘…

Page 44: Information  Complexity Lower Bounds

Multi-Party Set Disjointness

• Input: • Output: is ?• Braverman,Ellen,O.,Pitassi,Vaikuntanathan’13:

lower bound of bits

Page 45: Information  Complexity Lower Bounds

Reduction from DISJ tograph connectivity

• Given we want to– Choose vertices – Design inputs such that is connected iff

Page 46: Information  Complexity Lower Bounds

Reduction from DISJ tograph connectivity

1234

56

𝑝1

𝑝2

𝑝𝑘

(Players)

𝑋 𝑖

[𝑛 ]∖⋃𝑋 𝑖

(Elements)

input graph connected

Page 47: Information  Complexity Lower Bounds

Other Stuff

• Distributed computing

Page 48: Information  Complexity Lower Bounds

Other Stuff

• Compressing down to information cost• Number-on-forehead lower bounds• Open questions in communication complexity