Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. ·...

92
Quantum query complexity of entropy estimation Xiaodi Wu QuICS, University of Maryland MSR Redmond, July 19th, 2017 Joint Center for QuantumInformation and Computer Science

Transcript of Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. ·...

Page 1: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum query complexity of entropyestimation

Xiaodi Wu

QuICS, University of Maryland

MSR Redmond, July 19th, 2017

J o i n t C e n t e r f o rQuantum Information and Computer Science

Page 2: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Outline

Motivation and Problem Statements

Main Results

Techniques

Open Questions and On-going Work

Page 3: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Outline

Motivation and Problem Statements

Main Results

Techniques

Open Questions and On-going Work

Page 4: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Motivation

Why do I come across this problem?

I Quantum property testing [e.g., survey [Montanaro, deWolf]] (Quantum testers of classical properties !)

I Testing of distributional properties. A well-motivatedbranch in the classical property testing literature !! alsoconnected to learning problems.

I Objects are distributions, rather than boolean functions.Might bring new insights or call for new techniques!

Page 5: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Motivation

Why do I come across this problem?

I Quantum property testing [e.g., survey [Montanaro, deWolf]] (Quantum testers of classical properties !)

I Testing of distributional properties. A well-motivatedbranch in the classical property testing literature !! alsoconnected to learning problems.

I Objects are distributions, rather than boolean functions.Might bring new insights or call for new techniques!

Page 6: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Motivation

Why do I come across this problem?

I Quantum property testing [e.g., survey [Montanaro, deWolf]] (Quantum testers of classical properties !)

I Testing of distributional properties. A well-motivatedbranch in the classical property testing literature !! alsoconnected to learning problems.

I Objects are distributions, rather than boolean functions.Might bring new insights or call for new techniques!

Page 7: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Entropies

Given any distribution p over a discrete set X, the Shannonentropy of this distribution p is defined by

H(p) :=∑

x∈X:p(x)>0

−px log px.

One important generalization of Shannon entropy is the Renyientropy of order α, denoted Hα(p), which is defined by

Hα(p) =

1

1−α log∑

x∈X pαx , when α 6= 1.

H(p), when α = 1.

Page 8: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Entropies

Given any distribution p over a discrete set X, the Shannonentropy of this distribution p is defined by

H(p) :=∑

x∈X:p(x)>0

−px log px.

One important generalization of Shannon entropy is the Renyientropy of order α, denoted Hα(p), which is defined by

Hα(p) =

1

1−α log∑

x∈X pαx , when α 6= 1.

H(p), when α = 1.

Page 9: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Problem Statement

For convenience, assume X = [n].

A natural question: Given access to samples obtained by tak-ing independent draws from p, determine the necessary numberof independent draws to estimate H(p) or Hα(p) within errorε, with high probability.

Motivations: this is a theoretically appealing topic with intimateconnections to statistics, information theory, learning theory, andalgorithm design.

Page 10: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Problem Statement

For convenience, assume X = [n].

A natural question: Given access to samples obtained by tak-ing independent draws from p, determine the necessary numberof independent draws to estimate H(p) or Hα(p) within errorε, with high probability.

Motivations: this is a theoretically appealing topic with intimateconnections to statistics, information theory, learning theory, andalgorithm design.

Page 11: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Our Question

Main Question: is there any quantum speed-up ofestimation of Shannon and Renyi entropies ?

Our question aligns with the emerging topic called “quantumproperty testing” and focuses on investigating the quantumadvantage in testing classical statistical properties.

The first research paper on this topic is by Bravyi, Harrow, andHassidim [BHH 11], where they have discovered quantumspeed-ups of testing uniformity, orthogonality, and statisticaldifference on unknown distributions, followed by Chakrabortyet al. [CFMdW 10].

Page 12: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Our Question

Main Question: is there any quantum speed-up ofestimation of Shannon and Renyi entropies ?

Our question aligns with the emerging topic called “quantumproperty testing” and focuses on investigating the quantumadvantage in testing classical statistical properties.

The first research paper on this topic is by Bravyi, Harrow, andHassidim [BHH 11], where they have discovered quantumspeed-ups of testing uniformity, orthogonality, and statisticaldifference on unknown distributions, followed by Chakrabortyet al. [CFMdW 10].

Page 13: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Outline

Motivation and Problem Statements

Main Results

Techniques

Open Questions and On-going Work

Page 14: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical results

Classically, this fundamental question has been intensively stud-ied in both the communities of theoretical computer science andinformation theory.

If ε is a constant, then the classical query complexity

I for Shannon entropy [VV 11], [JVHW 15]: Θ( nlogn).

I for Renyi entropy [AOST 17]:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 15: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical results

Classically, this fundamental question has been intensively stud-ied in both the communities of theoretical computer science andinformation theory.

If ε is a constant, then the classical query complexity

I for Shannon entropy [VV 11], [JVHW 15]: Θ( nlogn).

I for Renyi entropy [AOST 17]:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 16: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical results

Classically, this fundamental question has been intensively stud-ied in both the communities of theoretical computer science andinformation theory.

If ε is a constant, then the classical query complexity

I for Shannon entropy [VV 11], [JVHW 15]: Θ( nlogn).

I for Renyi entropy [AOST 17]:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 17: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical results

Classically, this fundamental question has been intensively stud-ied in both the communities of theoretical computer science andinformation theory.

If ε is a constant, then the classical query complexity

I for Shannon entropy [VV 11], [JVHW 15]: Θ( nlogn).

I for Renyi entropy [AOST 17]:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 18: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum results

Our results:

α classical bounds quantum bounds (this talk)

0 < α < 1 O( n1α

logn), Ω(n1α−o(1)) [AOST 17] O(n

1α− 1

2 ), Ω(maxn17α−o(1), n

13 )

α = 1 Θ( nlogn) [VV 11, JVHW 15] O(

√n), Ω(n

13 )

α > 1, α /∈ N O( nlogn), Ω(n1−o(1)) [AOST 17] O(n1−

12α

), Ω(maxn

13 ,Ω(n

12− 1

2α ))α = 2 Θ(

√n)[AOST 17] Θ(n

13 )

α > 2, α ∈ N Θ(n1−1/α) [AOST 17] O(nν(1−1/α)), Ω(n12− 1

2α ), ν < 3/4

α =∞ Θ( nlogn) [VV 11] O(Q(dlog ne-distinctness)), Ω(

√n)

Table 1: Summary of classical and quantum query complexity ofHα(p) for α > 0, assuming ε = Θ(1).

Page 19: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum results

Quantum lower bounds

Quantum upper bounds

Classical tight bounds

0 1 2 3 4 5α0.0

0.5

1.0

1.5

2.0

exp of n

Figure 1: Visualization of classical and quantum query complexity ofHα(p). The x-axis represents α and the y-axis represents theexponent of n. Red curves and points represent quantum upperbounds. Green curves and points represent classical tight bounds.The Magenta curve represents quantum lower bounds.

Page 20: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum results: in my April talk at MSR

Quantum lower bounds

Quantum upper bounds

Classical tight bounds

0 1 2 3 4 5α0.0

0.5

1.0

1.5

2.0

deg n

Figure 2: Visualization of classical and quantum query complexity ofHα(p). The x-axis represents α and the y-axis represents theexponent of n. Red curves and points represent quantum upperbounds. Green curves and points represent classical tight bounds.The Magenta curve represents quantum lower bounds.

Page 21: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Outline

Motivation and Problem Statements

Main Results

Techniques

Open Questions and On-going Work

Page 22: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Sample vs Query model

Ref. [BHH 11] models any discrete distribution p = (pi)ni=1 on

[n] by an oracle Op : [S] → [n]. Any probability pi (i ∈ [n]) isthus proportional to the size of pre-image of i under Op:

pi =|s ∈ [S] : Op(s) = i|

S∀ i ∈ [n].

If one samples s uniformly from [S] and outputs Op(s), then oneobtains a sample drawn from distribution p.

It is shown in [BHH 11] that the query complexity of the oraclemodel above and the sample complexity of independent samplesare in fact equivalent classically.

Page 23: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Sample vs Query model

Ref. [BHH 11] models any discrete distribution p = (pi)ni=1 on

[n] by an oracle Op : [S] → [n]. Any probability pi (i ∈ [n]) isthus proportional to the size of pre-image of i under Op:

pi =|s ∈ [S] : Op(s) = i|

S∀ i ∈ [n].

If one samples s uniformly from [S] and outputs Op(s), then oneobtains a sample drawn from distribution p.

It is shown in [BHH 11] that the query complexity of the oraclemodel above and the sample complexity of independent samplesare in fact equivalent classically.

Page 24: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Sample vs Query model

Ref. [BHH 11] models any discrete distribution p = (pi)ni=1 on

[n] by an oracle Op : [S] → [n]. Any probability pi (i ∈ [n]) isthus proportional to the size of pre-image of i under Op:

pi =|s ∈ [S] : Op(s) = i|

S∀ i ∈ [n].

If one samples s uniformly from [S] and outputs Op(s), then oneobtains a sample drawn from distribution p.

It is shown in [BHH 11] that the query complexity of the oraclemodel above and the sample complexity of independent samplesare in fact equivalent classically.

Page 25: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum Query Model

Quantumly, Op is transformed into a unitary operator Op actingon CS ⊗ Cn+1 such that

Op|s〉|0〉 = |s〉|Op(s)〉 ∀ s ∈ [S].

This is more powerful than Op because we may take a superpo-sition of states as an input.

Page 26: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Roadmap of quantum entropy estimation: for all α > 0

Amplitude

amplification

Quantum speedup of

Chebyshev’s inequality

Quantum

counting

Chebyshev

cooling

Simulated

annealing

Phase

estimation

Learning graph

for k-distinctness

Span programs

α-Renyi entropyestimation (α /∈ N)

Shannon entropyestimation

Min-entropyestimation

α-Renyi entropyestimation (α ∈ N)

Page 27: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

A general distribution property estimation problem:

Given a discrete distribution p = (pi)ni=1 on [n] and a function

f : (0, 1]→ R, estimate F (p) :=∑

i∈[n] pif(pi) with smalladditive or multiplicative error with high success probability.

If f(x) = − log x, F (p) is the Shannon entropy H(p) (additiveerror); if f(x) = xα−1 for some α > 0, α 6= 1, Hα(p) ∝ logF (p)(multiplicative error).

Inspired by BHH, we formulate a framework for approximatingF (p).

Page 28: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

A general distribution property estimation problem:

Given a discrete distribution p = (pi)ni=1 on [n] and a function

f : (0, 1]→ R, estimate F (p) :=∑

i∈[n] pif(pi) with smalladditive or multiplicative error with high success probability.

If f(x) = − log x, F (p) is the Shannon entropy H(p) (additiveerror); if f(x) = xα−1 for some α > 0, α 6= 1, Hα(p) ∝ logF (p)(multiplicative error).

Inspired by BHH, we formulate a framework for approximatingF (p).

Page 29: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

A general distribution property estimation problem:

Given a discrete distribution p = (pi)ni=1 on [n] and a function

f : (0, 1]→ R, estimate F (p) :=∑

i∈[n] pif(pi) with smalladditive or multiplicative error with high success probability.

If f(x) = − log x, F (p) is the Shannon entropy H(p) (additiveerror); if f(x) = xα−1 for some α > 0, α 6= 1, Hα(p) ∝ logF (p)(multiplicative error).

Inspired by BHH, we formulate a framework for approximatingF (p).

Page 30: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

Algorithm: Estimate F (p) =∑

i pif(pi).

1 Set l,M ∈ N;2 Regard the following subroutine as A:3 Draw a sample i ∈ [n] according to p;4 Use amplitude estimation with M queries to obtain an estimation

pi of pi;5 Output X = f(pi);

6 Use A for l times in “quantum speedup of Chebyshev’s inequality” and

outputs an estimation F (p) of F (p);

Intuitively, performance depends on E(F (p)) and σ(F (p)).

Want: E(F (p)) close to F (p) and σ(F (p)) small.

Pros: intuitively correct; (tedious) calculation only !?

Cons: very limited quantum speedup with only the basic framework.

Page 31: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

Algorithm: Estimate F (p) =∑

i pif(pi).

1 Set l,M ∈ N;2 Regard the following subroutine as A:3 Draw a sample i ∈ [n] according to p;4 Use amplitude estimation with M queries to obtain an estimation

pi of pi;5 Output X = f(pi);

6 Use A for l times in “quantum speedup of Chebyshev’s inequality” and

outputs an estimation F (p) of F (p);

Intuitively, performance depends on E(F (p)) and σ(F (p)).

Want: E(F (p)) close to F (p) and σ(F (p)) small.

Pros: intuitively correct; (tedious) calculation only !?

Cons: very limited quantum speedup with only the basic framework.

Page 32: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

Algorithm: Estimate F (p) =∑

i pif(pi).

1 Set l,M ∈ N;2 Regard the following subroutine as A:3 Draw a sample i ∈ [n] according to p;4 Use amplitude estimation with M queries to obtain an estimation

pi of pi;5 Output X = f(pi);

6 Use A for l times in “quantum speedup of Chebyshev’s inequality” and

outputs an estimation F (p) of F (p);

Intuitively, performance depends on E(F (p)) and σ(F (p)).

Want: E(F (p)) close to F (p) and σ(F (p)) small.

Pros: intuitively correct; (tedious) calculation only !?

Cons: very limited quantum speedup with only the basic framework.

Page 33: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

Algorithm: Estimate F (p) =∑

i pif(pi).

1 Set l,M ∈ N;2 Regard the following subroutine as A:3 Draw a sample i ∈ [n] according to p;4 Use amplitude estimation with M queries to obtain an estimation

pi of pi;5 Output X = f(pi);

6 Use A for l times in “quantum speedup of Chebyshev’s inequality” and

outputs an estimation F (p) of F (p);

Intuitively, performance depends on E(F (p)) and σ(F (p)).

Want: E(F (p)) close to F (p) and σ(F (p)) small.

Pros: intuitively correct; (tedious) calculation only !?

Cons: very limited quantum speedup with only the basic framework.

Page 34: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

High-level Framework

Algorithm: Estimate F (p) =∑

i pif(pi).

1 Set l,M ∈ N;2 Regard the following subroutine as A:3 Draw a sample i ∈ [n] according to p;4 Use amplitude estimation with M queries to obtain an estimation

pi of pi;5 Output X = f(pi);

6 Use A for l times in “quantum speedup of Chebyshev’s inequality” and

outputs an estimation F (p) of F (p);

Intuitively, performance depends on E(F (p)) and σ(F (p)).

Want: E(F (p)) close to F (p) and σ(F (p)) small.

Pros: intuitively correct; (tedious) calculation only !?

Cons: very limited quantum speedup with only the basic framework.

Page 35: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Roadmap of quantum entropy estimation: for all α > 0

Amplitude

amplification

Quantum speedup of

Chebyshev’s inequality

Quantum

counting

Chebyshev

cooling

Simulated

annealing

Phase

estimation

Learning graph

for k-distinctness

Span programs

α-Renyi entropyestimation (α /∈ N)

Shannon entropyestimation

Min-entropyestimation

α-Renyi entropyestimation (α ∈ N)

Page 36: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for Shannon entropy estimation

Two new ingredients (c.f. BHH) when f(p) = − log p

I Montanaro’s quantum speedup of Monte Carlo methods.Let A be a quantum algorithm outputting X such thatVar(X) ≤ σ2. For ε s.t. 0 < ε < 4σ, by using O(σ/ε) timesof A and A−1, one can output estimate E(X) of E(X) s.t.

Pr[|E(X)− E(X)| ≥ ε

]≤ 1/4.

Classically, one needs to use Θ(σ2/ε2) times of A.

I a fine-tuned analysis to bound the value of log(1/p) whenp→ 0. Analyze the full distribution of amplitudeamplification.

Complexity: Classical Θ(n/ log(n)) vs Quantum O(√n).

Page 37: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for Shannon entropy estimation

Two new ingredients (c.f. BHH) when f(p) = − log p

I Montanaro’s quantum speedup of Monte Carlo methods.Let A be a quantum algorithm outputting X such thatVar(X) ≤ σ2. For ε s.t. 0 < ε < 4σ, by using O(σ/ε) timesof A and A−1, one can output estimate E(X) of E(X) s.t.

Pr[|E(X)− E(X)| ≥ ε

]≤ 1/4.

Classically, one needs to use Θ(σ2/ε2) times of A.

I a fine-tuned analysis to bound the value of log(1/p) whenp→ 0. Analyze the full distribution of amplitudeamplification.

Complexity: Classical Θ(n/ log(n)) vs Quantum O(√n).

Page 38: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for Shannon entropy estimation

Two new ingredients (c.f. BHH) when f(p) = − log p

I Montanaro’s quantum speedup of Monte Carlo methods.Let A be a quantum algorithm outputting X such thatVar(X) ≤ σ2. For ε s.t. 0 < ε < 4σ, by using O(σ/ε) timesof A and A−1, one can output estimate E(X) of E(X) s.t.

Pr[|E(X)− E(X)| ≥ ε

]≤ 1/4.

Classically, one needs to use Θ(σ2/ε2) times of A.

I a fine-tuned analysis to bound the value of log(1/p) whenp→ 0. Analyze the full distribution of amplitudeamplification.

Complexity: Classical Θ(n/ log(n)) vs Quantum O(√n).

Page 39: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for Shannon entropy estimation

Two new ingredients (c.f. BHH) when f(p) = − log p

I Montanaro’s quantum speedup of Monte Carlo methods.Let A be a quantum algorithm outputting X such thatVar(X) ≤ σ2. For ε s.t. 0 < ε < 4σ, by using O(σ/ε) timesof A and A−1, one can output estimate E(X) of E(X) s.t.

Pr[|E(X)− E(X)| ≥ ε

]≤ 1/4.

Classically, one needs to use Θ(σ2/ε2) times of A.

I a fine-tuned analysis to bound the value of log(1/p) whenp→ 0. Analyze the full distribution of amplitudeamplification.

Complexity: Classical Θ(n/ log(n)) vs Quantum O(√n).

Page 40: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Roadmap of quantum entropy estimation: for all α > 0

Amplitude

amplification

Quantum speedup of

Chebyshev’s inequality

Quantum

counting

Chebyshev

cooling

Simulated

annealing

Phase

estimation

Learning graph

for k-distinctness

Span programs

α-Renyi entropyestimation (α /∈ N)

Shannon entropyestimation

Min-entropyestimation

α-Renyi entropyestimation (α ∈ N)

Page 41: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α /∈ N

Issue w/ the basic framework for Renyi entropy

multiplicative errors (Montanaro’s algorithm yields no speedupin the worst case) → quantum advantage when 1/2 < α < 2.

New observations

I given E[X] ∈ [a, b] s.t. b = O(a), we develop a variant ofMontanaro’s algorithm with a quadratic speedup.

I Let Pα(p) =∑

i pαi , α1, α2 > 0 s.t. α1/α2 = 1± 1/ log(n),

Pα1(p) = Θ(Pα2(p)α1/α2)

Use Pα2 to estimate [a, b] for Pα1 where α2 is closer to 1.

I Recursively solve the Pα2 case until α2 ≈ 1 where aspeed-up is already known.

Page 42: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α /∈ N

Issue w/ the basic framework for Renyi entropy

multiplicative errors (Montanaro’s algorithm yields no speedupin the worst case) → quantum advantage when 1/2 < α < 2.

New observations

I given E[X] ∈ [a, b] s.t. b = O(a), we develop a variant ofMontanaro’s algorithm with a quadratic speedup.

I Let Pα(p) =∑

i pαi , α1, α2 > 0 s.t. α1/α2 = 1± 1/ log(n),

Pα1(p) = Θ(Pα2(p)α1/α2)

Use Pα2 to estimate [a, b] for Pα1 where α2 is closer to 1.

I Recursively solve the Pα2 case until α2 ≈ 1 where aspeed-up is already known.

Page 43: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α /∈ N

Issue w/ the basic framework for Renyi entropy

multiplicative errors (Montanaro’s algorithm yields no speedupin the worst case) → quantum advantage when 1/2 < α < 2.

New observations

I given E[X] ∈ [a, b] s.t. b = O(a), we develop a variant ofMontanaro’s algorithm with a quadratic speedup.

I Let Pα(p) =∑

i pαi , α1, α2 > 0 s.t. α1/α2 = 1± 1/ log(n),

Pα1(p) = Θ(Pα2(p)α1/α2)

Use Pα2 to estimate [a, b] for Pα1 where α2 is closer to 1.

I Recursively solve the Pα2 case until α2 ≈ 1 where aspeed-up is already known.

Page 44: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α /∈ N

Issue w/ the basic framework for Renyi entropy

multiplicative errors (Montanaro’s algorithm yields no speedupin the worst case) → quantum advantage when 1/2 < α < 2.

New observations

I given E[X] ∈ [a, b] s.t. b = O(a), we develop a variant ofMontanaro’s algorithm with a quadratic speedup.

I Let Pα(p) =∑

i pαi , α1, α2 > 0 s.t. α1/α2 = 1± 1/ log(n),

Pα1(p) = Θ(Pα2(p)α1/α2)

Use Pα2 to estimate [a, b] for Pα1 where α2 is closer to 1.

I Recursively solve the Pα2 case until α2 ≈ 1 where aspeed-up is already known.

Page 45: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α /∈ N

Recursively call roughly O(log(n)) times until α ≈ 1

I α > 1: α→ α(1 + 1logn)−1 → α(1 + 1

logn)−2 · · · ;I 0 < α < 1: α→ α(1− 1

logn)−1 → α(1− 1logn)−2 · · · .

Similarity and difference: cooling schedules in simulatedannealing, volume estimation

I multi-section, multiplicative factors, similar designprinciple.

I adapt this design principle to our context.

Page 46: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α /∈ N

Recursively call roughly O(log(n)) times until α ≈ 1

I α > 1: α→ α(1 + 1logn)−1 → α(1 + 1

logn)−2 · · · ;I 0 < α < 1: α→ α(1− 1

logn)−1 → α(1− 1logn)−2 · · · .

Similarity and difference: cooling schedules in simulatedannealing, volume estimation

I multi-section, multiplicative factors, similar designprinciple.

I adapt this design principle to our context.

Page 47: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Roadmap of quantum entropy estimation: for all α > 0

Amplitude

amplification

Quantum speedup of

Chebyshev’s inequality

Quantum

counting

Chebyshev

cooling

Simulated

annealing

Phase

estimation

Learning graph

for k-distinctness

Span programs

α-Renyi entropyestimation (α /∈ N)

Shannon entropyestimation

Min-entropyestimation

α-Renyi entropyestimation (α ∈ N)

Page 48: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

In the sample model, this problem has a good empiricalestimator based on α-th frequency moment.

Frequency moments

A sequence a1, . . . , aM ∈ [n] are given with the occurrences of1, . . . , n to be m1, . . . ,mn respectively. You are asked to give agood approximation of Fk =

∑i∈[n]m

ki for k ∈ N.

Empirical Estimation

Key observation:∑i∈[n]

pαi ≈∑i∈[n]

(mi

M)k = Fk/M

k

Note: this is not the best classical estimator. Why?

Page 49: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

In the sample model, this problem has a good empiricalestimator based on α-th frequency moment.

Frequency moments

A sequence a1, . . . , aM ∈ [n] are given with the occurrences of1, . . . , n to be m1, . . . ,mn respectively. You are asked to give agood approximation of Fk =

∑i∈[n]m

ki for k ∈ N.

Empirical Estimation

Key observation:∑i∈[n]

pαi ≈∑i∈[n]

(mi

M)k = Fk/M

k

Note: this is not the best classical estimator. Why?

Page 50: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

In the sample model, this problem has a good empiricalestimator based on α-th frequency moment.

Frequency moments

A sequence a1, . . . , aM ∈ [n] are given with the occurrences of1, . . . , n to be m1, . . . ,mn respectively. You are asked to give agood approximation of Fk =

∑i∈[n]m

ki for k ∈ N.

Empirical Estimation

Key observation:∑i∈[n]

pαi ≈∑i∈[n]

(mi

M)k = Fk/M

k

Note: this is not the best classical estimator. Why?

Page 51: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

In the sample model, this problem has a good empiricalestimator based on α-th frequency moment.

Frequency moments

A sequence a1, . . . , aM ∈ [n] are given with the occurrences of1, . . . , n to be m1, . . . ,mn respectively. You are asked to give agood approximation of Fk =

∑i∈[n]m

ki for k ∈ N.

Empirical Estimation

Key observation:∑i∈[n]

pαi ≈∑i∈[n]

(mi

M)k = Fk/M

k

Note: this is not the best classical estimator. Why?

Page 52: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

Want to use this empirical estimator in the query model !!

Issues and Solutions

I Issue 1: How to generate M samples? cannot query Mtimes (the same complexity as classical).

Key observation: treat our quantum oracle Op as asequence of S samples. The α-frequency moment ofOp(1), . . . , Op(S) is exactly Sα

∑i∈[n] p

αi .

I Issue 2: quantum algorithm for α-frequency moment withquery complexity (e.g., o(n

34(1− 1

α)) [Montanaro 16]) where

the ”n” is actually ”S” in this context.Key observation: Roughly replace S by αn in thealgorithm and redo the analysis.

Page 53: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

Want to use this empirical estimator in the query model !!

Issues and Solutions

I Issue 1: How to generate M samples? cannot query Mtimes (the same complexity as classical).Key observation: treat our quantum oracle Op as asequence of S samples. The α-frequency moment ofOp(1), . . . , Op(S) is exactly Sα

∑i∈[n] p

αi .

I Issue 2: quantum algorithm for α-frequency moment withquery complexity (e.g., o(n

34(1− 1

α)) [Montanaro 16]) where

the ”n” is actually ”S” in this context.Key observation: Roughly replace S by αn in thealgorithm and redo the analysis.

Page 54: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

Want to use this empirical estimator in the query model !!

Issues and Solutions

I Issue 1: How to generate M samples? cannot query Mtimes (the same complexity as classical).Key observation: treat our quantum oracle Op as asequence of S samples. The α-frequency moment ofOp(1), . . . , Op(S) is exactly Sα

∑i∈[n] p

αi .

I Issue 2: quantum algorithm for α-frequency moment withquery complexity (e.g., o(n

34(1− 1

α)) [Montanaro 16]) where

the ”n” is actually ”S” in this context.

Key observation: Roughly replace S by αn in thealgorithm and redo the analysis.

Page 55: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum algorithm for α-Renyi entropy: α ≥ 2 ∈ N

Want to use this empirical estimator in the query model !!

Issues and Solutions

I Issue 1: How to generate M samples? cannot query Mtimes (the same complexity as classical).Key observation: treat our quantum oracle Op as asequence of S samples. The α-frequency moment ofOp(1), . . . , Op(S) is exactly Sα

∑i∈[n] p

αi .

I Issue 2: quantum algorithm for α-frequency moment withquery complexity (e.g., o(n

34(1− 1

α)) [Montanaro 16]) where

the ”n” is actually ”S” in this context.Key observation: Roughly replace S by αn in thealgorithm and redo the analysis.

Page 56: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Roadmap of quantum entropy estimation: for all α > 0

Amplitude

amplification

Quantum speedup of

Chebyshev’s inequality

Quantum

counting

Chebyshev

cooling

Simulated

annealing

Phase

estimation

Learning graph

for k-distinctness

Span programs

α-Renyi entropyestimation (α /∈ N)

Shannon entropyestimation

Min-entropyestimation

α-Renyi entropyestimation (α ∈ N)

Page 57: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

The min-entropy (i.e., α = +∞) case: find the maxi pi

How about using integer α algorithm?

Intuitively, exists some α s.t., Hα(p) (sub-linear for any α) is agood enough approximation of Hmin(p).

NO, O(·) hides an exponential dependence on α.

Another reduction to log(n)-distinctness

I ”Poissonized sampling”: the numbers of occurrences of ithpoint are pair-wise independent Poisson distributions,parameterized by a guess value λ and pi, i ∈ [n].

I log(n) is a natural cut-off threshold for Poissondistribution.

I When λ ·maxi pi passes this threshold, a log(n)-collisioncan be found w.h.p.. Otherwise, update λ and try again.

Page 58: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

The min-entropy (i.e., α = +∞) case: find the maxi pi

How about using integer α algorithm?

Intuitively, exists some α s.t., Hα(p) (sub-linear for any α) is agood enough approximation of Hmin(p).

NO, O(·) hides an exponential dependence on α.

Another reduction to log(n)-distinctness

I ”Poissonized sampling”: the numbers of occurrences of ithpoint are pair-wise independent Poisson distributions,parameterized by a guess value λ and pi, i ∈ [n].

I log(n) is a natural cut-off threshold for Poissondistribution.

I When λ ·maxi pi passes this threshold, a log(n)-collisioncan be found w.h.p.. Otherwise, update λ and try again.

Page 59: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

The min-entropy (i.e., α = +∞) case: find the maxi pi

How about using integer α algorithm?

Intuitively, exists some α s.t., Hα(p) (sub-linear for any α) is agood enough approximation of Hmin(p).

NO, O(·) hides an exponential dependence on α.

Another reduction to log(n)-distinctness

I ”Poissonized sampling”: the numbers of occurrences of ithpoint are pair-wise independent Poisson distributions,parameterized by a guess value λ and pi, i ∈ [n].

I log(n) is a natural cut-off threshold for Poissondistribution.

I When λ ·maxi pi passes this threshold, a log(n)-collisioncan be found w.h.p.. Otherwise, update λ and try again.

Page 60: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

The min-entropy (i.e., α = +∞) case: find the maxi pi

How about using integer α algorithm?

Intuitively, exists some α s.t., Hα(p) (sub-linear for any α) is agood enough approximation of Hmin(p).

NO, O(·) hides an exponential dependence on α.

Another reduction to log(n)-distinctness

I ”Poissonized sampling”: the numbers of occurrences of ithpoint are pair-wise independent Poisson distributions,parameterized by a guess value λ and pi, i ∈ [n].

I log(n) is a natural cut-off threshold for Poissondistribution.

I When λ ·maxi pi passes this threshold, a log(n)-collisioncan be found w.h.p.. Otherwise, update λ and try again.

Page 61: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

The min-entropy (i.e., α = +∞) case: find the maxi pi

How about using integer α algorithm?

Intuitively, exists some α s.t., Hα(p) (sub-linear for any α) is agood enough approximation of Hmin(p).

NO, O(·) hides an exponential dependence on α.

Another reduction to log(n)-distinctness

I ”Poissonized sampling”: the numbers of occurrences of ithpoint are pair-wise independent Poisson distributions,parameterized by a guess value λ and pi, i ∈ [n].

I log(n) is a natural cut-off threshold for Poissondistribution.

I When λ ·maxi pi passes this threshold, a log(n)-collisioncan be found w.h.p.. Otherwise, update λ and try again.

Page 62: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum lower bounds

Any quantum algorithm that approximates α-Renyi entropy ofa discrete distribution on [n] with success probability at least2/3 must use

I Ω(n17α−o(1)) quantum queries when 0 < α < 3

7 .

I Ω(n13 ) quantum queries when 3

7 ≤ α ≤ 3.

I Ω(n12− 1

2α ) quantum queries when α ≥ 3.

I Ω(√n) quantum queries when α = +∞.

Techniques:

I Reductions to the collision, Hamming weight, symmetryfunction problems.

I The polynomial method inspired by the collision lowerbound for a better error dependence.

Page 63: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum lower bounds

Any quantum algorithm that approximates α-Renyi entropy ofa discrete distribution on [n] with success probability at least2/3 must use

I Ω(n17α−o(1)) quantum queries when 0 < α < 3

7 .

I Ω(n13 ) quantum queries when 3

7 ≤ α ≤ 3.

I Ω(n12− 1

2α ) quantum queries when α ≥ 3.

I Ω(√n) quantum queries when α = +∞.

Techniques:

I Reductions to the collision, Hamming weight, symmetryfunction problems.

I The polynomial method inspired by the collision lowerbound for a better error dependence.

Page 64: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Quantum lower bounds

Any quantum algorithm that approximates α-Renyi entropy ofa discrete distribution on [n] with success probability at least2/3 must use

I Ω(n17α−o(1)) quantum queries when 0 < α < 3

7 .

I Ω(n13 ) quantum queries when 3

7 ≤ α ≤ 3.

I Ω(n12− 1

2α ) quantum queries when α ≥ 3.

I Ω(√n) quantum queries when α = +∞.

Techniques:

I Reductions to the collision, Hamming weight, symmetryfunction problems.

I The polynomial method inspired by the collision lowerbound for a better error dependence.

Page 65: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Outline

Motivation and Problem Statements

Main Results

Techniques

Open Questions and On-going Work

Page 66: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Questions

k-distinctness for super-constant k?

A problem interesting on its own. Our reduction to themin-entropy case just adds another motivation!

No quantum speed-up from existing algorithms by Ambainis andBelovs when k = log(n)

differences between quantum and classical?

At a high level, we want testers with correct expectations butsmall variances. Many techniques are proposed classically, andso different from ours.

Can we leverage the design principle of classical testers?

Page 67: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Questions

k-distinctness for super-constant k?

A problem interesting on its own. Our reduction to themin-entropy case just adds another motivation!

No quantum speed-up from existing algorithms by Ambainis andBelovs when k = log(n)

differences between quantum and classical?

At a high level, we want testers with correct expectations butsmall variances. Many techniques are proposed classically, andso different from ours.

Can we leverage the design principle of classical testers?

Page 68: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Questions

k-distinctness for super-constant k?

A problem interesting on its own. Our reduction to themin-entropy case just adds another motivation!

No quantum speed-up from existing algorithms by Ambainis andBelovs when k = log(n)

differences between quantum and classical?

At a high level, we want testers with correct expectations butsmall variances. Many techniques are proposed classically, andso different from ours.

Can we leverage the design principle of classical testers?

Page 69: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Questions

k-distinctness for super-constant k?

A problem interesting on its own. Our reduction to themin-entropy case just adds another motivation!

No quantum speed-up from existing algorithms by Ambainis andBelovs when k = log(n)

differences between quantum and classical?

At a high level, we want testers with correct expectations butsmall variances. Many techniques are proposed classically, andso different from ours.

Can we leverage the design principle of classical testers?

Page 70: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Thank You!!

Q & A

Page 71: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Shannon entropy

A first choice: empirical estimator. If we take M samples and oc-curences of 1, . . . , n are m1, . . . ,mn respectively, then empiricallythe distribution is (m1

M , . . . , mnM ), and

Hemp(p) = −∑i∈[n]

mi

Mlog

mi

M.

To approximate H(p) within error ε with high probability, needM = Θ( n

ε2).

Not that bad compared to the best estimator, whereM = Θ( nε2 logn

).

Page 72: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Shannon entropy

A first choice: empirical estimator. If we take M samples and oc-curences of 1, . . . , n are m1, . . . ,mn respectively, then empiricallythe distribution is (m1

M , . . . , mnM ), and

Hemp(p) = −∑i∈[n]

mi

Mlog

mi

M.

To approximate H(p) within error ε with high probability, needM = Θ( n

ε2).

Not that bad compared to the best estimator, whereM = Θ( nε2 logn

).

Page 73: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Shannon entropy

A first choice: empirical estimator. If we take M samples and oc-curences of 1, . . . , n are m1, . . . ,mn respectively, then empiricallythe distribution is (m1

M , . . . , mnM ), and

Hemp(p) = −∑i∈[n]

mi

Mlog

mi

M.

To approximate H(p) within error ε with high probability, needM = Θ( n

ε2).

Not that bad compared to the best estimator, whereM = Θ( nε2 logn

).

Page 74: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Shannon entropy

A first choice: empirical estimator. If we take M samples and oc-curences of 1, . . . , n are m1, . . . ,mn respectively, then empiricallythe distribution is (m1

M , . . . , mnM ), and

Hemp(p) = −∑i∈[n]

mi

Mlog

mi

M.

To approximate H(p) within error ε with high probability, needM = Θ( n

ε2).

Not that bad compared to the best estimator, whereM = Θ( nε2 logn

).

Page 75: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Shannon entropy

Construction of the best estimator:

I [VV 11]: A very clever (but complicated) application oflinear programming under Poissonized samples (willexplain later)

I [JVHW 15]: polynomial approximation

Page 76: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Shannon entropy

Construction of the best estimator:

I [VV 11]: A very clever (but complicated) application oflinear programming under Poissonized samples (willexplain later)

I [JVHW 15]: polynomial approximation

Page 77: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Shannon entropy

Construction of the best estimator:

I [VV 11]: A very clever (but complicated) application oflinear programming under Poissonized samples (willexplain later)

I [JVHW 15]: polynomial approximation

Page 78: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

Still, we can first consider the empirical estimator: Hα,emp(p) =1

1−α log∑

i∈[n](miM

)α.

In [AOST 17], it is shown that the query complexity of the em-pirical estimator is

Θ(n1α ), when 0 < α < 1.

Θ(n), when α > 1.

Recall that the classical query complexity of Renyi entropy:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 79: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

Still, we can first consider the empirical estimator: Hα,emp(p) =1

1−α log∑

i∈[n](miM

)α.

In [AOST 17], it is shown that the query complexity of the em-pirical estimator is

Θ(n1α ), when 0 < α < 1.

Θ(n), when α > 1.

Recall that the classical query complexity of Renyi entropy:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 80: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

Still, we can first consider the empirical estimator: Hα,emp(p) =1

1−α log∑

i∈[n](miM

)α.

In [AOST 17], it is shown that the query complexity of the em-pirical estimator is

Θ(n1α ), when 0 < α < 1.

Θ(n), when α > 1.

Recall that the classical query complexity of Renyi entropy:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 81: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

Still, we can first consider the empirical estimator: Hα,emp(p) =1

1−α log∑

i∈[n](miM

)α.

In [AOST 17], it is shown that the query complexity of the em-pirical estimator is

Θ(n1α ), when 0 < α < 1.

Θ(n), when α > 1.

Recall that the classical query complexity of Renyi entropy:O( n

logn) and Ω(n1α−o(1)), when 0 < α < 1.

O( nlogn) and Ω(n1−o(1)), when α > 1, α /∈ N.

Θ(n1−1α ), when α > 1, α ∈ N.

Page 82: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

Main difference happens only when α > 1, α ∈ N.

This is not a minor point. In particular, 2-Renyi entropy (alsoknown as the collision entropy) has classical query complexityΘ(√n), which is much less than that of Shannon entropy (Θ( n

logn)).

Collision entropy gives a rough estimation of Shannon entropy,and it measures the quality of random number generators andkey derivation in cryptographic applications.

Page 83: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

Main difference happens only when α > 1, α ∈ N.

This is not a minor point. In particular, 2-Renyi entropy (alsoknown as the collision entropy) has classical query complexityΘ(√n), which is much less than that of Shannon entropy (Θ( n

logn)).

Collision entropy gives a rough estimation of Shannon entropy,and it measures the quality of random number generators andkey derivation in cryptographic applications.

Page 84: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

Main difference happens only when α > 1, α ∈ N.

This is not a minor point. In particular, 2-Renyi entropy (alsoknown as the collision entropy) has classical query complexityΘ(√n), which is much less than that of Shannon entropy (Θ( n

logn)).

Collision entropy gives a rough estimation of Shannon entropy,and it measures the quality of random number generators andkey derivation in cryptographic applications.

Page 85: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

When α > 1 and α ∈ N, [AOST 17] proposes the following “bias-corrected” estimator:

Hα,bias(p) =1

1− αlog∑i∈[n]

mi(mi − 1) · · · (mi − α+ 1)

Mα.

This is actually the best estimator with query complexity Θ(n1−1α ).

The analysis is mainly based on a “Possionization” technique.

Page 86: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

Classical estimators for Renyi entropy

When α > 1 and α ∈ N, [AOST 17] proposes the following “bias-corrected” estimator:

Hα,bias(p) =1

1− αlog∑i∈[n]

mi(mi − 1) · · · (mi − α+ 1)

Mα.

This is actually the best estimator with query complexity Θ(n1−1α ).

The analysis is mainly based on a “Possionization” technique.

Page 87: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

An intuition of Possionization

When you are taking M samples where M is fixed, then theoccurences of 1, . . . , n are not independent. But if you take M ∼Poi(λ), then the occurence of i is mi ∼ Poi(λpi), and all mi arepairwise independent.

Moreover, with high probability λ/2 ≤ M ≤ 2λ, so Possioniza-tion does not influence query complexity up to a constant.

A key reason to use bias-corrected estimator: if a random variableX ∼ Poi(λ), then ∀ r ∈ N,

E[X(X − 1) · · · (X − r + 1)] = λr.

This property helps a lot for the analysis in [AOST 17] based onChernoff-Hoeffding inequality.

Page 88: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

An intuition of Possionization

When you are taking M samples where M is fixed, then theoccurences of 1, . . . , n are not independent. But if you take M ∼Poi(λ), then the occurence of i is mi ∼ Poi(λpi), and all mi arepairwise independent.

Moreover, with high probability λ/2 ≤ M ≤ 2λ, so Possioniza-tion does not influence query complexity up to a constant.

A key reason to use bias-corrected estimator: if a random variableX ∼ Poi(λ), then ∀ r ∈ N,

E[X(X − 1) · · · (X − r + 1)] = λr.

This property helps a lot for the analysis in [AOST 17] based onChernoff-Hoeffding inequality.

Page 89: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

An intuition of Possionization

When you are taking M samples where M is fixed, then theoccurences of 1, . . . , n are not independent. But if you take M ∼Poi(λ), then the occurence of i is mi ∼ Poi(λpi), and all mi arepairwise independent.

Moreover, with high probability λ/2 ≤ M ≤ 2λ, so Possioniza-tion does not influence query complexity up to a constant.

A key reason to use bias-corrected estimator: if a random variableX ∼ Poi(λ), then ∀ r ∈ N,

E[X(X − 1) · · · (X − r + 1)] = λr.

This property helps a lot for the analysis in [AOST 17] based onChernoff-Hoeffding inequality.

Page 90: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

An intuition of Possionization

When you are taking M samples where M is fixed, then theoccurences of 1, . . . , n are not independent. But if you take M ∼Poi(λ), then the occurence of i is mi ∼ Poi(λpi), and all mi arepairwise independent.

Moreover, with high probability λ/2 ≤ M ≤ 2λ, so Possioniza-tion does not influence query complexity up to a constant.

A key reason to use bias-corrected estimator: if a random variableX ∼ Poi(λ), then ∀ r ∈ N,

E[X(X − 1) · · · (X − r + 1)] = λr.

This property helps a lot for the analysis in [AOST 17] based onChernoff-Hoeffding inequality.

Page 91: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

An intuition of Possionization

When you are taking M samples where M is fixed, then theoccurences of 1, . . . , n are not independent. But if you take M ∼Poi(λ), then the occurence of i is mi ∼ Poi(λpi), and all mi arepairwise independent.

Moreover, with high probability λ/2 ≤ M ≤ 2λ, so Possioniza-tion does not influence query complexity up to a constant.

A key reason to use bias-corrected estimator: if a random variableX ∼ Poi(λ), then ∀ r ∈ N,

E[X(X − 1) · · · (X − r + 1)] = λr.

This property helps a lot for the analysis in [AOST 17] based onChernoff-Hoeffding inequality.

Page 92: Quantum query complexity of entropy estimationxwu/talks/MSR_Entropy.pdf · 2017. 7. 19. · estimation of Shannon and R enyi entropies ? Our question aligns with the emerging topic

ReferencesG. Valiant and P. Valiant.

Estimating the unseen: an n/log (n)-sample estimator for entropy andsupport size, shown optimal via new CLTs.

Proceedings of the forty-third annual ACM Symposium on Theory ofComputing, ACM, 2011.

J. Jiao, K. Venkat, Y. Han, and T. Weissman.

Minimax estimation of functionals of discrete distributions.

IEEE Transactions on Information Theory, 61.5 (2015): 2835-2885.

J. Acharya, A. Orlitsky, A. T. Suresh, and H. Tyagi.

Estimating Renyi entropy of discrete distributions.

IEEE Transactions on Information Theory, 63.1 (2017): 38-56.

S. Bravyi, A. W. Harrow, and A. Hassidim.

Quantum algorithms for testing properties of distributions..

IEEE Transactions on Information Theory, 57.6 (2011): 3971-3981.

A. Montanaro.

Quantum speedup of Monte Carlo methods.

Proc. R. Soc. A, Vol. 471. No. 2181. The Royal Society, 2015.