A BLEND OF INFORMATION THEORY AND STATISTICS - Yale...

52
Some Principles of Information Theory and Statistics Shannon Capacity Determines Limits of Statistical Accuracy Penalized Likelihood Analysis Regression with 1 Penalty Summary A BLEND OF INFORMATION THEORY AND STATISTICS Andrew R. Barron YALE UNIVERSITY Collaborators: Cong Huang, Jonathan Li, Gerald Cheang, Xi Luo Frejus, France, September 1-5, 2008

Transcript of A BLEND OF INFORMATION THEORY AND STATISTICS - Yale...

Page 1: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

A BLEND OF INFORMATION THEORY AND STATISTICS

Andrew R. Barron

YALE UNIVERSITY

Collaborators: Cong Huang, Jonathan Li, Gerald Cheang, Xi Luo

Frejus, France, September 1-5, 2008

Page 2: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

A BLEND OF INFORMATION THEORY AND STATISTICS

Andrew R. Barron

Collaborators: Cong Huang, Jonathan Li, Gerald Cheang, Xi Luo

Mon: Greedy Alg. for `1 Penalization and Subset Selection

Tue: Information Theory and Statistical Risk Analysis:Upper Bounds and Adaptation

Wed: Lower Bounds and Minimax Rates

Thu: Role of Approximation Theory

Page 3: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Information Theory & Statistical Risk Analysis

Andrew Barron

Barron Information Theory and Statistical Risk Analysis

Page 4: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Outline1 Some Principles of Information Theory and Statistics

Data CompressionMinimum Description Length Principle

2 Shannon Capacity Determines Limits of Statistical AccuracyThis part is saved for Wednesday

3 Penalized Likelihood AnalysisTwo-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

4 Regression with `1 PenaltyFixed σ2 CaseUnknown σ2 Case

5 Summary

Barron Information Theory and Statistical Risk Analysis

Page 5: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Outline1 Some Principles of Information Theory and Statistics

Data CompressionMinimum Description Length Principle

2 Shannon Capacity Determines Limits of Statistical AccuracyThis part is saved for Wednesday

3 Penalized Likelihood AnalysisTwo-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

4 Regression with `1 PenaltyFixed σ2 CaseUnknown σ2 Case

5 Summary

Barron Information Theory and Statistical Risk Analysis

Page 6: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Shannon, Kraft, McMillan

Characterization of uniquely decodeable codelengths

L(x), x ∈ X ,∑

x

2−L(x) ≤ 1

L(x) = log 1/q(x) q(x) = 2−L(x)

Operational meaning of probability:

A distribution is given by a choice of code

Barron Information Theory and Statistical Risk Analysis

Page 7: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Codelength Comparison

Targets p are possible distributions

Compare codelength log 1/q(x) to targets log 1/p(x)

Redundancy or regret[log 1/q(x)− log 1/p(x)

]

Expected redundancy

D(PX‖QX ) = EP

[log

p(X )

q(X )

]Barron Information Theory and Statistical Risk Analysis

Page 8: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Universal Codes

MODELSFamily of coding strategies ⇔ Family of prob. distributionsIndexed by parameters or functions:{

Lθ(x) : θ ∈ Θ}⇔

{pθ(x) : θ ∈ Θ

}{

Lf (x) : f ∈ F}⇔

{pf (x) : f ∈ F

}Universal codes ⇔ Universal probabilities q(x)

L(x) = log 1/q(x)

Redundancy:[

log 1/q(x)− log 1/pθ(x)]

Want it small either uniformly in x , θ or in expectation

Barron Information Theory and Statistical Risk Analysis

Page 9: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Statistical Aim

Training data x ⇒ estimator p = pθ

Subsequent data x ′

Want log 1/p(x ′) to compare favorably to log 1/p(x ′)

For targets p close to or in the families

Barron Information Theory and Statistical Risk Analysis

Page 10: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Loss

Kullback Information-divergence:

D(PX ′‖QX ′) = E[

log p(X ′)/q(X ′)]

Bhattacharyya, Hellinger, Chernoff, Rényi divergence:

d(PX ′ , QX ′) = 2 log 1/E [q(X ′)/p(X ′)]1/2

Product model case: p(x ′) =∏n

i=1 p(x ′i )

D(PX ′‖QX ′) = n D(P‖Q)

Likewise d(PX ′ , QX ′) = n d(P, Q)

Barron Information Theory and Statistical Risk Analysis

Page 11: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Loss

Relationship:d(P, Q) ≤ D(P‖Q)

and, if the log density ratio is not more than B, then

D(P‖Q) ≤ CB d(P, Q)

with CB ≤ 2 + B

Barron Information Theory and Statistical Risk Analysis

Page 12: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Minimum Description Length (MDL) Principle

Universal coding brought into statistical play

Minimum Description Length Principle:

The shortest code for data gives the best statistical model

Barron Information Theory and Statistical Risk Analysis

Page 13: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

MDL: Two-stage Version

Two-stage codelength:

L(x) = minθ∈Θ

[log 1/pθ(x) + L(θ)

]bits for x given θ + bits for θ

The corresponding statistical estimator is p = pθ

Typically in d-dimensional families L(θ) is of order

d2

log n

Barron Information Theory and Statistical Risk Analysis

Page 14: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

MDL: Mixture Versions

Codelength based on a Bayes mixture

L(x) = log1

q(x)

whereq(x) =

∫p(x |θ)w(θ)dθ

average case optimal and pointwise optimal for a.e. θ

Codelength approximation (Barron 1985, Clarke and Barron 1990,1994)

log1

p(x |θ)+

d2

logn

2π+ log

|I(θ)|1/2

w(θ)

where I(θ) is the empirical Fisher Information at the MLE

Barron Information Theory and Statistical Risk Analysis

Page 15: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

MDL: Mixture Versions

Codelength based on a Bayes mixture

L(x) = log1

q(x)

whereq(x) =

∫p(x |θ)w(θ)dθ

average case optimal and pointwise optimal for a.e. θ

Corresponding statistical estimator is

p(x ′) = q(x ′|x) =

∫p(x ′|θ)p(x |θ)w(θ)dθ∫

p(x |θ)w(θ)dθ

which is a predictive distribution

Barron Information Theory and Statistical Risk Analysis

Page 16: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Predictive Versions

Chain rule, for joint densities

q(x) = q(x1)q(x2|x1) · · ·q(xn|x1, x2, . . . , xn−1)

Chain rule, for codelengths based on predictive distrib

L(x) = log1

q(x1)+ log

1q(x2|x1)

+ . . . log1

q(xn|x1, . . . , xn−1)

Corresponding statistical estimator at x ′ = xn+1

p(x ′) = q(xn+1|x1, . . . , xn)

Barron Information Theory and Statistical Risk Analysis

Page 17: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Data CompressionMinimum Description Length Principle

Chain Rule of Divergence

Expected redundancy of compression of x by pred. distrib.corresponds to cumulative risk for sample sizes n′ < n

D(PX‖QX ) = E[

logp(x1)

q(x1)+log

p(x2)

q(x2|x1)+. . . log

p(xn)

q(xn|x1 . . . xn−1)

]

=n−1∑n′=0

E[D(PX‖PX |X1,...,Xn′

)]

Individual risk E [D(PX‖PX |X1,...,Xn)]

Cumulative risk of d2 log n corresponds to individual risk of

d2n

Barron Information Theory and Statistical Risk Analysis

Page 18: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Outline1 Some Principles of Information Theory and Statistics

Data CompressionMinimum Description Length Principle

2 Shannon Capacity Determines Limits of Statistical AccuracyThis part is saved for Wednesday

3 Penalized Likelihood AnalysisTwo-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

4 Regression with `1 PenaltyFixed σ2 CaseUnknown σ2 Case

5 Summary

Barron Information Theory and Statistical Risk Analysis

Page 19: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Information Capacity

A Channel θ → Y is a family of probability distributions

{PY |θ : θ ∈ Θ}

Information Capacity

C = maxPθ

I(θ; Y )

Barron Information Theory and Statistical Risk Analysis

Page 20: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Communications Capacity

Ccom = maximum rate of reliable communication,measured in message bits per use of a channel

Shannon Channel Capacity Theorem (Shannon 1948)

Ccom = C

Barron Information Theory and Statistical Risk Analysis

Page 21: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Data Compression Capacity

Minimax Redundancy

Red = minQY

maxθ∈Θ

D(PY |θ‖QY )

Data Compression Capacity Theorem

Red = C

(Gallager, Davisson & Leon-Garcia, Ryabko)

Barron Information Theory and Statistical Risk Analysis

Page 22: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Statistical Risk Setting

Loss function`(θ, θ′)

Examples:

Kullback Loss

`(θ, θ′) = D(PY |θ‖PY |θ′)

Squared metric loss, e.g. squared Hellinger loss:

`(θ, θ′) = d2(θ, θ′)

Barron Information Theory and Statistical Risk Analysis

Page 23: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Statistical Capacity

Estimators: θn

Based on sample Y of size n

Minimax Risk (Wald):

rn = minθn

maxθ

E`(θ, θn)

Barron Information Theory and Statistical Risk Analysis

Page 24: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Ingredients for determination of Statistical Capacity

Kolmogorov Metric Entropy of S ⊂ Θ:

H(ε) = max{log Card(Θε) : d(θ, θ′) > ε for θ, θ′ ∈ Θε ⊂ S}

Loss Assumption, for θ, θ′ ∈ S:

`(θ, θ′) ∼ D(PY |θ‖PY |θ′) ∼ d2(θ, θ′)

Barron Information Theory and Statistical Risk Analysis

Page 25: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

This part is saved for Wednesday

Statistical Capacity Theorem

For infinite-dimensional Θ

With metric entropy evaluated a critical separation εn

Statistical Capacity TheoremMinimax Risk ∼ Info Capacity Rate ∼ Metric Entropy rate

rn ∼ Cn

n∼ H(εn)

n∼ ε2

n

Yang 1997, Yang and B. 1999, Haussler and Opper 1997

Barron Information Theory and Statistical Risk Analysis

Page 26: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Outline1 Some Principles of Information Theory and Statistics

Data CompressionMinimum Description Length Principle

2 Shannon Capacity Determines Limits of Statistical AccuracyThis part is saved for Wednesday

3 Penalized Likelihood AnalysisTwo-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

4 Regression with `1 PenaltyFixed σ2 CaseUnknown σ2 Case

5 Summary

Barron Information Theory and Statistical Risk Analysis

Page 27: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Two-stage Code Redundancy

Expected codelength minus target at pθ∗

Redundancy = E[

minθ∈Θ

{log

1pθ(x)

+ L(θ)

}− log

1pθ∗(x)

]Redundancy approx in smooth families

d2

logn

2π+ log

|I(θ∗)|1/2

w(θ∗)

Barron Information Theory and Statistical Risk Analysis

Page 28: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Redundancy and Resolvability

Redundancy = E minθ∈Θ

[log pθ∗ (x)

pθ(x) + L(θ))]

Resolvability = minθ∈Θ E[

log pθ∗ (x)pθ(x) + L(θ)

]= minθ∈Θ

[D(PX |θ∗‖PX |θ) + L(θ)

]Ideal tradeoff of Kullback approximation error & complexity

Population analogue of the two-stage code MDL criterion

Divide by n to express as a rate. In the i.i.d. case

Rn(θ∗) = min

θ∈Θ

[D(θ∗‖θ) +

L(θ)

n

]

Barron Information Theory and Statistical Risk Analysis

Page 29: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Risk of Estimator based on Two-stage Code

Estimator θ is the choice achieving the minimization

minθ∈Θ

{log

1pθ(x)

+ L(θ)

}Codelengths for θ are L(θ) = 2L(θ) with

∑θ∈Θ 2−L(θ) ≤ 1.

Total loss dn(θ∗, θ) with dn(θ

∗, θ) = d(PX ′|θ∗ , PX ′|θ)

Risk = E [dn(θ∗, θ)]

Info-Thy bound on risk: ( Barron 1985, Barron and Cover 1991, Jonathan Li 1999)

Risk ≤ Redundancy ≤ Resolvability

Barron Information Theory and Statistical Risk Analysis

Page 30: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Risk of Estimator based on Two-stage Code

Estimator θ achieves minθ∈Θ {log 1/pθ(x) + L(θ)}Codelengths require

∑θ∈F 2−L(θ) ≤ 1.

Risk ≤ Resolvability

Specialize to i.i.d. case:

Ed(θ∗, θ) ≤ minθ∈Θ

[D(θ∗‖θ) +

L(θ)

n

]As n ↗, tolerate more complex PX |θ if needed to get nearPX |θ∗

Rate is 1/n, or close to that rate if the target is simple

Drawback: Code interpretation entails countable Θ

Barron Information Theory and Statistical Risk Analysis

Page 31: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Key to Risk Analysis

log likelihood-ratio discrepancy at training x and future x ′[log

pθ∗(x)

pθ(x)− dn(θ

∗, θ)]

Proof shows, for L(θ) satisfying Kraft, that

minθ∈Θ

{[log

pθ∗(x)

pθ(x)− dn(θ

∗, θ)]

+ L(θ)

}has expectation ≥ 0. From which the risk bound follows.

Barron Information Theory and Statistical Risk Analysis

Page 32: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Information-theoretically Valid Penalties

Penalized Likelihood

minθ∈Θ

{log

1pθ(x)

+ Pen(θ)

}Possibly uncountable Θ

Yields data compression interpretation if there exists acountable Θ and L satisfying Kraft such that the above isnot less than

minθ∈Θ

{log

1pθ(x)

+ L(θ)

}

Barron Information Theory and Statistical Risk Analysis

Page 33: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Data-Compression Valid Penalties

Equivalently, Pen(θ) is valid for penalized likelihood withuncountable Θ to have a data compression interpretationif there is such a countable Θ and Kraft summable L(θ),such that, for every θ in Θ, there is a representor θ in Θsuch that

Pen(θ) ≥ L(θ) + logpθ(x)

pθ(x)

This is the link between uncountable and countable cases

Barron Information Theory and Statistical Risk Analysis

Page 34: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Statistical-Risk Valid Penalties

Penalized Likelihood

θ = argminθ∈Θ

{log

1pθ(x)

+ Pen(θ)

}Again: possibly uncountable Θ

Task: determine a condition on Pen(θ) such that the risk iscaptured by the population analogue

Edn(θ∗, θ) ≤ inf

θ∈Θ

{E log

pθ∗(X )

pθ(X )+ Pen(θ)

}

Barron Information Theory and Statistical Risk Analysis

Page 35: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Statistical-Risk Valid Penalty

For an uncountable Θ and a penalty Pen(θ), θ ∈ Θ,suppose there is a countable Θ and L(θ) = 2L(θ)where L(θ) satisfies Kraft, such that, for all x , θ∗,

minθ∈Θ

{[log

pθ∗(x)

pθ(x)− dn(θ

∗, θ)]

+ Pen(θ)

}≥ min

θ∈Θ

{[log

pθ∗(x)

pθ(x)− dn(θ

∗, θ)]

+ L(θ)

}Proof of the risk conclusion:The second expression has expectation ≥ 0,so the first expression does too.This condition and result is obtained with J. Li and X. Luo(in Rissanen Festschrift 2008)

Barron Information Theory and Statistical Risk Analysis

Page 36: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Variable Complexity, Variable Distortion Cover

Equivalent statement of the condition: Pen(θ) is a validpenalty if for each θ in Θ there is a representor Θ in Θ withcomplexity L(θ), distortion ∆n(θ, θ) and

Pen(θ) ≥ L(θ) + ∆n(θ, θ)

where the distortion ∆n(θ, θ) is the difference in thediscrepancies at θ and θ

∆n(θ, θ) = logpθ(x)

pθ(x)+ dn(θ, θ

∗)− dn(θ, θ∗)

Barron Information Theory and Statistical Risk Analysis

Page 37: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

A Setting for Regression and log-density Estimation:Linear Span of a Dictionary

G is a dictionary of candidate basis functionsE.g. wavelets, splines, polynomials, trigonometric terms,sigmoids, explanatory variables and their interactions

Candidate functions in the linear span

fθ(x) =∑g∈G

θg g(x)

weighted `1 norm of coefficients

‖θ‖1 =∑

g

ag |θg |

weights ag = ‖g‖n where ‖g‖2n = 1

n∑n

i=1 g2(xi)

Barron Information Theory and Statistical Risk Analysis

Page 38: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

Example Models

Regression (focus of current presentation)

pθ(y |x) = Normal(fθ(x), σ2)

Logistic regression with y ∈ {0, 1}

pθ(y |x) = Logistic(fθ(x)) for y = 1

Log-density estimation (focus of Festschrift paper)

pθ(x) =p0(x) exp{fθ(x)}

cf

Gaussian graphical models

Barron Information Theory and Statistical Risk Analysis

Page 39: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Two-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

`1 Penalty

pen(θ) = λ‖θ‖1 where θ are coeff of fθ(x) =∑

g∈G θg g(x)

Popular penalty: Chen & Donoho (96) Basis Pursuit;Tibshirani (96) LASSO; Efron et al (04) LARS;Precursors: Jones (92), B.(90,93,94) greedy algorithm andanalysis of combined `1 and `0 penaltyWe want to avoid cross-validation in choice of λ

Data-compression: specify valid λ for coding interpretationRisk analysis: specify valid λ for risk ≤ resolvabilityComputation analysis: bounds accuracy of `1-penalizedgreedy pursuit algorithm

Barron Information Theory and Statistical Risk Analysis

Page 40: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Outline1 Some Principles of Information Theory and Statistics

Data CompressionMinimum Description Length Principle

2 Shannon Capacity Determines Limits of Statistical AccuracyThis part is saved for Wednesday

3 Penalized Likelihood AnalysisTwo-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

4 Regression with `1 PenaltyFixed σ2 CaseUnknown σ2 Case

5 Summary

Barron Information Theory and Statistical Risk Analysis

Page 41: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Regression with `1 penalty

`1 penalized log-density estimation, i.i.d. case

θ = argminθ

{1n

log1

pfθ(x)+ λn‖θ‖1

}Regression with Gaussian model, fixed σ2

minθ

{1

2σ21n

n∑i=1

(Yi − fθ(xi))2 +

12

log 2πσ2 +λn

σ‖θ‖1

}

Valid for

λn ≥√

2 log(2MG)

nwith MG = Card(G)

Barron Information Theory and Statistical Risk Analysis

Page 42: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Adaptive risk bound

For log density estimation with suitable λn

Ed(f ∗, fθ) ≤ infθ

{D(f ∗||fθ) + λn‖θ‖1

}For regression with fixed design points xi , fixed σ, and

λn =√

2 log(2M)n ,

E‖f ∗ − fθ‖2n

4σ2 ≤ infθ

{‖f ∗ − fθ‖2n

2σ2 +λn

σ‖θ‖1

}

Barron Information Theory and Statistical Risk Analysis

Page 43: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Adaptive risk bound specialized to regression

Again for fixed design and λn =√

2 log 2Mn , multiplying

through by 4σ2,

E‖f ∗ − fθ‖2n ≤ inf

θ

{2‖f ∗ − fθ‖2

n + 4σλn‖θ‖1

}In particular for all targets f ∗ = fθ∗ with finite ‖θ∗‖ the risk

bound 4σλn‖θ∗‖ is of order√

log Mn

Details in Barron, Luo (proceedings Workshop on Information Theory Methods in Science & Eng. 2008),Tampere, Finland: last week

Barron Information Theory and Statistical Risk Analysis

Page 44: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Comments on proof

Likelihood discrepancy plus complexity

12σ2

n∑i=1

(Yi − fθ(xi))2 − 1

2σ2

n∑i=1

(Yi − fθ(xi))2 + K log(2M)

Representor fθ of fθ of following form, with v near ‖θ‖1

fθ(x) =vK

K∑k=1

gk (x)/‖gk‖

g1, . . . gK picked at random from G, independently, where garises with probability proportional to |θg |agShows exists representor with like. discrep. + complexity

nv2

2K+ K log(2M)

Barron Information Theory and Statistical Risk Analysis

Page 45: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Comments on proof

Optimizing it yields penalty proportional to `1 normPenalty λ‖θ‖1 is valid for both data compression andstatistical risk requirements for λ ≥ λn where

λn =

√2 log(2M)

n

Especially useful for very large dictionariesImprovement for small dictionaries gets rid of log factor:log(2M) may be replaced by log(2e max{ M√

n , 1})

Barron Information Theory and Statistical Risk Analysis

Page 46: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Comments on proof

Existence of representor shown by random draw is aShannon-like demonstration of the variable cover (code)Similar approximation in analysis of greedy computation of`1 penalized least squares

Barron Information Theory and Statistical Risk Analysis

Page 47: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Aside: Random design case

May allow Xi random with distribution PX

Presuming functions in the library G are uniformly boundedAnalogous risk bounds hold (in submitted paper by Huang,Cheang, and Barron)

E‖f ∗ − fθ‖2 ≤ inf

θ

{2‖f ∗ − fθ‖2 + cλn‖θ‖1

}Allows for libraries G of infinite cardinality, replacing thelog MG with a metric entropy of GIf the linear span of G is dense in L2 then for all L2functions f ∗ the approximation error ‖f ∗ − fθ‖2 can bearranged to go to zero as the size of v = ‖θ∗‖1 increases.

Barron Information Theory and Statistical Risk Analysis

Page 48: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Aside: Simultaneous Minimax Rate Optimality

The resolvability bound tends to zero as λn =√

2 log MGn

gets small

E‖f ∗ − fθ‖2 ≤ inf

θ

{2‖f ∗ − fθ‖2 + cλn‖θ‖1

}Current work with Cong Huang shows for a broad class ofapproximate rates, the risk rate obtained from thisresolvability bound is minimax optimal for the class oftarget functions {f ∗ : infθ:‖θ‖1=v‖f ∗ − fθ‖2 ≤ Av},simultaneously for all such classes

Barron Information Theory and Statistical Risk Analysis

Page 49: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Fixed σ versus unknown σ

MDL with `1 penalty for each possible σ. Recall

minθ

{1

2σ21n

n∑i=1

(Yi − fθ(xi))2 +

12

log 2πσ2 +λn

σ‖θ‖1

}Provides a family of fits indexed by σ.For unknown σ suggest optimization over σ as well as θ

minθ,σ

{1

2σ21n

n∑i=1

(Yi − fθ(xi))2 +

12

log 2πσ2 +λn

σ‖θ‖1

}Slight modification of this does indeed satisfy our conditionfor an information-theoretically valid penalty and risk bound(details in the WITMSE 2008 proceedings)

Barron Information Theory and Statistical Risk Analysis

Page 50: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Fixed σ2 CaseUnknown σ2 Case

Best σ

Best σ for each θ solves the quadratic equation

σ2 = σ λn‖θ‖1 +1n

n∑i=1

(Yi − fθ(xi)

)2

By the quadratic formula the solution is

σ =12λn‖θ‖1 +

√√√√[12λn‖θ‖1

]2+

1n

n∑i=1

(Yi − fθ(xi)

)2

Barron Information Theory and Statistical Risk Analysis

Page 51: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Outline1 Some Principles of Information Theory and Statistics

Data CompressionMinimum Description Length Principle

2 Shannon Capacity Determines Limits of Statistical AccuracyThis part is saved for Wednesday

3 Penalized Likelihood AnalysisTwo-stage Code Redundancy, Resolvability, and RiskPenalized Likelihood Analysis for Continuous Parameters`1 Penalties are Information-Theoretically Valid

4 Regression with `1 PenaltyFixed σ2 CaseUnknown σ2 Case

5 Summary

Barron Information Theory and Statistical Risk Analysis

Page 52: A BLEND OF INFORMATION THEORY AND STATISTICS - Yale …arb4/presentations/Barron_FrejusPresentationDa… · This part is saved for Wednesday Outline 1 Some Principles of Information

Some Principles of Information Theory and StatisticsShannon Capacity Determines Limits of Statistical Accuracy

Penalized Likelihood AnalysisRegression with `1 Penalty

Summary

Summary

Handle penalized likelihoods with continuous domains for θ

Information-theoretically valid penalties: Penalty exceedcomplexity plus distortion of optimized representor of θ

Yields MDL interpretation and stat. risk controlled byresolvability

`0 penalty dim2 log n classically analyzed

`1 penalty σλn ‖θ‖1 analyzed here: valid in regression for

λn ≥√

2(log2M)/n

Can handle fixed or unknown σ

Barron Information Theory and Statistical Risk Analysis