Adaptive Submodularity : A New Approach to Active Learning and Stochastic Optimization

18
1 fornia Institute of Technology er for the Mathematics of Information Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Daniel Golovin and Andreas Krause

description

Adaptive Submodularity : A New Approach to Active Learning and Stochastic Optimization. Daniel Golovin and Andreas Krause . Max K-Cover (Oil Spill Edition). Submodularity. Discrete diminishing returns property for set functions. ``Playing an action at an earlier stage - PowerPoint PPT Presentation

Transcript of Adaptive Submodularity : A New Approach to Active Learning and Stochastic Optimization

Page 1: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

1California Institute of Technology Center for the Mathematics of Information

Adaptive Submodularity:A New Approach to Active Learning and Stochastic

OptimizationDaniel Golovin and Andreas Krause

Page 2: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

2California Institute of Technology Center for the Mathematics of Information

Max K-Cover (Oil Spill Edition)

Page 3: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

3California Institute of Technology Center for the Mathematics of Information

SubmodularityTi

me

Tim

e

Discrete diminishing returns property for set functions.``Playing an action at an earlier stage only increases its marginal benefit''

Page 4: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

4California Institute of Technology Center for the Mathematics of Information

The Greedy Algorithm

Theorem [Nemhauser et al ‘78]

Page 5: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

5California Institute of Technology Center for the Mathematics of Information

Stochastic Max K-Cover

Asadpour & Saberi (`08): (1-1/e)-approx if sensors (independently) either work perfectly or fail completely.

Bayesian: Known failure distribution. Adaptive: Deploy a sensor and see what you get. Repeat K times.

0.5

0.2

0.3

At 1st location

Page 6: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

6California Institute of Technology Center for the Mathematics of Information

Adaptive SubmodularityTi

me

Playing an action at an earlier stage only increases its marginal benefit

expected(taken over its outcome)

Gain moreGain less

(i.e., at an ancestor)

Select ItemStochasticOutcome

Page 7: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

7California Institute of Technology Center for the Mathematics of Information

Adaptive Monotonicity

Page 8: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

8California Institute of Technology Center for the Mathematics of Information

What is it good for?

Allows us to generalize various results to the adaptive realm, including:

• (ln(n)+1)-approximation for Set Cover

• (1-1/e)-approximation for Max K-Cover, submodular maximization subject to a cardinality constraint

Page 9: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

9California Institute of Technology Center for the Mathematics of Information

Recall the Greedy Algorithm

Theorem [Nemhauser et al ‘78]

Page 10: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

10California Institute of Technology Center for the Mathematics of Information

The Adaptive-Greedy Algorithm

Theorem

Page 11: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

11California Institute of Technology Center for the Mathematics of Information

[Adapt-monotonicity] - -

( ) - [Adapt-submodularity]

Page 12: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

12California Institute of Technology Center for the Mathematics of Information

The world-state dictates which path in the tree we’ll take.

1. For each node at layer i+1, 2. Sample path to layer j, 3. Play the resulting layer j action at layer i+1.

How to play layer j at layer i+1

By adapt. submod.,playing a layer earlieronly increases it’s marginal benefit

Page 13: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

13California Institute of Technology Center for the Mathematics of Information

[Adapt-monotonicity] - -

( ) - ( ) -

[Def. of adapt-greedy]

( ) - [Adapt-submodularity]

Page 14: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

14California Institute of Technology Center for the Mathematics of Information

Page 15: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

15California Institute of Technology Center for the Mathematics of Information

2

1 3

Stochastic Max Cover is Adapt-Submod

1 3

Gain moreGain less

adapt-greedy is a (1-1/e) ≈ 63% approximation to the adaptive optimal solution.

Random sets distributedindependently.

Page 16: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

16California Institute of Technology Center for the Mathematics of Information

Stochastic Min Cost Cover Adaptively get a threshold amount of value. Minimize expected number of actions. If objective is adapt-submod and

monotone, we get a logarithmic approximation.

[Goemans & Vondrak, LATIN ‘06][Liu et al., SIGMOD ‘08]

[Feige, JACM ‘98]

[Guillory & Bilmes, ICML ‘10]c.f., Interactive Submodular Set Cover

Page 17: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

17California Institute of Technology Center for the Mathematics of Information

Optimal Decision Trees

x1

x2

x3

1

1

0

0 0

=

=

=

Garey & Graham, 1974; Loveland, 1985; Arkin et al., 1993; Kosaraju et al., 1999; Dasgupta, 2004; Guillory & Bilmes, 2009; Nowak, 2009; Gupta et al., 2010

“Diagnose the patient as cheaply as possible (w.r.t. expected cost)”

1

10

Page 18: Adaptive  Submodularity : A New Approach to Active Learning and Stochastic Optimization

18California Institute of Technology Center for the Mathematics of Information

Conclusions New structural property useful for design & analysis of

adaptive algorithms Powerful enough to recover and generalize many known

results in a unified manner. (We can also handle costs) Tight analyses and optimal approximation factors in

many cases.

2

1 3

x1

x2

x3

1

1

0

0 010.5

0.30.5 0.4

0.2

0.2 0.5