Ted Willke, Sr Principal Engineer, Intel at MLconf SEA - 5/20/16

Post on 13-Apr-2017

1.000 views 4 download

Transcript of Ted Willke, Sr Principal Engineer, Intel at MLconf SEA - 5/20/16

Can Cognitive Neuroscience Provide a Theory of Deep Learning Capacity?Ted Willke and the Mind’s Eye Team

Intel Labs

May 20, 2016

2

“Breakthrough innovation occurs when we bring down boundaries and encourage disciplines to learn from each other”

― Gyan Nagpal, Talent Economics: The Fine Line Between Winning and Losing the Global War for Talent

2

MIND’S EYE

3

Cognitive Neuroscience

4

Cognitive Neuroscience Is the study of the neurobiological mechanisms that underlie

cognitive processes, like attention, control, and decision making

Answer questions like: How does the brain coordinate behaviour to achieve goals? What are the brain structures upon which these functions depend? How does brain function differ amongst people?

Draws upon brain imaging/recordings and other observations to derive models

5

Context-Dependent Decision Making

Michael Shvartsman, Vibhav Srivatsava, Narayanan Sundaram, Jonathan D. Cohen, “Using behavior to decode allocation of attention in context dependent decision making”, accepted at International Conference on Cognitive Modeling, 2016.

6

Selective Forgetting

Kim, Ghootae and Lewis-Peacock, Jarrod A. and Norman, Kenneth A. and Turk-Browne, Nicholas B., “Pruning of memories by context-based prediction error,” Proceedings of the National Academy of Sciences, 2014

7

Production and comprehension of naturalistic narrative speech

Silbert LJ, Honey CJ, Simony E, Poeppel D, Hasson U (2014) Coupled neural systems underlie the production and comprehension of naturalistic narrative speech. Proc Natl Acad Sci USA 111:E4687-4696.

8

CRACKS APPEAR, DISRUPTIVE IDEAS 30 years on

MIT Press, 1986

9

Cognitive Neuroscience

Adapted from Marvin Minksy in Artificial Intelligence at MIT, Expanding Frontiers, Patrick H. Winston (Ed.), Vol.1, MIT Press, 1990. Reprinted in AI Magazine, Summer 1991

evolve

10

Neural networks

11

Neural Network preliminaries

http://wiki.apache.org/hama/MultiLayerPerceptron

12

Neural Network preliminaries

Lecun et al., “Deep Learning” in Nature (2015)

13

Arbitrary functions

https://upload.wikimedia.org/wikipedia/commons/7/7b/XOR_perceptron_net.png

14

The original tenets of parallel distributed processing (roughly)

1. Cognitive processes arise from the real-time propagation of activation via weighted connections

2. Active representations are patterns of activation distributed over ensembles of units

3. Processing is interactive (bidirectional)

4. Knowledge is encoded in the connection weights (not in a separate store)

5. Learning and long-term memory depend on changes to these weights

6. Processing, learning, and representation are graded and continuous

7. Processing, learning, and representation depend on the environment

T.T. Rogers, J.L. McClelland / Cognitive Science 28 (2014)

15

Brain-Inspired machine learning

Structure-Inspired Learning

Neurons (e.g., spiking models)

Networks (e.g., deep belief networks)

Architectures (e.g., Human Brain Project)

Cognitive-Inspired Learning

Reinforcement Learning

Context-based Memory

Noisy Decision Making

15

"Gray754" by Henry Vandyke Carter - Henry Gray(1918) Anatomy of the Human Body

16

Deep learning takes advantage of parallel distributed processing

http://www.amax.com/blog/wp-content/uploads/2015/12/blog_deeplearning3.jpg

17

Winning top spots in visual recognition challenges, etc.

(1) Lin et al., 2015, (2) https://www.cityscapes-dataset.com/dataset-overview/ (3) Deng et al., 2009 (4) http://lsun.cs.princeton.edu/2015.html

MS COCO (Common Objects in Context) CityScapes Datasets (Semantic Understanding)

ImageNet (Object Localization) LSUN (Saliency Prediction)

18

Yang et al. (2015)

What are sitting in the basket on a bicycle?

19

Yang et al. (2015)

Stacked Attention Networks for Image Question Answering

20

The Glory and the remaining mysteryWe have achieved…

Exceeding human-level performance on visual recognition tasks

Mastering more and more complex games (Go)

Demonstrating human-level control in reinforcement learning (Atari)

Question-answering and other AI services are upon us

but we still don’t know…

How learnt (feature) representations are encoded (or if they converge for the same networks trained on the same data)

The capacity for learning representations

The trade-off between efficiency of representation and flexibility of processing

How things learnt interfere with each other

21

Representations and Learning Capacity

22Li et al. (ICLR 2016)

Representation encoding: meaningful and consistent?

Can we reliably map feature representations between these networks?

23Li et al. (ICLR 2016)

Convergent Learning?

Conclusions:

1. Some features are learned reliably in multiple nets (some are not)

2. Units learn to span low-D subspaces, which are common (but specific basis vectors are not)

3. Representations are encoded as a mix of single unit and slightly distributed codes

4. Mean activation values across different networks converge to a nearly identical distribution

24

Can cognitive neuroscience provide any insight into the nature of learning

and task capacity?

25

The appeal of highly-parallel neural networks

Both cognitive neuroscience and machine learning applications exploit the following two features of neural networks to great benefit:

a) The ability to learn and process complex representations, taking into account a large number of interrelated and interacting constraints

b) The ability for the same network to process a wide range of potentially disparate representations (or tasks), sometimes called “multitask learning.”

But what are their limits??

26

The brain: The black box at the end of our necks• Facts:

Only 2% of body weight but uses up to 20% of energy

~200B neurons

Neurons fire up to ~10 kHz

1K to 10K connections per neuron

• Cerebral neocortex:

~20B neurons

~125 trillion synapses

There are more ways to organize the neocortex’s ~125 trillion synapses than stars in the known universe

27

The paradox – one task at a time

28

A fundamental puzzle concerning human processing

Why, in some circumstances is the brain capable of a remarkable degree of parallelism (e.g., locomotion, navigation, speech, and bimanual gesticulation), while in others it’s capacity for parallelism is radically limited (e.g., the inability to conduct mental arithmetic while constructing a grocery list at the same time)??!!

29

A theory

The difference in multitasking ability may reflect the degree to which different tasks rely on shared representations

The more that different processes interact, the stronger the imposition of seriality

May reflect a fundamental trade-off in neural network architectures between the efficiency of shared representations (and the capacity for generalization that they afford) and the effectiveness of multitasking.

30

Multi-tasking and cross-talk

Feng et al. (CABN 2014)

31

You will see a sequence of words. Quickly say the color of the letters.

32

SNOW

33

Ready!

34

BLUE

35

RED

36

BLACK

37

GREEN

41

BLACK

42

BLUE

46

GREEN

49

RED

52

BLUE

54

Now with the words upside down.

55

BLACK

56

GREEN

58

BLACK

59

RED

60

Were you faster to answer?

61

A Demonstration of interference

Stroop (1935)

62

multi-tasking interference (In the stroop test)

Cohen et al. (1990)

Color Word

Verbalize Task

63

Control-Demanding Behavior (Feng et al. 2014)

First to describe the trade-off between the efficiency of representation (“multiplexing”) and the simultaneous engagement of different processing pathways (“multitasking”)

Showed that even a modest amount of multiplexing rapidly introduces cross-talk among processing pathways

Proposed that the large advantage of efficient encoding have driven the human brain to favour this over the capacity for control-demanding processes.

64

Types of interference

65

Maximum independent set (MIS)

The MIS is the largest set of processes in the network that can be simultaneously executed without interference.

66

network structure (distribution complexity)

The network capacity for multitasking depends on the distribution of in-degrees and out-degrees of the network (we only play with in-degree of output components though)

We represent this with a “distribution complexity” symmetry measure (maximized for uniform distribution)

We study the characteristics of the network with DC fixed

67

Takeaway: Even modest amounts of process overlap impose dramatic constraints on parallel processing capability

68

Trade-off between generalization and parallelism: Feed-Forward simulation

69

Training/Test details

Training

20 network groups, 20 random initializations per group

All networks trained on same stimuli, 16 tasks

Trained to generate 1-hot task outputs (MSE < 0.0001)

Test

70/30 split

Generalization is MSE(ave) for ALL stimuli in test set

Parallel processing is measured response to (2,3,4) tasks simultaneously activated, measuring MSE for target pattern

70

Shared Representations

Smaller weights (a) Larger weights (b)

71

Generalization vs parallel processing capability

72

Parallel processing capability vs max initial weights

73

Future work

Extend analysis to weighted graphs

Study more complex networks (i.e., deeper structures, recurrent connections)

Study human performance (via neuroimaging data)!

74

C. elegans

74

The OpenWorm Project(image generated by neuroConstruct)

SINCE 1986

Thank you!