CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and...

32
CSC421 Lecture 1: Introduction Roger Grosse and Jimmy Ba Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 1 / 28

Transcript of CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and...

Page 1: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

CSC421 Lecture 1: Introduction

Roger Grosse and Jimmy Ba

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 1 / 28

Page 2: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Course information

Second course in machine learning, with a focus on neural networks

Replaces CSC321; similar content, but less introductory material andmore cutting-edge topicsAssumes knowledge of basic ML algorithms: linear regression, logisticregression, maximum likelihood, PCA, EM, etc.First 2/3: supervised learningLast 1/3: unsupervised learning and reinforcement learning

Two sections

Equivalent content, same assignments and exams

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 2 / 28

Page 3: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Course information

Formal prerequisites:

Multivariable Calculus: MAT235/MAT237/MAT257Linear Algebra: MAT221H1/MAT223H1/MAT240H1Machine Learning: CSC411/STA314

Prerequisites will be enforced, including for grad students.

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 3 / 28

Page 4: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Course information

Expectations and marking (undergrads)Written homeworks (20% of total mark)

Due Wednesday nights at 11:59pm, starting 1/242-3 short conceptual questionsUse material covered up through Tuesday of the preceding week

4 programming assignments (30% of total mark)

Python, PyTorch10-15 lines of codemay also involve some mathematical derivationsgive you a chance to experiment with the algorithms

Exams

midterm (15%)final (35%)

See Course Information handout for detailed policies

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 4 / 28

Page 5: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Course information

Expectations and marking (grad students)Same as undergrads:

Written homeworks: 20%Programming assignments: 30%Midterm: 15%

Final project: 35%

See Course Information handout for detailed policies

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 5 / 28

Page 6: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

How to get free GPUs

Colab (Recommended) Google Colab is a web-based iPythonNotebook service that has access to a free Nvidia K80 GPU perGoogle account.

GCE (Recommended) Google Compute Engine delivers virtualmachines running in Google’s data center. You get $300 free creditwhen you sign up.

CS Teaching Lab There are some very old GPUs in our CS TeachingLabs / CDF labs

See Course Information handout for the details

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 6 / 28

Page 7: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Course information

Course web page:http://www.cs.toronto.edu/~rgrosse/courses/csc421_2019/

Includes detailed course information handout

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 7 / 28

Page 8: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

What is machine learning?

For many problems, it’s difficult to program the correct behavior byhand

recognizing people and objectsunderstanding human speech

Machine learning approach: program an algorithm to automaticallylearn from data, or from experience

Some reasons you might want to use a learning algorithm:

hard to code up a solution by hand (e.g. vision, speech)system needs to adapt to a changing environment (e.g. spam detection)want the system to perform better than the human programmersprivacy/fairness (e.g. ranking search results)

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 8 / 28

Page 9: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

What is machine learning?

For many problems, it’s difficult to program the correct behavior byhand

recognizing people and objectsunderstanding human speech

Machine learning approach: program an algorithm to automaticallylearn from data, or from experience

Some reasons you might want to use a learning algorithm:

hard to code up a solution by hand (e.g. vision, speech)system needs to adapt to a changing environment (e.g. spam detection)want the system to perform better than the human programmersprivacy/fairness (e.g. ranking search results)

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 8 / 28

Page 10: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

What is machine learning?

For many problems, it’s difficult to program the correct behavior byhand

recognizing people and objectsunderstanding human speech

Machine learning approach: program an algorithm to automaticallylearn from data, or from experience

Some reasons you might want to use a learning algorithm:

hard to code up a solution by hand (e.g. vision, speech)system needs to adapt to a changing environment (e.g. spam detection)want the system to perform better than the human programmersprivacy/fairness (e.g. ranking search results)

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 8 / 28

Page 11: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

What is machine learning?

Types of machine learning

Supervised learning: have labeled examples of the correct behaviorReinforcement learning: learning system receives a reward signal,tries to learn to maximize the reward signalUnsupervised learning: no labeled examples – instead, looking forinteresting patterns in the data

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 9 / 28

Page 12: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Supervised learning examples

Supervised learning: have labeled examples of the correct behavior

e.g. Handwritten digit classification with the MNIST dataset

Task: given an image of a handwritten digit, predict the digit class

Input: the imageTarget: the digit class

Data: 70,000 images of handwritten digits labeled by humans

Training set: first 60,000 images, used to train the networkTest set: last 10,000 images, not available during training, used toevaluate performance

This dataset is the “fruit fly” of neural net research

Neural nets already achieved > 99% accuracy in the 1990s, but westill continue to learn a lot from it

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 10 / 28

Page 13: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Supervised learning examples

Supervised learning: have labeled examples of the correct behavior

e.g. Handwritten digit classification with the MNIST dataset

Task: given an image of a handwritten digit, predict the digit class

Input: the imageTarget: the digit class

Data: 70,000 images of handwritten digits labeled by humans

Training set: first 60,000 images, used to train the networkTest set: last 10,000 images, not available during training, used toevaluate performance

This dataset is the “fruit fly” of neural net research

Neural nets already achieved > 99% accuracy in the 1990s, but westill continue to learn a lot from it

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 10 / 28

Page 14: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Supervised learning examples

Supervised learning: have labeled examples of the correct behavior

e.g. Handwritten digit classification with the MNIST dataset

Task: given an image of a handwritten digit, predict the digit class

Input: the imageTarget: the digit class

Data: 70,000 images of handwritten digits labeled by humans

Training set: first 60,000 images, used to train the networkTest set: last 10,000 images, not available during training, used toevaluate performance

This dataset is the “fruit fly” of neural net research

Neural nets already achieved > 99% accuracy in the 1990s, but westill continue to learn a lot from it

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 10 / 28

Page 15: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Supervised learning examples

What makes a “2”?It is very hard to say what makes a 2

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 11 / 28

Page 16: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Supervised learning examples

Object recognitionSome examples from an earlier version of the net

(Krizhevsky and Hinton, 2012)

ImageNet dataset: thousands of categories, millions of labeled images

Lots of variability in viewpoint, lighting, etc.

Error rate dropped from 26% to under 4% over the course of a few years!

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 12 / 28

Page 17: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Supervised learning examples

Neural Machine Translation

(Wu et al., 2016)

Now the production model on Google TranslateRoger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 13 / 28

Page 18: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Supervised learning examples

Caption generation

(Xu et al., 2015)

Given: dataset of Flickr images with captionsMore examples at http://deeplearning.cs.toronto.edu/i2t

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 14 / 28

Page 19: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Unsupervised learning examples

In generative modeling, we want to learn a distribution over some dataset,such as natural images.We can evaluate a generative model by sampling from the model and seeingif it looks like the data.These results were considered impressive in 2014:

Denton et al., 2014, Deep generative image models using a Laplacian pyramid of adversarial networks

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 15 / 28

Page 20: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Unsupervised learning examples

Fast-forward to 2017:

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 16 / 28

Page 21: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Unsupervised learning examples

The progress of generative models:

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 17 / 28

Page 22: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Unsupervised learning examples

Recent exciting result: a model called the CycleGAN takes lots ofimages of one category (e.g. horses) and lots of images of anothercategory (e.g. zebras) and learns to translate between them.

https://github.com/junyanz/CycleGAN

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 18 / 28

Page 23: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Reinforcement learning

An agent interacts with an environment (e.g. game of Breakout)

In each time step,

the agent receives observations (e.g. pixels) which give it informationabout the state (e.g. positions of the ball and paddle)the agent picks an action (e.g. keystrokes) which affects the state

The agent periodically receives a reward (e.g. points)

The agent wants to learn a policy, or mapping from observations toactions, which maximizes its average reward over time

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 19 / 28

Page 24: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Reinforcement learning

DeepMind trained neural networks to play many different Atari games

given the raw screen as input, plus the score as a reward

single network architecture shared between all the games

in many cases, the networks learned to play better than humans (interms of points in the first minute)

https://www.youtube.com/watch?v=V1eYniJ0Rnk

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20 / 28

Page 25: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Reinforcement learning for control

Learning locomotion control from scratch

The reward is to run as far as possible over all the obstacles

single control policy that learns to adapt to different terrains

https://www.youtube.com/watch?v=hx_bgoTF7bs

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 21 / 28

Page 26: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

What are neural networks?

Most of the biological details aren’t essential, so we use vastlysimplified models of neurons.

While neural nets originally drew inspiration from the brain, nowadayswe mostly think about math, statistics, etc.

Neural networks are collections of thousands (or millions) of thesesimple processing units that together perform useful computations.

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 22 / 28

Page 27: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

What are neural networks?

Why neural nets?

inspiration from the brain

proof of concept that a neural architecture can see and hear!

very effective across a range of applications (vision, text, speech,medicine, robotics, etc.)

widely used in both academia and the tech industry

powerful software frameworks (PyTorch, TensorFlow, etc.) let usquickly implement sophisticated algorithms

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 23 / 28

Page 28: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

“Deep learning”

Deep learning: many layers (stages) of processing

E.g. this network which recognizes objects in images:

Figure 2: An illustration of the architecture of our CNN, explicitly showing the delineation of responsibilitiesbetween the two GPUs. One GPU runs the layer-parts at the top of the figure while the other runs the layer-partsat the bottom. The GPUs communicate only at certain layers. The network’s input is 150,528-dimensional, andthe number of neurons in the network’s remaining layers is given by 253,440–186,624–64,896–64,896–43,264–4096–4096–1000.

neurons in a kernel map). The second convolutional layer takes as input the (response-normalizedand pooled) output of the first convolutional layer and filters it with 256 kernels of size 5 ⇥ 5 ⇥ 48.The third, fourth, and fifth convolutional layers are connected to one another without any interveningpooling or normalization layers. The third convolutional layer has 384 kernels of size 3 ⇥ 3 ⇥256 connected to the (normalized, pooled) outputs of the second convolutional layer. The fourthconvolutional layer has 384 kernels of size 3 ⇥ 3 ⇥ 192 , and the fifth convolutional layer has 256kernels of size 3 ⇥ 3 ⇥ 192. The fully-connected layers have 4096 neurons each.

4 Reducing Overfitting

Our neural network architecture has 60 million parameters. Although the 1000 classes of ILSVRCmake each training example impose 10 bits of constraint on the mapping from image to label, thisturns out to be insufficient to learn so many parameters without considerable overfitting. Below, wedescribe the two primary ways in which we combat overfitting.

4.1 Data Augmentation

The easiest and most common method to reduce overfitting on image data is to artificially enlargethe dataset using label-preserving transformations (e.g., [25, 4, 5]). We employ two distinct formsof data augmentation, both of which allow transformed images to be produced from the originalimages with very little computation, so the transformed images do not need to be stored on disk.In our implementation, the transformed images are generated in Python code on the CPU while theGPU is training on the previous batch of images. So these data augmentation schemes are, in effect,computationally free.

The first form of data augmentation consists of generating image translations and horizontal reflec-tions. We do this by extracting random 224⇥ 224 patches (and their horizontal reflections) from the256⇥256 images and training our network on these extracted patches4. This increases the size of ourtraining set by a factor of 2048, though the resulting training examples are, of course, highly inter-dependent. Without this scheme, our network suffers from substantial overfitting, which would haveforced us to use much smaller networks. At test time, the network makes a prediction by extractingfive 224 ⇥ 224 patches (the four corner patches and the center patch) as well as their horizontalreflections (hence ten patches in all), and averaging the predictions made by the network’s softmaxlayer on the ten patches.

The second form of data augmentation consists of altering the intensities of the RGB channels intraining images. Specifically, we perform PCA on the set of RGB pixel values throughout theImageNet training set. To each training image, we add multiples of the found principal components,

4This is the reason why the input images in Figure 2 are 224 ⇥ 224 ⇥ 3-dimensional.

5

(Krizhevsky et al., 2012)

Each of the boxes consists of many neuron-like units similar to the one onthe previous slide!

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 24 / 28

Page 29: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

“Deep learning”

You can visualize what a learned feature is responding to by findingan image that excites it. (We’ll see how to do this.)

Higher layers in the network often learn higher-level, moreinterpretable representations

https://distill.pub/2017/feature-visualization/

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 25 / 28

Page 30: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

“Deep learning”

You can visualize what a learned feature is responding to by findingan image that excites it.Higher layers in the network often learn higher-level, moreinterpretable representations

https://distill.pub/2017/feature-visualization/

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 26 / 28

Page 31: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Software frameworks

Array processing (NumPy)

vectorize computations (express them in terms of matrix/vectoroperations) to exploit hardware efficiency

Neural net frameworks: PyTorch, TensorFlow, etc.

automatic differentiationcompiling computation graphslibraries of algorithms and network primitivessupport for graphics processing units (GPUs)

For this course:

Python, NumPyAutograd, a lightweight automatic differentiation package written byProfessor David Duvenaud and colleaguesPyTorch, a widely used neural net framework

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 27 / 28

Page 32: CSC421 Lecture 1: Introductionrgrosse/courses/csc421_2019/slides/lec01.pdf · Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 20/28. Reinforcement learning for control Learning

Software frameworks

Why take this class, if PyTorch does so much for you?

So you know what do to if something goes wrong!

Debugging learning algorithms requires sophisticated detective work,which requires understanding what goes on beneath the hood.

That’s why we derive things by hand in this class!

Roger Grosse and Jimmy Ba CSC421 Lecture 1: Introduction 28 / 28