Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor...
Transcript of Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor...
![Page 1: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/1.jpg)
Tensor Flow
Tensors: n-dimensional arrays
A sequence of tensor operationsDeep learning process are flows of tensors
Vector: 1-D tensorMatrix: 2-D tensor
Can represent also many machine learning algorithms
![Page 2: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/2.jpg)
A simple ReLU network
a1 b1 c1
a0 b0 c0
w
a1=a0wa,a+b0wb,a+c0wc,a
b1=a0wa,b+b0wb,b+c0wc,b
c1=a0wa,c+b0wb,c+c0wc,c
Apply relu(…) on a1, b1, c1
Slower approachPer-neuron operation
More efficient approachMatrix operation
![Page 3: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/3.jpg)
As matrix operations
a0
a1 b1 c1
a0 b0 c0
w
. =
=relu( )
b0 c0 a1 b1 c1
a1a1
=relu( )b1b1
=relu( )c1c1
wa,a
wb,a
wc,a
wa,b
wb,b
wc,b
wa,c
wb,c
wc,c
![Page 4: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/4.jpg)
With TensorFlow
a1 b1 c1
a0 b0 c0
w
out = tf.nn.relu(y)
y = tf.matmul(x, w)
x w
a0 . =b0 c0
wa,a
wb,a
wc,a
wa,b
wb,b
wc,b
wa,c
wb,c
wc,c
a1 b1 c1
=relu( )a1a1
=relu( )b1b1
=relu( )c1c1
import tensorflow as tf
![Page 5: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/5.jpg)
Define Tensors
xa,a
xb,a
xc,a
xa,b
xb,b
xc,b
xa,c
xb,c
xc,c
w Variable(<initial-value>, name=<optional-name>)
w = tf.Variable(tf.random_normal([3, 3]), name='w')import tensorflow as tf
y = tf.matmul(x, w)relu_out = tf.nn.relu(y)
Variable stores the state of current execution
Others are operations
![Page 6: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/6.jpg)
TensorFlow
Code so far defines a data flow graph
MatMul
ReLU
Variable
x
w = tf.Variable(tf.random_normal([3, 3]), name='w')import tensorflow as tf
y = tf.matmul(x, w)relu_out = tf.nn.relu(y)
Each variable corresponds to a node in the graph, not the result
Can be confusing at the beginning
![Page 7: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/7.jpg)
TensorFlow
Code so far defines a data flow graphNeeds to specify how we want to execute the graph MatMul
ReLU
Variable
x
SessionManage resource for graph execution
w = tf.Variable(tf.random_normal([3, 3]), name='w')sess = tf.Session()
y = tf.matmul(x, w)relu_out = tf.nn.relu(y)
import tensorflow as tf
result = sess.run(relu_out)
![Page 8: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/8.jpg)
Graph
Fetch
Retrieve content from a node
w = tf.Variable(tf.random_normal([3, 3]), name='w')sess = tf.Session()
y = tf.matmul(x, w)relu_out = tf.nn.relu(y)
import tensorflow as tf
print sess.run(relu_out)
MatMul
ReLU
Variable
x
Fetch
We have assembled the pipesFetch the liquid
![Page 9: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/9.jpg)
Graph
sess = tf.Session()
y = tf.matmul(x, w)relu_out = tf.nn.relu(y)
import tensorflow as tf
print sess.run(relu_out)sess.run(tf.initialize_all_variables())
w = tf.Variable(tf.random_normal([3, 3]), name='w')
Initialize Variable
Variable is an empty node
MatMul
ReLU
Variable
x
Fetch
Fill in the content of a Variable node
![Page 10: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/10.jpg)
Graph
sess = tf.Session()
y = tf.matmul(x, w)relu_out = tf.nn.relu(y)
import tensorflow as tf
print sess.run(relu_out)sess.run(tf.initialize_all_variables())
w = tf.Variable(tf.random_normal([3, 3]), name='w')x = tf.placeholder("float", [1, 3])
Placeholder
How about x?
MatMul
ReLU
Variable
x
Fetch
placeholder(<data type>, shape=<optional-shape>, name=<optional-name>)Its content will be fed
![Page 11: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/11.jpg)
Graph
import numpy as npimport tensorflow as tfsess = tf.Session()x = tf.placeholder("float", [1, 3])w = tf.Variable(tf.random_normal([3, 3]), name='w')y = tf.matmul(x, w)relu_out = tf.nn.relu(y)sess.run(tf.initialize_all_variables())print sess.run(relu_out, feed_dict={x:np.array([[1.0, 2.0, 3.0]])})
Feed
MatMul
ReLU
Variable
x
FetchPump liquid into the pipe
Feed
![Page 12: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/12.jpg)
Session management
Needs to release resource after usesess.close()
Common usage
with tf.Session() as sess: …
Interactivesess = InteractiveSession()
![Page 13: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/13.jpg)
Prediction
import numpy as npimport tensorflow as tf
with tf.Session() as sess: x = tf.placeholder("float", [1, 3]) w = tf.Variable(tf.random_normal([3, 3]), name='w') relu_out = tf.nn.relu(tf.matmul(x, w)) softmax = tf.nn.softmax(relu_out) sess.run(tf.initialize_all_variables()) print sess.run(softmax, feed_dict={x:np.array([[1.0, 2.0, 3.0]])})
SoftmaxMake predictions for n targets that sum to 1
![Page 14: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/14.jpg)
Prediction Difference
import numpy as npimport tensorflow as tf
with tf.Session() as sess: x = tf.placeholder("float", [1, 3]) w = tf.Variable(tf.random_normal([3, 3]), name='w') relu_out = tf.nn.relu(tf.matmul(x, w)) softmax = tf.nn.softmax(relu_out) sess.run(tf.initialize_all_variables()) answer = np.array([[0.0, 1.0, 0.0]]) print answer - sess.run(softmax, feed_dict={x:np.array([[1.0, 2.0, 3.0]])})
![Page 15: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/15.jpg)
Learn parameters: Loss
Define loss function
Loss function for softmax
softmax_cross_entropy_with_logits( logits, labels, name=<optional-name>)
labels = tf.placeholder("float", [1, 3])cross_entropy = tf.nn.softmax_cross_entropy_with_logits( relu_out, labels, name='xentropy')
![Page 16: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/16.jpg)
Learn parameters: Optimization
Gradient descent
class GradientDescentOptimizerGradientDescentOptimizer(learning rate)
labels = tf.placeholder("float", [1, 3])cross_entropy = tf.nn.softmax_cross_entropy_with_logits( relu_out, labels, name='xentropy')optimizer = tf.train.GradientDescentOptimizer(0.1)train_op = optimizer.minimize(cross_entropy)sess.run(train_op, feed_dict= {x:np.array([[1.0, 2.0, 3.0]]), labels:answer})
learning rate = 0.1
![Page 17: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/17.jpg)
Iterative update
labels = tf.placeholder("float", [1, 3])cross_entropy = tf.nn.softmax_cross_entropy_with_logits( relu_out, labels, name=‘xentropy')optimizer = tf.train.GradientDescentOptimizer(0.1)train_op = optimizer.minimize(cross_entropy)for step in range(10): sess.run(train_op, feed_dict= {x:np.array([[1.0, 2.0, 3.0]]), labels:answer})
Gradient descent usually needs more than one step
Run multiple times
![Page 18: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/18.jpg)
Add parameters for Softmax
…softmax_w = tf.Variable(tf.random_normal([3, 3]))logit = tf.matmul(relu_out, softmax_w)softmax = tf.nn.softmax(logit)…cross_entropy = tf.nn.softmax_cross_entropy_with_logits( logit, labels, name=‘xentropy')…
Do not want to use only non-negative input
Softmax layer
![Page 19: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/19.jpg)
Add biases
…w = tf.Variable(tf.random_normal([3, 3]))b = tf.Variable(tf.zeros([1, 3]))relu_out = tf.nn.relu(tf.matmul(x, w) + b)softmax_w = tf.Variable(tf.random_normal([3, 3]))softmax_b = tf.Variable(tf.zeros([1, 3]))logit = tf.matmul(relu_out, softmax_w) + softmax_bsoftmax = tf.nn.softmax(logit)…
Biases initialized to zero
![Page 20: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/20.jpg)
Make it deep
…x = tf.placeholder("float", [1, 3])relu_out = xnum_layers = 2for layer in range(num_layers): w = tf.Variable(tf.random_normal([3, 3])) b = tf.Variable(tf.zeros([1, 3])) relu_out = tf.nn.relu(tf.matmul(relu_out, w) + b)…
Add layers
![Page 21: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/21.jpg)
Visualize the graph
TensorBoard
writer = tf.train.SummaryWriter( '/tmp/tf_logs', sess.graph_def)
tensorboard --logdir=/tmp/tf_logs
![Page 22: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/22.jpg)
Improve naming, improve visualization
name_scope(name)Help specify hierarchical names
…for layer in range(num_layers): with tf.name_scope('relu'): w = tf.Variable(tf.random_normal([3, 3])) b = tf.Variable(tf.zeros([1, 3])) relu_out = tf.nn.relu(tf.matmul(relu_out, w) + b)…
Will help visualizer to better understand hierarchical relation
Move to outside the loop?
![Page 23: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/23.jpg)
Add name_scope for softmax
Before After
![Page 24: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/24.jpg)
Add regularization to the loss
eg. L2 regularize on the Softmax layer parameters
…l2reg = tf.reduce_sum(tf.square(softmax_w))loss = cross_entropy + l2regtrain_op = optimizer.minimize(loss)…print sess.run(l2reg)…
Add it to the lossAutomatic gradient calculation
![Page 25: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/25.jpg)
Add a parallel path
![Page 26: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/26.jpg)
Use activation as bias
Everything is a tensor
![Page 27: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/27.jpg)
Residual learning
ILSVRC 2015 classification task winerHe et al. 2015
![Page 28: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/28.jpg)
Visualize states
Add summaries scalar_summary histogram_summary
merged_summaries = tf.merge_all_summaries()results = sess.run([train_op, merged_summaries], feed_dict=…)writer.add_summary(results[1], step)
![Page 29: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/29.jpg)
Save and load models
tf.train.Saver(…) Default will associate with all variables all_variables()
save(sess, save_path, …)
restore(sess, save_path, …) Replace initialization That’s why we need to run initialization separately
![Page 30: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/30.jpg)
Convolution
conv2d(input, filter, strides, padding, use_cudnn_on_gpu=None, name=None)
![Page 31: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/31.jpg)
LSTM
# Parameters of gates are concatenated into one multiply for efficiency.c, h = array_ops.split(1, 2, state)concat = linear([inputs, h], 4 * self._num_units, True)# i = input_gate, j = new_input, f = forget_gate, o = output_gatei, j, f, o = array_ops.split(1, 4, concat)new_c = c * sigmoid(f + self._forget_bias) + sigmoid(i) * tanh(j)new_h = tanh(new_c) * sigmoid(o)
BasicLSTMCell
![Page 32: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/32.jpg)
Word2Vec with TensorFlow
# Look up embeddings for inputs.embeddings = tf.Variable( tf.random_uniform([vocabulary_size, embedding_size], -1.0, 1.0))embed = tf.nn.embedding_lookup(embeddings, train_inputs)# Construct the variables for the NCE lossnce_weights = tf.Variable( tf.truncated_normal([vocabulary_size, embedding_size], stddev=1.0 / math.sqrt(embedding_size)))nce_biases = tf.Variable(tf.zeros([vocabulary_size]))# Compute the average NCE loss for the batch.# tf.nce_loss automatically draws a new sample of the negative labels each# time we evaluate the loss.loss = tf.reduce_mean( tf.nn.nce_loss(nce_weights, nce_biases, embed, train_labels, num_sampled, vocabulary_size))
![Page 33: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/33.jpg)
Reuse Pre-trained models
Image recognition
Inception-v3
military uniform (866): 0.647296suit (794): 0.0477196academic gown (896): 0.0232411bow tie (817): 0.0157356bolo tie (940): 0.0145024
![Page 34: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/34.jpg)
Try it on your Android
github.com/tensorflow/tensorflow/tree/master/tensorflow/examples/android
Uses a Google Inception model to classify camera frames in real-time, displaying the top results in an overlay on the camera image.
Tensorflow Android Camera Demo
![Page 35: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/35.jpg)
github.com/nivwusquorum/tensorflow-deepq
Reinforcement Learning using Tensor Flow
![Page 36: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/36.jpg)
github.com/asrivat1/DeepLearningVideoGames
Using Deep Q Networks to Learn Video Game Strategies
![Page 40: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/40.jpg)
github.com/jazzsaxmafia/show_and_tell.tensorflow
![Page 42: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/42.jpg)
Google Brain Residency Program
Learn to conduct deep learning research w/experts in our teamFixed one-year employment with salary, benefits, ...
Interesting problems, TensorFlow, and access to computational resources
Goal after one year is to have conducted several research projects
New one year immersion program in deep learning research
![Page 43: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/43.jpg)
Google Brain Residency Program
Who should apply? People with BSc, MSc or PhD, ideally in CS, mathematics or statisticsCompleted coursework in calculus, linear algebra, and probability, or equiv.
Motivated, hard working, and have a strong interest in deep learning
Programming experience
![Page 44: Tensor Flow - Deep Learning Gardendeeplearning.lipingyang.org/.../2016/...TensorFlow.pdf · Tensor Flow Tensors: n-dimensional arrays A sequence of tensor operations Deep learning](https://reader033.fdocuments.in/reader033/viewer/2022042612/5f3e0ea6f77e0128bf0d9936/html5/thumbnails/44.jpg)
Google Brain Residency Program
Program Application & Timeline
DEADLINE: January 15, 2016
Thanks for your attention!