Outline

59
Artificiel Neural Networks 2 Morten Nielsen Department of Systems Biology, DTU IIB-INTECH, UNSAM, Argentina

description

Artificiel Neural Networks 2 Morten Nielsen Department of Systems Biology , DTU IIB-INTECH, UNSAM, Argentina. Outline. Optimization procedures Gradient decent (this you already know) Network training back propagation cross-validation Over-fitting examples. - PowerPoint PPT Presentation

Transcript of Outline

Page 1: Outline

Artificiel Neural Networks 2

Morten NielsenDepartment of Systems Biology,

DTUIIB-INTECH, UNSAM, Argentina

Page 2: Outline

Outline

• Optimization procedures – Gradient decent (this you already know)

• Network training– back propagation– cross-validation– Over-fitting– examples

Page 3: Outline

Neural network. Error estimate

I1 I2

w1 w2

Linear function

o

Page 4: Outline

Neural networks

Page 5: Outline

Gradient decent (from wekipedia)

Gradient descent is based on the observation that if the real-valued function F(x) is defined and differentiable in a neighborhood of a point a, then F(x) decreases fastest if one goes from a in the direction of the negative gradient of F at a. It follows that, if

for > 0 a small enough number, then F(b)<F(a)

Page 6: Outline

Gradient decent (example)

Page 7: Outline

Gradient decent. Example

Weights are changed in the opposite direction of the gradient of the error

I1 I2

w1 w2

Linear function

o

Page 8: Outline

Network architecture

Input layer

Hidden layer

Output layer

Ik

hj

Hj

oO

vjk

wj

Page 9: Outline

What about the hidden layer?

Page 10: Outline

Hidden to output layer

Page 11: Outline

Hidden to output layer

Page 12: Outline

Hidden to output layer

Page 13: Outline

Input to hidden layer

Page 14: Outline

Input to hidden layer

Page 15: Outline

Summary

Page 16: Outline

Or

Page 17: Outline

Or

Ii=X[0][k]

Hj=X[1][j]

Oi=X[2][i]

Page 18: Outline

Can you do it your self?

v22=1v12=1

v11=1v21=-1

w1=-1 w2=1

h2H2

h1H1

oO

I1=1 I2=1

What is the output (O) from the network?What are the wij and vjk values if the target value is 0 and =0.5?

Page 19: Outline

Can you do it your self (=0.5). Has the error decreased?

v22=1v12=1

v11=1v21=-1

w1=-1 w2=1

h2=H2=

h1=H1=

o=O=

I1=1 I2=1

v22=.

v12=V11=

v21=

w1= w2=

h2=H2=

h1=H1=

o=O=

I1=1 I2=1Before After

Page 20: Outline

Can you do it your self?

v22=1v12=1

v11=1v21=-1

w1=-1 w2=1

h2=0H2=0.5

h1=2H1=0.88

o=-0.38O=0.41

I1=1 I2=1

• What is the output (O) from the network?

• What are the wij and vjk values if the target value is 0?

Page 21: Outline

Can you do it your self (=0.5). Has the error decreased?

v22=1

v12=1v11=1

v21=-1

w1=-1 w2=1

h2=0H2=0.5

h1=2H1=0.88

o=-0.38O=0.41

I1=1 I2=1

v22=.0.99v12=1.005

v11=1.005v21=-1.01

w1=-1.043w2=0.975

h2=-0.02H2=0.495

h1=2.01H1=0.882

o=-0.44O=0.39

I1=1 I2=1

Page 22: Outline

Example

Page 23: Outline

Sequence encoding

• Change in weight is linearly dependent on input value

• “True” sparse encoding is therefore highly inefficient

• Sparse is most often encoded as– +1/-1 or 0.9/0.05

Page 24: Outline

Sequence encoding - rescaling

• Rescaling the input valuesIf the input (o or h) is too large or too small, g´ is zero and the weight are not changed. Optimal performance is when o and h are close to 0.5

Page 25: Outline

Training and error reduction

Page 26: Outline

Training and error reduction

Page 27: Outline

Training and error reduction

Size matters

Page 28: Outline

Example

Page 29: Outline

Neural network training. Cross validation

Cross validation

Train on 4/5 of dataTest on 1/5=>Produce 5 different neural networks each with a different prediction focus

Page 30: Outline

Neural network training curve

Maximum test set performanceMost cable of generalizing

Page 31: Outline

5 fold training

Which network to choose?

Page 32: Outline

5 fold training

Page 33: Outline

Conventional 5 fold cross validation

Page 34: Outline

“Nested” 5 fold cross validation

Page 35: Outline

When to be careful

• When data is scarce, the performance obtained used “conventional” versus “Nested” cross validation can be very large

• When data is abundant the difference is small, and “nested” cross validation might even be higher than “conventional” cross validation due to the ensemble aspect of the “nested” cross validation approach

Page 36: Outline

Do hidden neurons matter?

• The environment matters

NetMHCpan

Page 37: Outline

Context matters• FMIDWILDA YFAMYGEKVAHTHVDTLYVRYHYYTWAVLAYTWY 0.89 A0201• FMIDWILDA YFAMYQENMAHTDANTLYIIYRDYTWVARVYRGY 0.08 A0101• DSDGSFFLY YFAMYGEKVAHTHVDTLYVRYHYYTWAVLAYTWY 0.08 A0201• DSDGSFFLY YFAMYQENMAHTDANTLYIIYRDYTWVARVYRGY 0.85 A0101

Page 38: Outline

Summary• Gradient decent is used to determine the

updates for the synapses in the neural network

• Some relatively simple math defines the gradients– Networks without hidden layers can be solved on

the back of an envelope (SMM exercise)– Hidden layers are a bit more complex, but still ok

• Always train networks using a test set to stop training– Be careful when reporting predictive performance

• Use “nested” cross-validation for small data sets• And hidden neurons do matter (sometimes)

Page 39: Outline

And some more stuff for the long cold winter nights• Can it might be made differently?

Page 40: Outline

Predicting accuracy

• Can it be made differently?

Reliability

Page 41: Outline

• Identification of position specific receptor ligand interactions by use of artificial neural network decomposition. An investigation of interactions in the MHC:peptide system

Making sense of ANN weights

Page 42: Outline

Making sense of ANN weights

Page 43: Outline

Making sense of ANN weights

Page 44: Outline

Making sense of ANN weights

Page 45: Outline

Making sense of ANN weights

Page 46: Outline

Making sense of ANN weights

Page 47: Outline

Making sense of ANN weights

Page 48: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 49: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 50: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 51: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 52: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 53: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Auto encoder

Page 54: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 55: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 56: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 57: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 58: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks

Page 59: Outline

Deep learning

http://www.slideshare.net/hammawan/deep-neural-networks