Artificial Neural Networks: Introduction To Computational Neuroscience
Artificial Neural Networks: Introduction To Computational Neuroscience
Machine translation
Learning objectives
How artificial neural networks work?
x1 w1
w2 Σ
x2 z
b
1
Let’s try it out!
x1 x2 y = x1 or x2
0 0 0
0 1 1
1 0 1
1 1 1
Algorithm: repeat
1, if x1w1 x2 w2 b 0
z
0, otherwise
w1 w1 ( y z ) x1
w2 w2 ( y z ) x2
b b ( y z)
until y=z holds for entire dataset
Perceptron limitations
+1 b1
+1 c
b2 a1 x1w11 x2 w21 b1
h1 (a1 ) z h1v1 h2v2 c
w11 v1
x1 Σ Σ
w12
a2 x1w12 x2 w22 b2
w21 h2 (a2 )
1
( x)
x2
w22
Σ v2 1 e x
Loss function
1
• Function approximation: L ( z y)2
2
( z 10) 2
L L z h2 a2
b2 z h2 a2 b2 L L z h1 L
( z y)v1h1 (1 h1 ) zy
a1 z h1 a1 z
L L z h1 a1
w11 z h1 a1 w11
x1 Σ L L z Σ
L L z h2 a2 ( z y )h1
v1 z v1
w12 z h2 a2 w12
L L z h2
L L z h1 a1 ( z y)v2 h2 (1 h2 )
a2 z h2 a2
w21 z h1 a1 w21
x2 Σ L L z
L L z h2 a2 ( z y )h2
v2 z v2
w22 z h2 a2 w22
{wij , v j , b j , c}
L
learning rate
p ( z)
L y log( p) (1 y) log(1 p)
log(1 p)
• Multi-class classification:
e zi
p softmax( z ), pi
e
zj
j 1
( x)
1 e x
L yi log pi log pk
i
Things to remember...
Perceptron was the first artificial neuron model
invented in late 1950s.
Perceptron can learn only linearly separable
classification problems.
Feed-forward networks with non-linear activation
functions and hidden layers can overcome
limitations of perceptrons.
Multi-layer artificial neural networks are trained
using backpropagation and gradient descent.
Part 2
weights: 1 0 -1
Hubel & Wiesel (1959)
• Performed experiments
with anesthetized cat.
• Discovered topographical
mapping, sensitivity to
orientation and
hierarchical processing.
Convolution
https://developer.apple.com/library/ios/documentation/Performance/Conceptual/vImage/ConvolutionOperations/ConvolutionOperations.html
Pooling
– speech recognition,
– machine translation,
– language modeling,
– any time series. INPUT LAYER
Backpropagation through time
y1 y2 y3 y4
1
L ( z y)2
2
OUTPUT LAYER z1 z2 z3 z4
HIDDEN LAYER h0 h1 h2 h3 h4
INPUT LAYER x1 x2 x3 x4
time
Different configurations
Autoencoders
• Architecture: OUTPUT LAYER = INPUT LAYER
– Input and output layers
are the same.
– Hidden layer functions
as a “bottleneck”.
– Network is trained to HIDDEN LAYER
reconstruct input from
hidden layer activations.
• Used for:
– image semantic hashing
– dimensionality reduction INPUT LAYER
We didn’t talk about...
• Long Short Term Memory (LSTMs)
• Restricted Boltzmann Machines (RBMs)
• Echo State Networks / Liquid State Machines
• Hopfield Network
• Self-organizing maps (SOMs)
• Radial basis function networks (RBFs)
https://www.youtube.com/watch?v=OOT3UIXZztE
https://github.com/matterport/Mask_RCNN
Image Captioning
Image Captioning Errors
Reinforcement
screen learning
score
https://www.youtube.com/watch?v=NhxCg2PA3ZI
Adversarial Examples
https://www.youtube.com/watch?v=XaQu7kkQBPc
Things to remember...
Artificial neural networks are state-of-the-art
in image recognition, speech recognition,
machine translation and many other fields.
Anything that you can do in 1 second,
probably we can train a neural network to do
the same, i.e. neural nets can do perception.
But in the end they are just reactive function
approximators and can be easily fooled. In
particular they do not think like humans (yet).
Thank you!