Lecture 02 - Artificial Neural Network
Lecture 02 - Artificial Neural Network
Delta rule attempts to minimize the local error and uses the
derivative of the error to find the slope of the error space in
the region local to a particular point
NEURAL NETWORKS
Delta rule:
wi = -c (Error/ wi)
Backpropagation Algorithm
wki
k
i
wki
k
i
wki
k
i
j
wij
Oi = xi
i
Design Choices:
Input Encoding
- 30 x 32 pixel image
- A coarse resolution of the 120 x 128 pixel image
- Every 4 x 4 pixels are replaced by their mean value
- The pixel intensity is linearly scaled from 0 to 1 so
that inputs, hidden units and output units have
the same range
NEURAL NETWORKS
Design Choices:
Output Encoding
- Learning Task: Direction in which person is looking
- Only one neuron could have been used with outputs
0.2, 0.4, 0.6, and 0.8 to encode the four possible
values
- But we use 4 output neurons, so that measure of
confidence in the ANN’s decision can be
obtained
- Output vector:
1 for true & 0 for false; e.g. [1, 0, 0, 0]
NEURAL NETWORKS
Design Choices:
Network Structure
- How many Layers?
Usually one hidden layer is enough
- How many units in the hidden layer
More than necessary units result in over-fitting
Less units result in failure of training
Trial & error: Start with a number and prune
the units with the help of a cross-validation set
NEURAL NETWORKS
Adding Momentum
The first term on the right of the equation is just the weight
update rule described before
NEURAL NETWORKS
Adding Momentum
Adding Momentum
Chapter 4 of T. Mitchell
http://www-2.cs.cmu.edu/afs/cs/project/ai-repository/
ai/areas/neural/systems/nevprop/np.c
Assignment
Home Work