NNDL Lecture 1
NNDL Lecture 1
Learning
Dendrites: Input
Cell body: Processor
Synaptic: Link
Axon: Output
How do our brains work?
A processing element
The axon endings almost touch the dendrites or cell body of the
next neuron.
How do our brains work?
A processing element
Processing ∑
∑= X1+X2 + ….+Xm =y
Output y
How do ANNs work?
Not all inputs are equal
xm ........... x2 x1
.
Input
wm ....
weights .
w2 w1
Processing ∑
∑= X1w1+X2w2 + ….+Xmwm
=y
Output y
How do ANNs work?
The signal is not passed down to the
next neuron verbatim
xm ........... x2 x1
.
Input
wm .... w2 w1
weights .
Processing ∑
Transfer Function
f(vk)
(Activation Function)
Output y
The output is a function of the input, that is
affected by the weights, and the transfer functions
Artificial Neural Networks
An ANN can:
1. compute any computable function, by the appropriate
selection of the network topology and weights values.
2. learn from experience!
Specifically, by trial‐and‐error
Learning by trial‐and‐error
Many neurons:
Higher accuracy
Slower
Risk of over‐fitting
Memorizing, rather than understanding
The network will be useless with new problems.
Few neurons:
Lower accuracy
Inability to learn at all
Optimal number.
Data representation
Usually input/output data needs pre ‐processing
Pictures
Pixel intensity
Text:
A pattern
Size of training set
No one‐fits‐all formula
Over fitting can occur if a “good” training set is not
chosen
What constitutes a “good” training set?
Samples must represent the general population.
Samples must contain members of each class.
Samples in each class must contain a wide range of
variations or noise effect.
The size of the training set is related to the number of
hidden neurons
Learning Paradigms
Supervised learning
Unsupervised learning
Reinforcement learning
Supervised learning
This is what we have seen so far!
A network is fed with a set of training samples (inputs
and corresponding output), and it uses these samples
to learn the general relationship between the inputs
and the outputs.
This relationship is represented by the values of the
weights of the trained network.
Unsupervised learning
No desired output is associated with the training
data!
Faster than supervised learning
Used to find out structures within data:
Clustering
Compression
Reinforcement learning
Like supervised learning, but:
Weights adjusting is not directly related to the error
value.
The error value is used to randomly, shuffle weights!
Relatively slow learning due to ‘randomness’.
Applications Areas
Function approximation
including time series prediction and modeling.
Classification
including patterns and sequences recognition, novelty
detection and sequential decision making.
(radar systems, face identification, handwritten text recognition)
Data processing
including filtering, clustering blinds source separation and
compression.
(data mining, e-mail Spam filtering)
Advantages / Disadvantages
Advantages
Adapt to unknown situations
Powerful, it can model complex functions.
Ease of use, learns by example, and very little user
domain‐specific expertise needed
Disadvantages
Forgets
Not exact
Large complexity of the network structure
Conclusion
Artificial Neural Networks are an imitation of the biological
neural networks, but much simpler ones.
The computing would have a lot to gain from neural networks.
Their ability to learn by example makes them very flexible and
powerful furthermore there is need to device an algorithm in
order to perform a specific task.
Conclusion
Neural networks also contributes to area of research such a
neurology and psychology. They are regularly used to model
parts of living organizations and to investigate the internal
mechanisms of the brain.
Many factors affect the performance of ANNs, such as the
transfer functions, size of training sample, network topology,
weights adjusting algorithm, …
Thank You