Ict L2 PDF
Ict L2 PDF
5. One should never lose sight of how crude the approximations are, and
how over-simplified our ANNs are compared to real brains.
2
Brains versus Computers
1. There are approximately 10 billion neurons in the human cortex, compared with
10 of thousands of processors in the most powerful parallel computers.
2. Each biological neuron is connected to several thousands of other neurons,
similar to the connectivity in powerful parallel computers.
3. Lack of processing units can be compensated by speed. The typical operating
speeds of biological neurons is measured in milliseconds (10-3 s), while a silicon
chip can operate in nanoseconds (10-9 s).
4. The human brain is extremely energy efficient, using approximately 10-16 joules
per operation per second, whereas the best computers today use around 10-6
joules per operation per second.
5. Brains have been evolving for tens of millions of years, computers have been
evolving for tens of decades.
6. Brain is capable of adaptation by changing the connectivity. But computer is hard
to be adaptive.
The brain uses massively parallel computation
– »1011 neurons in the brain
– »104 connections per neuron
3
Brains versus Computers
• receives input signals generated by
Biological neuron other neurons through its
dendrites,
• integrates these signals in its body,
• then generates its own signal (a
series of electric pulses) that travel
along the axon which in turn makes
contacts with dendrites of other
neurons.
• The points of contact between
neurons are called synapses.
• Incoming impulses can be
excitatory if they cause firing, or
inhibitory if they hinder the firing
of the response.
After carrying a pulse, an axon is in a state of non-excitability for a certain time called
the refractory period.
4
Brains versus Computers
Biological neuron
5
Brains versus Computers
Brain Computation
6
Biological Neural Networks (BNN)
7
Artificial Neural Networks (ANN)
8
Definitions
Haykin :
A neural network is a massively parallel distributed processor that has a natural
propensity for storing experiential knowledge and making it available for use. It
resembles the brain in two respects:
– Knowledge is acquired by the network through a learning process.
– Interneuron connection strengths known as synaptic weights are used to store the
knowledge.
Zurada:
Artificial neural systems, or neural networks, are physical cellular systems
which can acquire, store, and utilize experiential knowledge.
9
Definitions
Mohamad H Hasssoun :
Neural Networks are neural in the sense that they may have been inspired by
neuroscience but not necessarily because they are faithful models of biologic
neural or cognitive phenomena
J.A. Anderson :
It is not absolutely necessary to believe that neural network models have
anything to do with the nervous system, but it helps. Because, we are able to use
a large body of ideas and facts from
10
Importance of ANN
* The property that enables a system to continue operating properly in the event
of the failure of some of its components.
11
Artificial Neural Net
W1
Y
X1
W2
X2
The figure shows a simple artificial neural net with two input
neurons (X1, X2) and one output neuron (Y). The inter connected
weights are given by W1 and W2.
12
Artificial Neural Net
13
Artificial Neural Net
1. A set of links, describing the neuron inputs, with weights W1, W2,
…, Wm.
x1-x2= -1
x2 x1-x2=0
x1-x2= 1
x1
15
Operation of a Neural Network
-
x0 w0j
x1 w1j
f
Output y
xn wnj
16
McCulloch-Pitts neuron model (1943)
Activation function
0
17
McCulloch-Pitts neuron model (1943)
18
Networks of McCulloch-Pitts Neurons
19
Building Blocks of Artificial Neural Net
• Activation Function
20
Network Architecture
21
Training Process
22
Activation Function
23
Activation Function
Activation functions:
(A) Identity
(F) Ramp
24
Decision Boundaries/Linear Separability
Linear separability is the concept wherein the separation of the input space into
regions is based on whether the network response is positive or negative.
The decision boundary is the surface at which the output of the unit is precisely
equal to the threshold.
x1
w1
1 w1
slope =
y w2
2 w2
w2
0 2
x2 W1=1, w2=2, 2
25
Learning and Generalization
26
Perceptron
27
Perceptron Network
28
Perceptron Network
29
Perceptron Network
30
Perceptron Learning
wi = wi + wi or wi = (t - o) xi
where
t = c(x) is the target value, o is the perceptron output, is a small
constant (e.g., 0.1) called learning rate.
If the output is correct (t = o) the weights wi are not changed
If the output is incorrect (t o) the weights wi are changed such that the
output of the perceptron for the new weights is closer to t.
32
Learning Rules
33
Learning Rules
34
Supervised Hebbian Learning
Linear Associator
35
Supervised Hebbian Learning
36
Supervised Hebbian Learning
37
Steepest Descent Method
38
Steepest Descent Method
39
Steepest Descent Method
Steepest Descent with Minimization Along a Line Trajectory for Newton’s Method
40
Conjugate Gradient
41
Widrow-Hoff Learning
ADALINE Network
42
Widrow-Hoff Learning
Mean Square Error
LMS Algorithm
Convergence Point
Stable Learning Rate
43
Widrow-Hoff Learning
Adaptive Filter ADALINE
Tapped Delay Line
44
Multilayer Perceptrons
Three-Layer Network
45
Pattern Classification
Two-Layer XOR Network
46
Function Approximation
Function Approximation Network
47
Function Approximation
Nominal Response of Network
48
Function Approximation
Effect of Parameter Changes on Network Response
49