0% found this document useful (0 votes)
57 views

Gaurav Ann PDF

Artificial neural networks (ANNs) are parallel distributed processing systems composed of interconnected units (neurons) that can learn. Key aspects of ANNs include processing units, activation states, weighted connections between units, a propagation rule to determine unit input, an activation function, learning rules, and an environment providing input/error signals. ANNs operate similarly to biological brains through massively parallel distributed processing and adaptive learning. Information is transmitted between neurons as electric signals (spikes) encoding frequency and phase. ANNs learn by adjusting connection weights to minimize error between actual and predicted outputs.

Uploaded by

Jagdeep Pandey
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
57 views

Gaurav Ann PDF

Artificial neural networks (ANNs) are parallel distributed processing systems composed of interconnected units (neurons) that can learn. Key aspects of ANNs include processing units, activation states, weighted connections between units, a propagation rule to determine unit input, an activation function, learning rules, and an environment providing input/error signals. ANNs operate similarly to biological brains through massively parallel distributed processing and adaptive learning. Information is transmitted between neurons as electric signals (spikes) encoding frequency and phase. ANNs learn by adjusting connection weights to minimize error between actual and predicted outputs.

Uploaded by

Jagdeep Pandey
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 75

Artificial Neural Network

Artificial Neural Network


 A set of major aspects of a parallel distributed model include:
 a set of processing units (cells).
 a state of activation for every unit, which equivalent to the output of the unit.
 connections between the units. Generally each connection is defined by a weight.
 a propagation rule, which determines the effective input of a unit from its external
inputs.
 an activation function, which determines the new level of activation based on the
effective input and the current activation.
 an external input for each unit.
 a method for information gathering (the learning rule).
 an environment within which the system must operate, providing input signals and
_ if necessary _ error signals.
Why Artificial Neural Networks?
 There are two basic reasons why we are interested in building artificial neural
networks (ANNs):

• Technical viewpoint: Some problems such as


character recognition or the prediction of future
states of a system require massively parallel and
adaptive processing.

• Biological viewpoint: ANNs can be used to


replicate and simulate components of the human
(or animal) brain, thereby giving us insight into
natural information processing.
Artificial Neural Networks
• The “building blocks” of neural networks are the neurons.
• In technical systems, we also refer to them as units or nodes.

• Basically, each neuron


 receives input from many other neurons.
 changes its internal state (activation) based on the current input.
 sends one output signal to many other neurons, possibly including its input
neurons (recurrent network).
Artificial Neural Networks
• Information is transmitted as a series of electric impulses, so-called
spikes.

• The frequency and phase of these spikes encodes the information.

• In biological systems, one neuron can be connected to as many as


10,000 other neurons.

• Usually, a neuron receives its information from other neurons in a


confined area, its so-called receptive field.
How do ANNs work?
How do our brains work?
The Brain is A massively parallel information
processing system.
Our brains are a huge network of processing
elements. A typical brain contains a network
of 10 billion neurons.
Dendrites: Input
Cell body: Processor
Synaptic: Link
Axon: Output
How do ANNs work?
• Now, let us have a look at the model of an artificial neuron.
Transfer Functions
 Linear: The output is proportional to the total weighted
input.
 Threshold: The output is set at one of two values, depending
on whether the total weighted input is greater than or less
than some threshold value.
 Non‐linear: The output varies continuously but not linearly
as the input changes.
Error Estimation
 The root mean square error (RMSE) is a frequently-
used measure of the differences between values
predicted by a model or an estimator and the values
actually observed from the thing being modeled or
estimated
Weights Adjusting
 After each iteration, weights should be adjusted to
minimize the error.
– All possible weights
– Back propagation
Back Propagation
 Back-propagation is an example of supervised learning
is used at each layer to minimize the error between
the layer’s response and the actual data
 The error at each hidden layer is an average of the
evaluated error
 Hidden layer networks are trained this way
Learning Paradigms
 Supervised learning
 Unsupervised learning
 Reinforcement learning
Supervised learning
 This is what we have seen so far!
 A network is fed with a set of training samples (inputs and
corresponding output), and it uses these samples to learn the general
relationship between the inputs and the outputs.
 This relationship is represented by the values of the weights of the
trained network.
Unsupervised learning
 No desired output is associated with the training data!
 Faster than supervised learning
 Used to find out structures within data:
 Clustering
 Compression
Reinforcement learning
 Like supervised learning, but:
 Weights adjusting is not directly related to the error value.
 The error value is used to randomly, shuffle weights!
 Relatively slow learning due to ‘randomness’.
Applications Areas
 Function approximation
 including time series prediction and modeling.
 Classification
 including patterns and sequences recognition, novelty detection and
sequential decision making.
 (radar systems, face identification, handwritten text recognition)

 Data processing
 including filtering, clustering blinds source separation and compression.
 (data mining, e-mail Spam filtering)
HOW ANN WORKS
HOW DO NN LEARN
GRADIENT DESCENT

You might also like