Artificial neural networks (ANNs) are parallel distributed processing systems composed of interconnected units (neurons) that can learn. Key aspects of ANNs include processing units, activation states, weighted connections between units, a propagation rule to determine unit input, an activation function, learning rules, and an environment providing input/error signals. ANNs operate similarly to biological brains through massively parallel distributed processing and adaptive learning. Information is transmitted between neurons as electric signals (spikes) encoding frequency and phase. ANNs learn by adjusting connection weights to minimize error between actual and predicted outputs.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0 ratings0% found this document useful (0 votes)
57 views
Gaurav Ann PDF
Artificial neural networks (ANNs) are parallel distributed processing systems composed of interconnected units (neurons) that can learn. Key aspects of ANNs include processing units, activation states, weighted connections between units, a propagation rule to determine unit input, an activation function, learning rules, and an environment providing input/error signals. ANNs operate similarly to biological brains through massively parallel distributed processing and adaptive learning. Information is transmitted between neurons as electric signals (spikes) encoding frequency and phase. ANNs learn by adjusting connection weights to minimize error between actual and predicted outputs.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 75
Artificial Neural Network
Artificial Neural Network
A set of major aspects of a parallel distributed model include: a set of processing units (cells). a state of activation for every unit, which equivalent to the output of the unit. connections between the units. Generally each connection is defined by a weight. a propagation rule, which determines the effective input of a unit from its external inputs. an activation function, which determines the new level of activation based on the effective input and the current activation. an external input for each unit. a method for information gathering (the learning rule). an environment within which the system must operate, providing input signals and _ if necessary _ error signals. Why Artificial Neural Networks? There are two basic reasons why we are interested in building artificial neural networks (ANNs):
• Technical viewpoint: Some problems such as
character recognition or the prediction of future states of a system require massively parallel and adaptive processing.
• Biological viewpoint: ANNs can be used to
replicate and simulate components of the human (or animal) brain, thereby giving us insight into natural information processing. Artificial Neural Networks • The “building blocks” of neural networks are the neurons. • In technical systems, we also refer to them as units or nodes.
• Basically, each neuron
receives input from many other neurons. changes its internal state (activation) based on the current input. sends one output signal to many other neurons, possibly including its input neurons (recurrent network). Artificial Neural Networks • Information is transmitted as a series of electric impulses, so-called spikes.
• The frequency and phase of these spikes encodes the information.
• In biological systems, one neuron can be connected to as many as
10,000 other neurons.
• Usually, a neuron receives its information from other neurons in a
confined area, its so-called receptive field. How do ANNs work? How do our brains work? The Brain is A massively parallel information processing system. Our brains are a huge network of processing elements. A typical brain contains a network of 10 billion neurons. Dendrites: Input Cell body: Processor Synaptic: Link Axon: Output How do ANNs work? • Now, let us have a look at the model of an artificial neuron. Transfer Functions Linear: The output is proportional to the total weighted input. Threshold: The output is set at one of two values, depending on whether the total weighted input is greater than or less than some threshold value. Non‐linear: The output varies continuously but not linearly as the input changes. Error Estimation The root mean square error (RMSE) is a frequently- used measure of the differences between values predicted by a model or an estimator and the values actually observed from the thing being modeled or estimated Weights Adjusting After each iteration, weights should be adjusted to minimize the error. – All possible weights – Back propagation Back Propagation Back-propagation is an example of supervised learning is used at each layer to minimize the error between the layer’s response and the actual data The error at each hidden layer is an average of the evaluated error Hidden layer networks are trained this way Learning Paradigms Supervised learning Unsupervised learning Reinforcement learning Supervised learning This is what we have seen so far! A network is fed with a set of training samples (inputs and corresponding output), and it uses these samples to learn the general relationship between the inputs and the outputs. This relationship is represented by the values of the weights of the trained network. Unsupervised learning No desired output is associated with the training data! Faster than supervised learning Used to find out structures within data: Clustering Compression Reinforcement learning Like supervised learning, but: Weights adjusting is not directly related to the error value. The error value is used to randomly, shuffle weights! Relatively slow learning due to ‘randomness’. Applications Areas Function approximation including time series prediction and modeling. Classification including patterns and sequences recognition, novelty detection and sequential decision making. (radar systems, face identification, handwritten text recognition)
Data processing including filtering, clustering blinds source separation and compression. (data mining, e-mail Spam filtering) HOW ANN WORKS HOW DO NN LEARN GRADIENT DESCENT