Introduction To ANN
Introduction To ANN
University of Karachi
University of Karachi
According to Nigrin (1993), p. 11: A neural network is a circuit composed of a very large number of simple processing elements that are neurally based. Each element operates only on local information. Furthermore each element operates asynchronously; thus there is no overall system clock. According to Zurada (1992), p. xv: Artificial neural systems, or NNs, are physical cellular systems which can acquire, store, and utilize experiential knowledge.
University of Karachi
solve. But computers would be so useful if they could do things that we do not exactly know how to do. NNs process information in the similar way the human brain does. The network is composed of a large number of highly interconnected processing elements (neurons) working in parallel to perform a specific task. The examples must be selected carefully otherwise useful time is wasted or even worse the network might be functioning incorrectly. The disadvantage is that because the network finds out how to solve the problem by itself. Its operation can be unpredictable. On the other hand conventional computers use a cognitive approach to problem solving; the way the problem is to solved must be known and stated on small unambiguous instructions. Hence instructions are then converted to high-level language program and then into machine code that he computer can understand. These machines are completely predictable, if anything goes wrong is due to the software or hardware fault. NNs and conventional algorithmic computers are not on competition but complement to each other. There are tasks that are more suited to an algorithmic approach like arithmetic operations and tasks that are more suited to NNs. Even more, a large number of tasks, require systems that use a combination of the two approaches (normally a conventional computer is used to supervise the NNs) in order to perform at maximum efficiency.
University of Karachi
Dendri tes
An Artificial Neuron
Computer scientists want to find out about the properties of non-symbolic information processing with neural nets and about learning systems in general. Statisticians use neural nets as flexible, nonlinear regression and classification models. Engineers of many kinds exploit the capabilities of NNs in many areas, such as signal processing and automatic control. Cognitive scientists view NNs as a possible apparatus to describe models of thinking and consciousness (High-level brain function). Neuro-physiologists use NNs to describe and explore medium-level brain function (e.g. memory, sensory system, robotics). Physicists use NNs to model phenomena in statistical mechanics and for a lot of other tasks. Biologists use NNs to interpret nucleotide sequences. Philosophers and some other people may also be interested in NNs for various reasons.
University of Karachi
University of Karachi
13. Acoustic and seismic analysis o Ocean platform detection and classification. Seismic discrimination. Oil and water fields detection by OSA 14. Military systems Radar: Reentry vehicle trajectory prediction. Radar imagery target classification. Detection and identification of tactical targets. Radar pulse classification Infrared: Target acquisition and aim-point selection. LANDSAT scene classification Ultrasonics and acoustics emission: Ultrasonic imaging. Feedwater nozzle detection. Turbine rotor inspection. Ultrasonic pipe inspection. Monitoring of crack-growth activity Missile guidance: Equation of hyroscope systems dynamics identification using sorting of difference equations. Air-to-air guidance law synthesis
Image Processing and Computer Vision Including image matching, preprocessing, segmentation and analysis, computer vision e.g., circuit board inspection), image compression, stereo vision and processing and understanding of time varying images. Signal Processing Including seismic signal analysis and morphology. Pattern Recognition Including feature extraction, radar signal classification and analysis, speech recognition and understanding, fingerprint identification, character (letter or number) recognition, and handwriting analysis. Medical Including electrocardiograph signal analysis and understanding, diagnosis of many diseases, and medical image processing. Financial Systems Including stock market analysis, real estate appraisal, credit card authorization, and securities trading Planning, Control and Search Including parallel implementation of constraint scarification problems (CSPs), solutions to traveling salesman and control and robotics. Power Systems Including system state estimation, transient detection and classification fault detection and recovery, load forecasting, and security assessment. Human Factors Interfacing
University of Karachi
An unsupervised method can learn a summary of a probability distribution, then that summarized distribution can be used to make predictions. Furthermore, supervised methods come in two subvarieties: auto-associative and hetero-associative. In autoassociative learning, the target values are the same as the inputs, whereas in heteroassociative learning, the targets are generally different from the inputs. Many unsupervised methods are equivalent to auto-associative supervised methods.
University of Karachi
Feed backward NNs: In a feedback or recurrent NN, there are cycles in the connections. In some feedback NNs, each time an input is presented, the NN must iterate for a potentially long time before it produces a response. Feedback NNs are usually more difficult to train than feedforward NNs.
University of Karachi
Linear NNs, Multilayer perceptron, RBF networks - Bishop (1995), Moody and Darken (1989), Orr (1996), CMAC: Cerebellar Model Articulation Controller, Classification only , Regression only
b. Feedback
c. Competitive
2. Unsupervised
a. Competitive
Vector Quantization, Self-Organizing Map, Adaptive resonance theory, DCL: Differential Competitive Learning - Kosko (1992)
b. Dimension Reduction
c. Autoassociation
3. Nonlearning
Hopfield Various networks for optimization
University of Karachi
A Layer of Neurons
Input Layer of neurons
Where
a1 R =Number of elements in input vector. S = Number of neurons in layer
P1 P2 P3 PR
W1,1 b1 b2 WS,R b3
n1 n2
f
a2
f f
a3
n3 a = f (WP +b)
In this network, each element of the input vector is connected to each neuron input through the weight matrix W. The ith neuron has a summer that gathers its weighted inputs and bias to form its own scalar output n (i). The various n (i) taken together form a S-element net input vector n. Finally, the neuron layer output form a column vector a.
University of Karachi
University of Karachi
11. Transfer Function: It is common practice in Statistical regression modeling that the experimenter tries to model population using sample data points. Thus the modeling of sample data using regression analysis requires parameter estimation using any procedure. But, ANNs are adaptive systems with the power of a universal computer, i.e. they can realize an arbitrary mapping (association) of one vector space (inputs) to the other vector space (outputs). They differ in many respects, one of the most important characteristics being the transfer functions performed by each neuron. The choice of transfer functions in neural networks is of crucial importance to their performance. There is a growing understanding that the choice of transfer functions is at least as important as the network architecture and learning algorithm. Neural networks are used either to approximate a posteriori probabilities for classification or to approximate the probability densities of the training data. Viewing the problem of learning from geometrical point of view the purpose of the transfer functions performed by the neural network nodes is to enable the proper use of the parameter space in the most flexible way using the lowest number of adaptive parameters.
The activation and the output functions of the input and the output layers may be of different type than those of the hidden layers, in particular frequently linear functions are used for inputs and outputs and non-linear transfer functions for hidden layers.
The behavior of an Artificial Neural Network depends highly on the transfer function used. There are a number of transfer function that typically fall into one of the following three categories. 1. 2. 3. Linear (or ramp) Threshold Sigmoid
(Hard limit transfer function) (Positive Linear transfer function) (Linear transfer function) (Hyperbolic tangent transfer function)
University of Karachi
+1 n
0
-1
The hardlim transfer function forces a neuron to output a 1 if its net input reaches a threshold otherwise it outputs 0. This allows a neuron to make a decision or classification. It can say yes or no. This kind of neuron is often trained with the perceptron learning rule.
+1 n
0
-1
The transfer function poslin returns the output n if n is greater than equal to zero and o if n is less than equal to zero. purelin transfer function:
a
+1 n
0
-1
3.4
+1 n
0
-1
University of Karachi
tansig is named after the hyperbolic tangent, which has the same shape. However, tanh may be more accurate and is recommended for applications that require the hyperbolic tangent. tansig(N) calculates its output according to:
n = 2 / (1 + exp (-2 * n) ) - 1
This is mathematically equivalent to tanh (N).