Artificial neural network for machine learninggrinu
An Artificial Neurol Network (ANN) is a computational model. It is based on the structure and functions of biological neural networks. It works like the way human brain processes information. ANN includes a large number of connected processing units that work together to process information. They also generate meaningful results from it.
1. A perceptron is a basic artificial neural network that can learn linearly separable patterns. It takes weighted inputs, applies an activation function, and outputs a single binary value.
2. Multilayer perceptrons can learn non-linear patterns by using multiple layers of perceptrons with weighted connections between them. They were developed to overcome limitations of single-layer perceptrons.
3. Perceptrons are trained using an error-correction learning rule called the delta rule or the least mean squares algorithm. Weights are adjusted to minimize the error between the actual and target outputs.
The document provides an overview of artificial neural networks (ANNs). It discusses how ANNs are modeled after biological neural networks and neurons. The key concepts covered include the basic structure and functioning of artificial neurons, different types of learning in ANNs, commonly used network architectures, and applications of ANNs. Examples of applications discussed are classification, recognition, assessment, forecasting and prediction. The document also notes how ANNs are used across various fields including computer science, statistics, engineering, cognitive science, neurophysiology, physics and biology.
The document provides an overview of artificial neural networks and their learning capabilities. It discusses:
- How biological neural networks in the brain inspired artificial neural networks
- The basic structure of artificial neurons and how they are connected in a network
- Single layer perceptrons and how they can be trained to learn simple tasks using supervised learning algorithms like the perceptron learning rule
- Multilayer neural networks with one or more hidden layers that can learn more complex patterns using backpropagation to modify weights.
The document discusses artificial neural networks. It describes their basic structure and components, including dendrites that receive input signals, a soma that processes the inputs, and an axon that transmits output signals. It also explains how neurons are connected at synapses to transfer signals between neurons. Finally, it mentions different types of activation functions that can be used in neural networks.
Artificial Intelligence: Artificial Neural NetworksThe Integral Worm
This document summarizes artificial neural networks (ANN), which were inspired by biological neural networks in the human brain. ANNs consist of interconnected computational units that emulate neurons and pass signals to other units through connections with variable weights. ANNs are arranged in layers and learn by modifying the weights between units based on input and output data to minimize error. Common ANN algorithms include backpropagation for supervised learning to predict outputs from inputs.
This document discusses classifying handwritten digits using the MNIST dataset with a simple linear machine learning model. It begins by introducing the MNIST dataset of images and corresponding labels. It then discusses using a linear model with weights and biases to make predictions for each image. The weights represent a filter to distinguish digits. The model is trained using gradient descent to minimize the cross-entropy cost function by adjusting the weights and biases based on batches of training data. The goal is to improve the model's ability to correctly classify handwritten digit images.
Introduction Of Artificial neural networkNagarajan
The document summarizes different types of artificial neural networks including their structure, learning paradigms, and learning rules. It discusses artificial neural networks (ANN), their advantages, and major learning paradigms - supervised, unsupervised, and reinforcement learning. It also explains different mathematical synaptic modification rules like backpropagation of error, correlative Hebbian, and temporally-asymmetric Hebbian learning rules. Specific learning rules discussed include the delta rule, the pattern associator, and the Hebb rule.
The document provides an introduction to neural networks, including:
- Biological neural networks transmit signals via neurons connected by synapses and axons.
- Artificial neural networks are composed of simple processing elements (neurons) that operate in parallel and are determined by network structure and connection strengths (weights).
- Multilayer neural networks consist of an input layer, hidden layers, and output layer connected by weights to solve complex problems. Learning involves updating weights so the network can efficiently perform tasks.
The document describes biological neurons and artificial neural networks. Biological neurons receive and sum signals through dendrites, and if the summation crosses a threshold, the neuron fires a signal through its axon. Learning is stored in the connections (synapses) between neurons. An artificial neuron model can implement logical functions like AND and OR by setting weights and thresholds. Multilayer perceptrons are needed to model non-linear problems. Training involves propagating error backwards to adjust weights to minimize error using the backpropagation algorithm.
This document provides an introduction to neural networks. It discusses how neural networks have recently achieved state-of-the-art results in areas like image and speech recognition and how they were able to beat a human player at the game of Go. It then provides a brief history of neural networks, from the early perceptron model to today's deep learning approaches. It notes how neural networks can automatically learn features from data rather than requiring handcrafted features. The document concludes with an overview of commonly used neural network components and libraries for building neural networks today.
Artificial neural networks are a form of artificial intelligence inspired by biological neural networks. They are composed of interconnected processing units that can learn patterns from data through training. Neural networks are well-suited for tasks like pattern recognition, classification, and prediction. They learn by example without being explicitly programmed, similarly to how the human brain learns.
This document provides an overview of associative memories and discrete Hopfield networks. It begins with introductions to basic concepts like autoassociative and heteroassociative memory. It then describes linear associative memory, which uses a Hebbian learning rule to form associations between input-output patterns. Next, it covers Hopfield's autoassociative memory, a recurrent neural network for associating patterns to themselves. Finally, it discusses performance analysis of recurrent autoassociative memories. The document presents key concepts in associative memory theory and different models like linear associative memory and Hopfield networks.
Part 1 of the Deep Learning Fundamentals Series, this session discusses the use cases and scenarios surrounding Deep Learning and AI; reviews the fundamentals of artificial neural networks (ANNs) and perceptrons; discuss the basics around optimization beginning with the cost function, gradient descent, and backpropagation; and activation functions (including Sigmoid, TanH, and ReLU). The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
Neural networks are inspired by biological neural networks and are composed of interconnected processing elements called neurons. Neural networks can learn complex patterns and relationships through a learning process without being explicitly programmed. They are widely used for applications like pattern recognition, classification, forecasting and more. The document discusses neural network concepts like architecture, learning methods, activation functions and applications. It provides examples of biological and artificial neurons and compares their characteristics.
The document provides an overview of artificial neural networks and supervised learning techniques. It discusses the biological inspiration for neural networks from neurons in the brain. Single-layer perceptrons and multilayer backpropagation networks are described for classification tasks. Methods to accelerate learning such as momentum and adaptive learning rates are also summarized. Finally, it briefly introduces recurrent neural networks like the Hopfield network for associative memory applications.
This document provides an overview of neural networks. It discusses how the human brain works and how artificial neural networks are modeled after the human brain. The key components of a neural network are neurons which are connected and can be trained. Neural networks can perform tasks like pattern recognition through a learning process that adjusts the connections between neurons. The document outlines different types of neural network architectures and training methods, such as backpropagation, to configure neural networks for specific applications.
Introduction to Adaptive Resonance Theory (ART) neural networks including:
Introduction (Stability-Plasticity Dilemma)
ART Network
ART Types
Basic ART network Architecture
ART Algorithm and Learning
ART Computational Example
ART Application
Conclusion
Main References
The document discusses the perceptron, which is a single processing unit of a neural network that was first proposed by Rosenblatt in 1958. A perceptron uses a step function to classify its input into one of two categories, returning +1 if the weighted sum of inputs is greater than or equal to 0 and -1 otherwise. It operates as a linear threshold unit and can be used for binary classification of linearly separable data, though it cannot model nonlinear functions like XOR. The document also outlines the single layer perceptron learning algorithm.
Deep Learning Tutorial | Deep Learning TensorFlow | Deep Learning With Neural...Simplilearn
This Deep Learning presentation will help you in understanding what is Deep Learning, why do we need Deep learning, what is neural network, applications of Deep Learning, what is perceptron, implementing logic gates using perceptron, types of neural networks. At the end of the video, you will get introduced to TensorFlow along with a usecase implementation on recognizing hand-written digits. Deep Learning is inspired by the integral function of the human brain specific to artificial neural networks. These networks, which represent the decision-making process of the brain, use complex algorithms that process data in a non-linear way, learning in an unsupervised manner to make choices based on the input. Deep Learning, on the other hand, uses advanced computing power and special type of neural networks and applies them to large amounts of data to learn, understand, and identify complicated patterns. W will also understand neural networks and how they work in this Deep Learning tutorial video. This Deep Learning tutorial is ideal for professionals with beginner to intermediate level of experience. Now, let us dive deep into this topic and understand what Deep Learning actually is.
Below topics are explained in this Deep Learning presentation:
1. What is Deep Learning?
2. Why do we need Deep Learning?
3. What is Neural network?
4. What is Perceptron?
5. Implementing logic gates using Perceptron
6. Types of Neural networks
7. Applications of Deep Learning
8. Working of Neural network
9. Introduction to TensorFlow
10. Use case implementation using TensorFlow
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Artificial Neural Network Lect4 : Single Layer Perceptron ClassifiersMohammed Bennamoun
This document provides an overview of single layer perceptrons (SLPs) and classification. It defines a perceptron as the simplest form of neural network consisting of adjustable weights and a bias. SLPs can perform binary classification of linearly separable patterns by adjusting weights during training. The document outlines limitations of SLPs, including their inability to represent non-linearly separable functions like XOR. It introduces Bayesian decision theory and how it can be used for optimal classification by comparing posterior probabilities given prior probabilities and likelihood functions. Decision boundaries are defined for dividing a feature space into non-overlapping regions to classify patterns.
The document discusses the concepts of soft computing and artificial neural networks. It defines soft computing as an emerging approach to computing that parallels the human mind in dealing with uncertainty and imprecision. Soft computing consists of fuzzy logic, neural networks, and genetic algorithms. Neural networks are simplified models of biological neurons that can learn from examples to solve problems. They are composed of interconnected processing units, learn via training, and can perform tasks like pattern recognition. The document outlines the basic components and learning methods of artificial neural networks.
This document discusses kernel methods and radial basis function (RBF) networks. It begins with an introduction and overview of Cover's theory of separability of patterns. It then revisits the XOR problem and shows how it can be solved using Gaussian hidden functions. The interpolation problem is explained and how RBF networks can perform strict interpolation through a set of training data points. Radial basis functions that satisfy Micchelli's theorem allowing for a nonsingular interpolation matrix are presented. Finally, the structure and training of RBF networks using k-means clustering and recursive least squares estimation is covered.
The document discusses various neural network learning rules:
1. Error correction learning rule (delta rule) adapts weights based on the error between the actual and desired output.
2. Memory-based learning stores all training examples and classifies new inputs based on similarity to nearby examples (e.g. k-nearest neighbors).
3. Hebbian learning increases weights of simultaneously active neuron connections and decreases others, allowing patterns to emerge from correlations in inputs over time.
4. Competitive learning (winner-take-all) adapts the weights of the neuron most active for a given input, allowing unsupervised clustering of similar inputs across neurons.
The document provides an overview of perceptrons and neural networks. It discusses how neural networks are modeled after the human brain and consist of interconnected artificial neurons. The key aspects covered include the McCulloch-Pitts neuron model, Rosenblatt's perceptron, different types of learning (supervised, unsupervised, reinforcement), the backpropagation algorithm, and applications of neural networks such as pattern recognition and machine translation.
The document discusses fundamentals of neural networks and artificial intelligence. It provides an overview of topics covered in lectures 37 and 38, including the biological neuron model, artificial neuron model, neural network architectures, learning methods in neural networks, single-layer neural network systems, and applications of neural networks. It also includes details on the McCulloch-Pitts neuron model and the basic elements of an artificial neuron, such as weights, thresholds, and activation functions.
The document provides an introduction to neural networks, including:
- Biological neural networks transmit signals via neurons connected by synapses and axons.
- Artificial neural networks are composed of simple processing elements (neurons) that operate in parallel and are determined by network structure and connection strengths (weights).
- Multilayer neural networks consist of an input layer, hidden layers, and output layer connected by weights to solve complex problems. Learning involves updating weights so the network can efficiently perform tasks.
The document describes biological neurons and artificial neural networks. Biological neurons receive and sum signals through dendrites, and if the summation crosses a threshold, the neuron fires a signal through its axon. Learning is stored in the connections (synapses) between neurons. An artificial neuron model can implement logical functions like AND and OR by setting weights and thresholds. Multilayer perceptrons are needed to model non-linear problems. Training involves propagating error backwards to adjust weights to minimize error using the backpropagation algorithm.
This document provides an introduction to neural networks. It discusses how neural networks have recently achieved state-of-the-art results in areas like image and speech recognition and how they were able to beat a human player at the game of Go. It then provides a brief history of neural networks, from the early perceptron model to today's deep learning approaches. It notes how neural networks can automatically learn features from data rather than requiring handcrafted features. The document concludes with an overview of commonly used neural network components and libraries for building neural networks today.
Artificial neural networks are a form of artificial intelligence inspired by biological neural networks. They are composed of interconnected processing units that can learn patterns from data through training. Neural networks are well-suited for tasks like pattern recognition, classification, and prediction. They learn by example without being explicitly programmed, similarly to how the human brain learns.
This document provides an overview of associative memories and discrete Hopfield networks. It begins with introductions to basic concepts like autoassociative and heteroassociative memory. It then describes linear associative memory, which uses a Hebbian learning rule to form associations between input-output patterns. Next, it covers Hopfield's autoassociative memory, a recurrent neural network for associating patterns to themselves. Finally, it discusses performance analysis of recurrent autoassociative memories. The document presents key concepts in associative memory theory and different models like linear associative memory and Hopfield networks.
Part 1 of the Deep Learning Fundamentals Series, this session discusses the use cases and scenarios surrounding Deep Learning and AI; reviews the fundamentals of artificial neural networks (ANNs) and perceptrons; discuss the basics around optimization beginning with the cost function, gradient descent, and backpropagation; and activation functions (including Sigmoid, TanH, and ReLU). The demos included in these slides are running on Keras with TensorFlow backend on Databricks.
Neural networks are inspired by biological neural networks and are composed of interconnected processing elements called neurons. Neural networks can learn complex patterns and relationships through a learning process without being explicitly programmed. They are widely used for applications like pattern recognition, classification, forecasting and more. The document discusses neural network concepts like architecture, learning methods, activation functions and applications. It provides examples of biological and artificial neurons and compares their characteristics.
The document provides an overview of artificial neural networks and supervised learning techniques. It discusses the biological inspiration for neural networks from neurons in the brain. Single-layer perceptrons and multilayer backpropagation networks are described for classification tasks. Methods to accelerate learning such as momentum and adaptive learning rates are also summarized. Finally, it briefly introduces recurrent neural networks like the Hopfield network for associative memory applications.
This document provides an overview of neural networks. It discusses how the human brain works and how artificial neural networks are modeled after the human brain. The key components of a neural network are neurons which are connected and can be trained. Neural networks can perform tasks like pattern recognition through a learning process that adjusts the connections between neurons. The document outlines different types of neural network architectures and training methods, such as backpropagation, to configure neural networks for specific applications.
Introduction to Adaptive Resonance Theory (ART) neural networks including:
Introduction (Stability-Plasticity Dilemma)
ART Network
ART Types
Basic ART network Architecture
ART Algorithm and Learning
ART Computational Example
ART Application
Conclusion
Main References
The document discusses the perceptron, which is a single processing unit of a neural network that was first proposed by Rosenblatt in 1958. A perceptron uses a step function to classify its input into one of two categories, returning +1 if the weighted sum of inputs is greater than or equal to 0 and -1 otherwise. It operates as a linear threshold unit and can be used for binary classification of linearly separable data, though it cannot model nonlinear functions like XOR. The document also outlines the single layer perceptron learning algorithm.
Deep Learning Tutorial | Deep Learning TensorFlow | Deep Learning With Neural...Simplilearn
This Deep Learning presentation will help you in understanding what is Deep Learning, why do we need Deep learning, what is neural network, applications of Deep Learning, what is perceptron, implementing logic gates using perceptron, types of neural networks. At the end of the video, you will get introduced to TensorFlow along with a usecase implementation on recognizing hand-written digits. Deep Learning is inspired by the integral function of the human brain specific to artificial neural networks. These networks, which represent the decision-making process of the brain, use complex algorithms that process data in a non-linear way, learning in an unsupervised manner to make choices based on the input. Deep Learning, on the other hand, uses advanced computing power and special type of neural networks and applies them to large amounts of data to learn, understand, and identify complicated patterns. W will also understand neural networks and how they work in this Deep Learning tutorial video. This Deep Learning tutorial is ideal for professionals with beginner to intermediate level of experience. Now, let us dive deep into this topic and understand what Deep Learning actually is.
Below topics are explained in this Deep Learning presentation:
1. What is Deep Learning?
2. Why do we need Deep Learning?
3. What is Neural network?
4. What is Perceptron?
5. Implementing logic gates using Perceptron
6. Types of Neural networks
7. Applications of Deep Learning
8. Working of Neural network
9. Introduction to TensorFlow
10. Use case implementation using TensorFlow
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Artificial Neural Network Lect4 : Single Layer Perceptron ClassifiersMohammed Bennamoun
This document provides an overview of single layer perceptrons (SLPs) and classification. It defines a perceptron as the simplest form of neural network consisting of adjustable weights and a bias. SLPs can perform binary classification of linearly separable patterns by adjusting weights during training. The document outlines limitations of SLPs, including their inability to represent non-linearly separable functions like XOR. It introduces Bayesian decision theory and how it can be used for optimal classification by comparing posterior probabilities given prior probabilities and likelihood functions. Decision boundaries are defined for dividing a feature space into non-overlapping regions to classify patterns.
The document discusses the concepts of soft computing and artificial neural networks. It defines soft computing as an emerging approach to computing that parallels the human mind in dealing with uncertainty and imprecision. Soft computing consists of fuzzy logic, neural networks, and genetic algorithms. Neural networks are simplified models of biological neurons that can learn from examples to solve problems. They are composed of interconnected processing units, learn via training, and can perform tasks like pattern recognition. The document outlines the basic components and learning methods of artificial neural networks.
This document discusses kernel methods and radial basis function (RBF) networks. It begins with an introduction and overview of Cover's theory of separability of patterns. It then revisits the XOR problem and shows how it can be solved using Gaussian hidden functions. The interpolation problem is explained and how RBF networks can perform strict interpolation through a set of training data points. Radial basis functions that satisfy Micchelli's theorem allowing for a nonsingular interpolation matrix are presented. Finally, the structure and training of RBF networks using k-means clustering and recursive least squares estimation is covered.
The document discusses various neural network learning rules:
1. Error correction learning rule (delta rule) adapts weights based on the error between the actual and desired output.
2. Memory-based learning stores all training examples and classifies new inputs based on similarity to nearby examples (e.g. k-nearest neighbors).
3. Hebbian learning increases weights of simultaneously active neuron connections and decreases others, allowing patterns to emerge from correlations in inputs over time.
4. Competitive learning (winner-take-all) adapts the weights of the neuron most active for a given input, allowing unsupervised clustering of similar inputs across neurons.
The document provides an overview of perceptrons and neural networks. It discusses how neural networks are modeled after the human brain and consist of interconnected artificial neurons. The key aspects covered include the McCulloch-Pitts neuron model, Rosenblatt's perceptron, different types of learning (supervised, unsupervised, reinforcement), the backpropagation algorithm, and applications of neural networks such as pattern recognition and machine translation.
The document discusses fundamentals of neural networks and artificial intelligence. It provides an overview of topics covered in lectures 37 and 38, including the biological neuron model, artificial neuron model, neural network architectures, learning methods in neural networks, single-layer neural network systems, and applications of neural networks. It also includes details on the McCulloch-Pitts neuron model and the basic elements of an artificial neuron, such as weights, thresholds, and activation functions.
This document provides an introduction to artificial neural networks (ANN). It discusses how ANN are inspired by biological neural networks in the brain and consist of interconnected artificial neurons that can learn patterns from data. The document covers the basic components of ANN including the neuron model, network architecture, learning rules, and training processes. It also compares key aspects of ANN and biological neural networks such as parallelism, learning, fault tolerance and information storage.
The document provides an overview of artificial neural networks and biological neural networks. It discusses the components and functions of the human nervous system including the central nervous system made up of the brain and spinal cord, as well as the peripheral nervous system. The four main parts of the brain - cerebrum, cerebellum, diencephalon, and brainstem - are described along with their roles in processing sensory information and controlling bodily functions. A brief history of artificial neural networks is also presented.
This document provides an overview of artificial neural networks (ANNs). It discusses the history of ANNs beginning in the 1940s and important developments like the perceptron in 1957 and backpropagation algorithms in the 1970s and 1980s. The document defines ANNs as consisting of interconnected processing units (neurons) that communicate by sending signals to each other via weighted connections, and learns from experience through training. It also compares ANNs to the human brain in using a highly parallel and distributed approach to problem solving.
1) Artificial neural networks (ANNs) are processing systems inspired by biological neural networks, consisting of interconnected nodes that process information via algorithms or hardware components. ANNs can accurately model functions like visual processing in the retina.
2) ANNs are useful for problems like facial recognition that are difficult to solve with algorithms due to their ability to learn from examples in a way similar to the human brain.
3) ANNs have many applications, including pattern recognition, modeling complex relationships in large datasets, and real-time systems due to their parallel architecture.
The document provides an overview of neural networks including:
- Their history from early models in the 1940s to the breakthrough of backpropagation in the 1980s.
- What a neural network is and how it works at the level of individual neurons and when connected together.
- Common applications of neural networks like prediction, classification, and clustering.
- Key considerations in choosing an appropriate neural network architecture and training data for a given problem.
The document discusses artificial neural networks (ANNs). It defines ANNs as computational models inspired by biological neural networks. The basic structure and types of ANNs are explained, including feed forward and feedback networks. The document also covers ANN learning methods like supervised, unsupervised, and reinforcement learning. Applications of ANNs span various domains like aerospace, automotive, military, electronics, and more. While ANNs can perform complex tasks, they require extensive training and processing power for large networks.
The document discusses the history and concepts of artificial neural networks. It provides an overview of the key topics to be covered, including the history of ANNs dating back to the 1940s and important developments like the perceptron in 1957 and backpropagation in 1974. The document defines an artificial neuron and its basic components. It also compares biological neurons to artificial neurons and outlines some of the major differences between biological and artificial neural networks.
Artificial neural networks are fundamental means for providing an attempt at modelling the information
processing capabilities of artificial nervous system which plays an important role in the field of cognitive
science. This paper focuses the features of artificial neural networks studied by reviewing the existing research
works, these features were then assessed and evaluated and comparative analysis. The study and literature
survey metrics such as functional capabilities of neurons, learning capabilities, style of computation, processing
elements, processing speed, connections, strength, information storage, information transmission,
communication media selection, signal transduction and fault tolerance were used as basis for comparison. A
major finding in this paper showed that artificial neural networks served as the platform for neuron computing
technology in the field of cognitive science.
The document provides an introduction to neural networks including:
1) It describes real neural networks in the brain and how they transmit electrical signals between neurons.
2) It then explains artificial neural networks which are computing systems inspired by biological neural networks and how they are composed of interconnected processing elements that can learn tasks.
3) Various types of neural networks are discussed like feedforward and feedback networks as well as learning processes like supervised and unsupervised learning. Transfer functions that simulate neuron signals are also covered.
Neural networks are computational models inspired by the human brain. They consist of interconnected nodes that process information using parallel computing. Neural networks can be trained to recognize patterns and classify data through a process of learning from examples. The human brain provides inspiration for neural networks through its use of neurons, synaptic plasticity which enables learning and adaptation, and its ability to reorganize through experience-dependent neuroplasticity. Key aspects of biological neurons like dendrites, synapses, and axons are replicated in the basic unit of artificial neural networks, called the artificial neuron.
Neural networks are computational models inspired by the human brain. They consist of interconnected nodes that process information using parallel computing. Neural networks can be trained to recognize patterns and classify data through a process of learning from examples. The human brain provides inspiration for neural networks through its use of neurons, synaptic plasticity which enables learning and adaptation, and its ability to reorganize through experience-dependent neuroplasticity. Key aspects of biological neurons like dendrites, synapses, and axons are replicated in the basic unit of artificial neural networks, called the artificial neuron.
This document provides information about the CS407 Neural Computation course. It outlines the lecturer, timetable, assessment, textbook recommendations, and covers topics from today's lecture including an introduction to neural networks, their inspiration from the brain, a brief history, applications, and an overview of topics to be covered in the course.
Artificial neural networks (ANNs) are computational models inspired by the human brain that are used for predictive analytics and nonlinear statistical modeling. ANNs can learn complex patterns and relationships from large datasets through a process of training, and then make predictions on new data. The three most common types of ANN architectures are multilayer perceptrons, radial basis function networks, and self-organizing maps. ANNs have been successfully applied across many domains, including finance, medicine, engineering, and biology, to solve problems involving classification, prediction, and nonlinear pattern recognition.
1. The document describes an introductory course on neural networks. It includes information on topics covered, textbooks, assignments, and report topics.
2. The main topics covered are comprehensive introduction, learning algorithms, and types of neural networks. Report topics include the McCulloch-Pitts model, applications of neural networks, and various learning algorithms.
3. The document also provides background information on biological neural networks and the basic components and functioning of artificial neural networks at a high level.
3. What is an ANN Describe various types of ANN. Which ANN do you p.pdfivylinvaydak64229
3. What is an ANN? Describe various types of ANN. Which ANN do you prefer amidst of the
variety of ANNs? Justify the reason beyond this.
Solution
What is ANN?:
An artificial neuron network (ANN) is a computational model based on the structure and
functions of biological neural networks. Information that flows through the network affects the
structure of the ANN because a neural network changes - or learns, in a sense - based on that
input and output.
ANNs have three layers that are interconnected. The first layer consists of input neurons. Those
neurons send data on to the second layer, which in turn sends the output neurons to the third
layer.
Types of artificial neural networks:
There are two Artificial Neural Network topologies FreeForward and Feedback.
FeedForward ANN
The information flow is unidirectional. A unit sends information to other unit from which it does
not receive any information. There are no feedback loops. They are used in pattern
generation/recognition/classification. They have fixed inputs and outputs.
Feedback.:
Here, feedback loops are allowed. They are used in content addressable memories.
Radial Basis Function (RBF) Neural Network –
Radial basis functions are powerful techniques for interpolation in multidimensional space. A
RBF is a function which has built into a distance criterion with respect to a center. RBF neural
networks have the advantage of not suffering from local minima in the same way as Multi-Layer
Perceptrons. RBF neural networks have the disadvantage of requiring good coverage of the input
space by radial basis functions.
Kohonen Self-organizing Neural Network – The self-organizing map (SOM) performs a form of
unsupervised learning. A set of artificial neurons learn to map points in an input space to
coordinates in an output space. The input space can have different dimensions and topology from
the output space, and the SOM will attempt to preserve these.
Recurrent Neural Networks – Recurrent neural networks (RNNs) are models with bi-directional
data flow. Recurrent neural networks can be used as general sequence processors. Various types
of Recurrent neural networks are Fully recurrent network (Hopfield network and Boltzmann
machine), Simple recurrent networks, Echo state network, Long short term memory network, Bi-
directional RNN, Hierarchical RNN, and Stochastic neural networks.
Modular Neural Network : Biological studies have shown that the human brain functions not as a
single massive network, but as a collection of small networks. This realization gave birth to the
concept of modular neural networks, in which several small networks cooperate or compete to
solve problems.
Physical Neural Network :
A physical neural network includes electrically adjustable resistance material to simulate
artificial synapses.
Feed Forwad is mostly used ANN Network due to its different applications:
1)Physiological feed-forward system:In physiology, feed-forward control is exemplified by the
normal anticipatory regu.
This document discusses graph algorithms and directed acyclic graphs (DAGs). It explains that the edges in a graph can be identified as tree, back, forward, or cross edges based on the color of vertices during depth-first search (DFS). It also defines DAGs as directed graphs without cycles and describes how to perform a topological sort of a DAG by inserting vertices into a linked list based on their finishing times from DFS. Finally, it discusses how to find strongly connected components (SCCs) in a graph using DFS on the original graph and its transpose.
This document discusses string matching algorithms. It begins with an introduction to the naive string matching algorithm and its quadratic runtime. Then it proposes three improved algorithms: FC-RJ, FLC-RJ, and FMLC-RJ, which attempt to match patterns by restricting comparisons based on the first, first and last, or first, middle, and last characters, respectively. Experimental results show that these three proposed algorithms outperform the naive algorithm by reducing execution time, with FMLC-RJ working best for three-character patterns.
The document discusses shortest path problems and algorithms. It defines the shortest path problem as finding the minimum weight path between two vertices in a weighted graph. It presents the Bellman-Ford algorithm, which can handle graphs with negative edge weights but detects negative cycles. It also presents Dijkstra's algorithm, which only works for graphs without negative edge weights. Key steps of the algorithms include initialization, relaxation of edges to update distance estimates, and ensuring the shortest path property is satisfied.
The document discusses strongly connected component decomposition (SCCD) which uses depth-first search (DFS) to separate a directed graph into subsets of mutually reachable vertices. It describes running DFS on the original graph and its transpose to find these subsets in Θ(V+E) time, then provides an example applying the three step process of running DFS on the graph and transpose, finding two strongly connected components.
Red-black trees are self-balancing binary search trees. They guarantee an O(log n) running time for operations by ensuring that no path from the root to a leaf is more than twice as long as any other. Nodes are colored red or black, and properties of the coloring are designed to keep the tree balanced. Inserting and deleting nodes may violate these properties, so rotations are used to restore the red-black properties and balance of the tree.
This document discusses recurrences and the master method for solving recurrence relations. It defines a recurrence as an equation that describes a function in terms of its value on smaller functions. The master method provides three cases for solving recurrences of the form T(n) = aT(n/b) + f(n). If f(n) is asymptotically smaller than nlogba, the solution is Θ(nlogba). If f(n) is Θ(nlogba), the solution is Θ(nlogba lgn). If f(n) is asymptotically larger and the regularity condition holds, the solution is Θ(f(n)). It provides examples of applying
The document discusses the Rabin-Karp algorithm for string matching. It defines Rabin-Karp as a string search algorithm that compares hash values of strings rather than the strings themselves. It explains that Rabin-Karp works by calculating a hash value for the pattern and text subsequences to compare, and only does a brute force comparison when hash values match. The worst-case complexity is O(n-m+1)m but the average case is O(n+m) plus processing spurious hits. Real-life applications include bioinformatics to find protein similarities.
The document discusses minimum spanning trees (MST) and two algorithms for finding them: Prim's algorithm and Kruskal's algorithm. It begins by defining an MST as a spanning tree (connected acyclic graph containing all vertices) with minimum total edge weight. Prim's algorithm grows a single tree by repeatedly adding the minimum weight edge connecting the growing tree to another vertex. Kruskal's algorithm grows a forest by repeatedly merging two components via the minimum weight edge connecting them. Both algorithms produce optimal MSTs by adding only "safe" edges that cannot be part of a cycle.
This document discusses the analysis of insertion sort and merge sort algorithms. It covers the worst-case and average-case analysis of insertion sort. For merge sort, it describes the divide-and-conquer technique, the merge sort algorithm including recursive calls, how it works to merge elements, and analyzes merge sort through constructing a recursion tree to prove its runtime is O(n log n).
The document discusses loop invariants and uses insertion sort as an example. The invariant for insertion sort is that at the start of each iteration of the outer for loop, the elements in A[1...j-1] are sorted. It shows that this invariant is true before the first iteration, remains true after each iteration by how insertion sort works, and when the loops terminate the entire array A[1...n] will be sorted, proving correctness.
Linear sorting algorithms like counting sort, bucket sort, and radix sort can sort arrays of numbers in linear O(n) time by exploiting properties of the data. Counting sort works for integers within a range [0,r] by counting the frequency of each number and using the frequencies to place numbers in the correct output positions. Bucket sort places numbers uniformly distributed between 0 and 1 into buckets and sorts the buckets. Radix sort treats multi-digit numbers as strings by sorting based on individual digit positions from least to most significant.
The document discusses heap data structures and algorithms. A heap is a binary tree that satisfies the heap property of a parent being greater than or equal to its children. Common operations on heaps like building
Greedy algorithms make locally optimal choices at each step in the hope of finding a globally optimal solution. The activity selection problem involves choosing a maximum set of activities that do not overlap in time. The greedy algorithm for this problem sorts activities by finish time and selects the earliest finishing activity at each step. This algorithm is optimal because the activity selection problem exhibits the optimal substructure property and the greedy algorithm satisfies the greedy-choice property at each step.
#StandardsGoals for 2025: Standards & certification roundup - Tech Forum 2025BookNet Canada
Book industry standards are evolving rapidly. In the first part of this session, we’ll share an overview of key developments from 2024 and the early months of 2025. Then, BookNet’s resident standards expert, Tom Richardson, and CEO, Lauren Stewart, have a forward-looking conversation about what’s next.
Link to recording, transcript, and accompanying resource: https://bnctechforum.ca/sessions/standardsgoals-for-2025-standards-certification-roundup/
Presented by BookNet Canada on May 6, 2025 with support from the Department of Canadian Heritage.
What is Model Context Protocol(MCP) - The new technology for communication bw...Vishnu Singh Chundawat
The MCP (Model Context Protocol) is a framework designed to manage context and interaction within complex systems. This SlideShare presentation will provide a detailed overview of the MCP Model, its applications, and how it plays a crucial role in improving communication and decision-making in distributed systems. We will explore the key concepts behind the protocol, including the importance of context, data management, and how this model enhances system adaptability and responsiveness. Ideal for software developers, system architects, and IT professionals, this presentation will offer valuable insights into how the MCP Model can streamline workflows, improve efficiency, and create more intuitive systems for a wide range of use cases.
Procurement Insights Cost To Value Guide.pptxJon Hansen
Procurement Insights integrated Historic Procurement Industry Archives, serves as a powerful complement — not a competitor — to other procurement industry firms. It fills critical gaps in depth, agility, and contextual insight that most traditional analyst and association models overlook.
Learn more about this value- driven proprietary service offering here.
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep DiveScyllaDB
Want to learn practical tips for designing systems that can scale efficiently without compromising speed?
Join us for a workshop where we’ll address these challenges head-on and explore how to architect low-latency systems using Rust. During this free interactive workshop oriented for developers, engineers, and architects, we’ll cover how Rust’s unique language features and the Tokio async runtime enable high-performance application development.
As you explore key principles of designing low-latency systems with Rust, you will learn how to:
- Create and compile a real-world app with Rust
- Connect the application to ScyllaDB (NoSQL data store)
- Negotiate tradeoffs related to data modeling and querying
- Manage and monitor the database for consistently low latencies
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdfAbi john
Analyze the growth of meme coins from mere online jokes to potential assets in the digital economy. Explore the community, culture, and utility as they elevate themselves to a new era in cryptocurrency.
Quantum Computing Quick Research Guide by Arthur MorganArthur Morgan
This is a Quick Research Guide (QRG).
QRGs include the following:
- A brief, high-level overview of the QRG topic.
- A milestone timeline for the QRG topic.
- Links to various free online resource materials to provide a deeper dive into the QRG topic.
- Conclusion and a recommendation for at least two books available in the SJPL system on the QRG topic.
QRGs planned for the series:
- Artificial Intelligence QRG
- Quantum Computing QRG
- Big Data Analytics QRG
- Spacecraft Guidance, Navigation & Control QRG (coming 2026)
- UK Home Computing & The Birth of ARM QRG (coming 2027)
Any questions or comments?
- Please contact Arthur Morgan at [email protected].
100% human made.
AI EngineHost Review: Revolutionary USA Datacenter-Based Hosting with NVIDIA ...SOFTTECHHUB
I started my online journey with several hosting services before stumbling upon Ai EngineHost. At first, the idea of paying one fee and getting lifetime access seemed too good to pass up. The platform is built on reliable US-based servers, ensuring your projects run at high speeds and remain safe. Let me take you step by step through its benefits and features as I explain why this hosting solution is a perfect fit for digital entrepreneurs.
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungenpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-nomad-web-best-practices-und-verwaltung-von-multiuser-umgebungen/
HCL Nomad Web wird als die nächste Generation des HCL Notes-Clients gefeiert und bietet zahlreiche Vorteile, wie die Beseitigung des Bedarfs an Paketierung, Verteilung und Installation. Nomad Web-Client-Updates werden “automatisch” im Hintergrund installiert, was den administrativen Aufwand im Vergleich zu traditionellen HCL Notes-Clients erheblich reduziert. Allerdings stellt die Fehlerbehebung in Nomad Web im Vergleich zum Notes-Client einzigartige Herausforderungen dar.
Begleiten Sie Christoph und Marc, während sie demonstrieren, wie der Fehlerbehebungsprozess in HCL Nomad Web vereinfacht werden kann, um eine reibungslose und effiziente Benutzererfahrung zu gewährleisten.
In diesem Webinar werden wir effektive Strategien zur Diagnose und Lösung häufiger Probleme in HCL Nomad Web untersuchen, einschließlich
- Zugriff auf die Konsole
- Auffinden und Interpretieren von Protokolldateien
- Zugriff auf den Datenordner im Cache des Browsers (unter Verwendung von OPFS)
- Verständnis der Unterschiede zwischen Einzel- und Mehrbenutzerszenarien
- Nutzung der Client Clocking-Funktion
Increasing Retail Store Efficiency How can Planograms Save Time and Money.pptxAnoop Ashok
In today's fast-paced retail environment, efficiency is key. Every minute counts, and every penny matters. One tool that can significantly boost your store's efficiency is a well-executed planogram. These visual merchandising blueprints not only enhance store layouts but also save time and money in the process.
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptxJustin Reock
Building 10x Organizations with Modern Productivity Metrics
10x developers may be a myth, but 10x organizations are very real, as proven by the influential study performed in the 1980s, ‘The Coding War Games.’
Right now, here in early 2025, we seem to be experiencing YAPP (Yet Another Productivity Philosophy), and that philosophy is converging on developer experience. It seems that with every new method we invent for the delivery of products, whether physical or virtual, we reinvent productivity philosophies to go alongside them.
But which of these approaches actually work? DORA? SPACE? DevEx? What should we invest in and create urgency behind today, so that we don’t find ourselves having the same discussion again in a decade?
AI and Data Privacy in 2025: Global TrendsInData Labs
In this infographic, we explore how businesses can implement effective governance frameworks to address AI data privacy. Understanding it is crucial for developing effective strategies that ensure compliance, safeguard customer trust, and leverage AI responsibly. Equip yourself with insights that can drive informed decision-making and position your organization for success in the future of data privacy.
This infographic contains:
-AI and data privacy: Key findings
-Statistics on AI data privacy in the today’s world
-Tips on how to overcome data privacy challenges
-Benefits of AI data security investments.
Keep up-to-date on how AI is reshaping privacy standards and what this entails for both individuals and organizations.
Big Data Analytics Quick Research Guide by Arthur MorganArthur Morgan
This is a Quick Research Guide (QRG).
QRGs include the following:
- A brief, high-level overview of the QRG topic.
- A milestone timeline for the QRG topic.
- Links to various free online resource materials to provide a deeper dive into the QRG topic.
- Conclusion and a recommendation for at least two books available in the SJPL system on the QRG topic.
QRGs planned for the series:
- Artificial Intelligence QRG
- Quantum Computing QRG
- Big Data Analytics QRG
- Spacecraft Guidance, Navigation & Control QRG (coming 2026)
- UK Home Computing & The Birth of ARM QRG (coming 2027)
Any questions or comments?
- Please contact Arthur Morgan at [email protected].
100% human made.
Big Data Analytics Quick Research Guide by Arthur MorganArthur Morgan
Fundamentals of Neural Network (Soft Computing)
1. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
Fundamentals of Neural Networks : Soft Computing Course Lecture 7 – 14, notes, slides
www.myreaders.info/ , RC Chakraborty, e-mail [email protected] , Dec. 01, 2010
http://www.myreaders.info/html/soft_computing.html
Fundamentals of Neural Networks
Soft Computing
www.myreaders.info
Return to Website
Neural network, topics : Introduction, biological neuron model,
artificial neuron model, neuron equation. Artificial neuron : basic
elements, activation and threshold function, piecewise linear and
sigmoidal function. Neural network architectures : single layer feed-
forward network, multi layer feed-forward network, recurrent
networks. Learning methods in neural networks : unsupervised
Learning - Hebbian learning, competitive learning; Supervised
learning - stochastic learning, gradient descent learning; Reinforced
learning. Taxonomy of neural network systems : popular neural
network systems, classification of neural network systems as per
learning methods and architecture. Single-layer NN system : single
layer perceptron, learning algorithm for training perceptron,
linearly separable task, XOR problem, ADAptive LINear Element
(ADALINE) - architecture, and training. Applications of neural
networks: clustering, classification, pattern recognition, function
approximation, prediction systems.
2. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
Fundamentals of Neural Networks
Soft Computing
Topics
(Lectures 07, 08, 09, 10, 11, 12, 13, 14 8 hours)
Slides
1. Introduction
Why neural network ?, Research History, Biological Neuron model,
Artificial Neuron model, Notations, Neuron equation.
03-12
2. Model of Artificial Neuron
Artificial neuron - basic elements, Activation functions – Threshold
function, Piecewise linear function, Sigmoidal function, Example.
13-19
3. Neural Network Architectures
Single layer Feed-forward network, Multi layer Feed-forward network,
Recurrent networks.
20-23
4. Learning Methods in Neural Networks
Learning algorithms: Unsupervised Learning - Hebbian Learning,
Competitive learning; Supervised Learning : Stochastic learning,
Gradient descent learning; Reinforced Learning;
24-29
5. Taxonomy Of Neural Network Systems
Popular neural network systems; Classification of neural network
systems with respect to learning methods and architecture types.
30-32
6. Single-Layer NN System
Single layer perceptron : Learning algorithm for training Perceptron,
Linearly separable task, XOR Problem; ADAptive LINear Element
(ADALINE) : Architecture, Training.
32-39
7. Applications of Neural Networks
Clustering, Classification / pattern recognition, Function approximation,
Prediction systems.
39
8. References : 40
02
3. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
Fundamentals of Neural Networks
What is Neural Net ?
• A neural net is an artificial representation of the human brain that
tries to simulate its learning process. An artificial neural network
(ANN) is often called a "Neural Network" or simply Neural Net (NN).
• Traditionally, the word neural network is referred to a network of
biological neurons in the nervous system that process and transmit
information.
• Artificial neural network is an interconnected group of artificial neurons
that uses a mathematical model or computational model for information
processing based on a connectionist approach to computation.
• The artificial neural networks are made of interconnecting artificial
neurons which may share some properties of biological neural networks.
• Artificial Neural network is a network of simple processing elements
(neurons) which can exhibit complex global behavior, determined by the
connections between the processing elements and element parameters.
03
4. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
1. Introduction
Neural Computers mimic certain processing capabilities of the human brain.
- Neural Computing is an information processing paradigm, inspired by
biological system, composed of a large number of highly interconnected
processing elements (neurons) working in unison to solve specific problems.
- Artificial Neural Networks (ANNs), like people, learn by example.
- An ANN is configured for a specific application, such as pattern recognition or
data classification, through a learning process.
- Learning in biological systems involves adjustments to the synaptic
connections that exist between the neurons. This is true of ANNs as well.
04
5. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
1.1 Why Neural Network
Neural Networks follow a different paradigm for computing.
■ The conventional computers are good for - fast arithmetic and does
what programmer programs, ask them to do.
■ The conventional computers are not so good for - interacting with
noisy data or data from the environment, massive parallelism, fault
tolerance, and adapting to circumstances.
■ The neural network systems help where we can not formulate an
algorithmic solution or where we can get lots of examples of the
behavior we require.
■ Neural Networks follow different paradigm for computing.
The von Neumann machines are based on the processing/memory
abstraction of human information processing.
The neural networks are based on the parallel architecture of
biological brains.
■ Neural networks are a form of multiprocessor computer system, with
- simple processing elements ,
- a high degree of interconnection,
- simple scalar messages, and
- adaptive interaction between elements.
05
6. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
1.2 Research History
The history is relevant because for nearly two decades the future of
Neural network remained uncertain.
McCulloch and Pitts (1943) are generally recognized as the designers of the
first neural network. They combined many simple processing units together
that could lead to an overall increase in computational power. They
suggested many ideas like : a neuron has a threshold level and once that
level is reached the neuron fires. It is still the fundamental way in which
ANNs operate. The McCulloch and Pitts's network had a fixed set of weights.
Hebb (1949) developed the first learning rule, that is if two neurons are
active at the same time then the strength between them should be
increased.
In the 1950 and 60's, many researchers (Block, Minsky, Papert, and
Rosenblatt worked on perceptron. The neural network model could be
proved to converge to the correct weights, that will solve the problem. The
weight adjustment (learning algorithm) used in the perceptron was found
more powerful than the learning rules used by Hebb. The perceptron caused
great excitement. It was thought to produce programs that could think.
Minsky & Papert (1969) showed that perceptron could not learn those
functions which are not linearly separable.
The neural networks research declined throughout the 1970 and until mid
80's because the perceptron could not learn certain important functions.
Neural network regained importance in 1985-86. The researchers, Parker
and LeCun discovered a learning algorithm for multi-layer networks called
back propagation that could solve problems that were not linearly
separable.
06
7. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
1.3 Biological Neuron Model
The human brain consists of a large number, more than a billion of
neural cells that process information. Each cell works like a simple
processor. The massive interaction between all cells and their parallel
processing only makes the brain's abilities possible.
Fig. Structure of Neuron
Dendrites are branching fibers that
extend from the cell body or soma.
Soma or cell body of a neuron contains
the nucleus and other structures, support
chemical processing and production of
neurotransmitters.
Axon is a singular fiber carries
information away from the soma to the
synaptic sites of other neurons (dendrites
and somas), muscles, or glands.
Axon hillock is the site of summation
for incoming information. At any
moment, the collective influence of all
neurons that conduct impulses to a given
neuron will determine whether or not an
action potential will be initiated at the
axon hillock and propagated along the axon.
Myelin Sheath consists of fat-containing cells that insulate the axon from electrical
activity. This insulation acts to increase the rate of transmission of signals. A gap
exists between each myelin sheath cell along the axon. Since fat inhibits the
propagation of electricity, the signals jump from one gap to the next.
Nodes of Ranvier are the gaps (about 1 µm) between myelin sheath cells long axons
are Since fat serves as a good insulator, the myelin sheaths speed the rate of
transmission of an electrical impulse along the axon.
Synapse is the point of connection between two neurons or a neuron and a muscle or
a gland. Electrochemical communication between neurons takes place at these
junctions.
Terminal Buttons of a neuron are the small knobs at the end of an axon that release
chemicals called neurotransmitters.
07
8. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
• Information flow in a Neural Cell
The input /output and the propagation of information are shown below.
Fig. Structure of a neural cell in the human brain
■ Dendrites receive activation from other neurons.
■ Soma processes the incoming activations and converts them into
output activations.
■ Axons act as transmission lines to send activation to other neurons.
■ Synapses the junctions allow signal transmission between the
axons and dendrites.
■ The process of transmission is by diffusion of chemicals called
neuro-transmitters.
McCulloch-Pitts introduced a simplified model of this real neurons.
08
9. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
1.4 Artificial Neuron Model
An artificial neuron is a mathematical function conceived as a simple
model of a real (biological) neuron.
• The McCulloch-Pitts Neuron
This is a simplified model of real neurons, known as a Threshold Logic Unit.
Input1
Input 2
Input n
■ A set of input connections brings in activations from other neurons.
■ A processing unit sums the inputs, and then applies a non-linear
activation function (i.e. squashing / transfer / threshold function).
■ An output line transmits the result to other neurons.
In other words ,
- The input to a neuron arrives in the form of signals.
- The signals build up in the cell.
- Finally the cell discharges (cell fires) through the output .
- The cell can start building up signals again.
09
Σ Output
10. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
1.5 Notations
Recaps : Scalar, Vectors, Matrices and Functions
• Scalar : The number xi can be added up to give a scalar number.
s = x1 + x2 + x3 + . . . . + xn = xi
• Vectors : An ordered sets of related numbers. Row Vectors (1 x n)
X = ( x1 , x2 , x3 , . . ., xn ) , Y = ( y1 , y2 , y3 , . . ., yn )
Add : Two vectors of same length added to give another vector.
Z = X + Y = (x1 + y1 , x2 + y2 , . . . . , xn + yn)
Multiply: Two vectors of same length multiplied to give a scalar.
p = X . Y = x1 y1 + x2 y2 + . . . . + xnyn = xi yi
10
Σ
i=1
n
Σ
i=1
n
11. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
• Matrices : m x n matrix , row no = m , column no = n
w11 w11 . . . . w1n
w21 w21 . . . . w21
W = . . . . . . .
. . . . . . .
wm1 w11 . . . . wmn
Add or Subtract : Matrices of the same size are added or subtracted
component by component. A + B = C , cij = aij + bij
a11 a12 b11 b12 c11 = a11+b11 c12 = a12+b12
a21 a22 b21 b22 C21 = a21+b21 C22 = a22 +b22
Multiply : matrix A multiplied by matrix B gives matrix C.
(m x n) (n x p) (m x p)
elements cij = aik bkj
a11 a12 b11 b12 c11 c12
a21 a22 b21 b22 c21 c22
c11 = (a11 x b11) + (a12 x B21)
c12 = (a11 x b12) + (a12 x B22)
C21 = (a21 x b11) + (a22 x B21)
C22 = (a21 x b12) + (a22 x B22)
11
+ =
Σ
k=1
n
x =
12. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Introduction
1.6 Functions
The Function y= f(x) describes a relationship, an input-output mapping,
from x to y.
■ Threshold or Sign function : sgn(x) defined as
1 if x ≥ 0
sgn (x) =
0 if x < 0
Sign(x)
O/P
-4 -3 -2 -1 0 1 2 3 4 I/P
■ Threshold or Sign function : sigmoid(x) defined as a smoothed
(differentiable) form of the threshold function
1
sigmoid (x) =
1 + e
-x
Sign(x)
O/P
-4 -3 -2 -1 0 1 2 3 4 I/P
12
0
1
.2
.6
.4
.8
0
1
.2
.6
.8
13. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Artificial Neuron Model
2. Model of Artificial Neuron
A very simplified model of real neurons is known as a Threshold Logic
Unit (TLU). The model is said to have :
- A set of synapses (connections) brings in activations from other neurons.
- A processing unit sums the inputs, and then applies a non-linear activation
function (i.e. squashing / transfer / threshold function).
- An output line transmits the result to other neurons.
2.1 McCulloch-Pitts (M-P) Neuron Equation
McCulloch-Pitts neuron is a simplified model of real biological neuron.
Input 1
Input 2
Input n
Simplified Model of Real Neuron
(Threshold Logic Unit)
The equation for the output of a McCulloch-Pitts neuron as a function
of 1 to n inputs is written as
Output = sgn ( Input i - Φ )
where Φ is the neuron’s activation threshold.
If Input i ≥ Φ then Output = 1
If Input i < Φ then Output = 0
In this McCulloch-Pitts neuron model, the missing features are :
- Non-binary input and output,
- Non-linear summation,
- Smooth thresholding,
- Stochastic, and
- Temporal information processing.
13
Σ Output
Σ
i=1
n
Σ
i=1
n
Σ
i=1
n
14. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Artificial Neuron Model
2.2 Artificial Neuron - Basic Elements
Neuron consists of three basic components - weights, thresholds, and a
single activation function.
Fig Basic Elements of an Artificial Linear Neuron
■ Weighting Factors w
The values w1 , w2 , . . . wn are weights to determine the strength of
input vector X = [x1 , x2 , . . . , xn]
T
. Each input is multiplied by the
associated weight of the neuron connection XT
W. The +ve weight
excites and the -ve weight inhibits the node output.
I = X
T
.W = x1 w1 + x2 w2 + . . . . + xnwn = xi wi
■ Threshold Φ
The node’s internal threshold Φ is the magnitude offset. It affects the
activation of the node output y as:
Y = f (I) = f { xi wi - Φk }
To generate the final output Y , the sum is passed on to a non-linear
filter f called Activation Function or Transfer function or Squash function
which releases the output Y.
14
W1
Σ
W2
Wn
x1
x2
xn
Activation
Function
i=1
Synaptic Weights
Φ
Threshold
y
Σ
i=1
n
Σ
i=1
n
15. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Artificial Neuron Model
■ Threshold for a Neuron
In practice, neurons generally do not fire (produce an output) unless
their total input goes above a threshold value.
The total input for each neuron is the sum of the weighted inputs
to the neuron minus its threshold value. This is then passed through
the sigmoid function. The equation for the transition in a neuron is :
a = 1/(1 + exp(- x)) where
x = ai wi - Q
a is the activation for the neuron
ai is the activation for neuron i
wi is the weight
Q is the threshold subtracted
■ Activation Function
An activation function f performs a mathematical operation on the
signal output. The most common activation functions are:
- Linear Function,
- Piecewise Linear Function,
- Tangent hyperbolic function
- Threshold Function,
- Sigmoidal (S shaped) function,
The activation functions are chosen depending upon the type of
problem to be solved by the network.
15
Σ
i
16. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Artificial Neuron Model
2.2 Activation Functions f - Types
Over the years, researches tried several functions to convert the input into
an outputs. The most commonly used functions are described below.
- I/P Horizontal axis shows sum of inputs .
- O/P Vertical axis shows the value the function produces ie output.
- All functions f are designed to produce values between 0 and 1.
• Threshold Function
A threshold (hard-limiter) activation function is either a binary type or
a bipolar type as shown below.
binary threshold
O/p
I/P
Output of a binary threshold function produces :
1 if the weighted sum of the inputs is +ve,
0 if the weighted sum of the inputs is -ve.
1 if I ≥ 0
Y = f (I) =
0 if I < 0
bipolar threshold
O/p
I/P
Output of a bipolar threshold function produces :
1 if the weighted sum of the inputs is +ve,
-1 if the weighted sum of the inputs is -ve.
1 if I ≥ 0
Y = f (I) =
-1 if I < 0
Neuron with hard limiter activation function is called McCulloch-Pitts model.
16
1
1
-1
17. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Artificial Neuron Model
• Piecewise Linear Function
This activation function is also called saturating linear function and can
have either a binary or bipolar range for the saturation limits of the output.
The mathematical model for a symmetric saturation function is described
below.
Piecewise Linear
O/p
I/P
This is a sloping function that produces :
-1 for a -ve weighted sum of inputs,
1 for a +ve weighted sum of inputs.
∝ I proportional to input for values between +1
and -1 weighted sum,
1 if I ≥ 0
Y = f (I) = I if -1 ≥ I ≥ 1
-1 if I < 0
17
+1
-1
18. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Artificial Neuron Model
• Sigmoidal Function (S-shape function)
The nonlinear curved S-shape function is called the sigmoid function.
This is most common type of activation used to construct the neural
networks. It is mathematically well behaved, differentiable and strictly
increasing function.
Sigmoidal function A sigmoidal transfer function can be
written in the form:
1
Y = f (I) = , 0 ≤ f(I) ≤ 1
1 + e -α I
= 1/(1 + exp(-α I)) , 0 ≤ f(I) ≤ 1
This is explained as
≈ 0 for large -ve input values,
1 for large +ve values, with
a smooth transition between the two.
α is slope parameter also called shape
parameter; symbol the λ is also used to
represented this parameter.
The sigmoidal function is achieved using exponential equation.
By varying α different shapes of the function can be obtained which
adjusts the abruptness of the function as it changes between the two
asymptotic values.
18
1 O/P
0.5
I/P
-4 -2 0 1 2
α = 1.0
α = 0.5
α = 2.0
19. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Artificial Neuron Model
• Example :
The neuron shown consists of four inputs with the weights.
Fig Neuron Structure of Example
The output I of the network, prior to the activation function stage, is
+1
+1
I = XT
. W = 1 2 5 8 = 14
-1
+2
= (1 x 1) + (2 x 1) + (5 x -1) + (8 x 2) = 14
With a binary activation function the outputs of the neuron is:
y (threshold) = 1;
19
+1
Σ
+1
+2
-1
x1=1
x2=2
xn=8
Activation
Function
Summing
Junction
Synaptic
Weights
Φ = 0
Threshold
y
X3=5
I
20. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Architecture
3. Neural Network Architectures
An Artificial Neural Network (ANN) is a data processing system, consisting
large number of simple highly interconnected processing elements as
artificial neuron in a network structure that can be represented using a
directed graph G, an ordered 2-tuple (V, E) , consisting a set V of vertices
and a set E of edges.
- The vertices may represent neurons (input/output) and
- The edges may represent synaptic links labeled by the weights attached.
Example :
Fig. Directed Graph
Vertices V = { v1 , v2 , v3 , v4, v5 }
Edges E = { e1 , e2 , e3 , e4, e5 }
20
V1 V3
V2 V4
V5
e3
e2
e5
e4
e5
21. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Architecture
3.1 Single Layer Feed-forward Network
The Single Layer Feed-forward Network consists of a single layer of
weights , where the inputs are directly connected to the outputs, via a
series of weights. The synaptic links carrying weights connect every input
to every output , but not other way. This way it is considered a network of
feed-forward type. The sum of the products of the weights and the inputs
is calculated in each neuron node, and if the value is above some threshold
(typically 0) the neuron fires and takes the activated value (typically 1);
otherwise it takes the deactivated value (typically -1).
Fig. Single Layer Feed-forward Network
21
w21
w11
w12
wn2
wn1
w1m
w2m
wnm
w22
y1
y2
ym
x1
x2
xn
output yj
input xi weights wij
Single layer
Neurons
22. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Architecture
3.2 Multi Layer Feed-forward Network
The name suggests, it consists of multiple layers. The architecture of
this class of network, besides having the input and the output layers,
also have one or more intermediary layers called hidden layers. The
computational units of the hidden layer are known as hidden neurons.
Fig. Multilayer feed-forward network in (ℓ – m – n) configuration.
- The hidden layer does intermediate computation before directing the
input to output layer.
- The input layer neurons are linked to the hidden layer neurons; the
weights on these links are referred to as input-hidden layer weights.
- The hidden layer neurons and the corresponding weights are referred to
as output-hidden layer weights.
- A multi-layer feed-forward network with ℓ input neurons, m1 neurons in
the first hidden layers, m2 neurons in the second hidden layers, and n
output neurons in the output layers is written as (ℓ - m1 - m2 – n ).
The Fig. above illustrates a multilayer feed-forward network with a
configuration (ℓ - m – n).
22
w11
w12
v21
v11
w1m
vn1
v1m
v2m
Vℓm
w11
x1
x2
xℓ
y3
y1
y2
yn
y1
ym
Hidden Layer
neurons yj
Output Layer
neurons zk
Input Layer
neurons xi
Input
hidden layer
weights vij
Output
hidden layer
weights wjk
23. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network – Architecture
3.3 Recurrent Networks
The Recurrent Networks differ from feed-forward architecture. A Recurrent
network has at least one feed back loop.
Example :
Fig Recurrent Neural Network
There could be neurons with self-feedback links; that is the output of a
neuron is fed back into it self as input.
23
x1
x2
Xℓ
y2
y1
Yn
y1
ym
Hidden Layer
neurons yj
Output Layer
neurons zk
Input Layer
neurons xi
Feedback
links
24. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Learning methods
4. Learning Methods in Neural Networks
The learning methods in neural networks are classified into three basic types :
- Supervised Learning,
- Unsupervised Learning and
- Reinforced Learning
These three types are classified based on :
- presence or absence of teacher and
- the information provided for the system to learn.
These are further categorized, based on the rules used, as
- Hebbian,
- Gradient descent,
- Competitive and
- Stochastic learning.
24
25. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Learning methods
• Classification of Learning Algorithms
Fig. below indicate the hierarchical representation of the algorithms
mentioned in the previous slide. These algorithms are explained in
subsequent slides.
Fig. Classification of learning algorithms
25
Neural Network
Learning algorithms
Unsupervised Learning
Supervised Learning
(Error based)
Reinforced Learning
(Output based)
Error Correction
Gradient descent
Stochastic
Back
Propagation
Least Mean
Square
Hebbian Competitive
26. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Learning methods
• Supervised Learning
- A teacher is present during learning process and presents expected
output.
- Every input pattern is used to train the network.
- Learning process is based on comparison, between network's computed
output and the correct expected output, generating "error".
- The "error" generated is used to change network parameters that result
improved performance.
• Unsupervised Learning
- No teacher is present.
- The expected or desired output is not presented to the network.
- The system learns of it own by discovering and adapting to the structural
features in the input patterns.
• Reinforced learning
- A teacher is present but does not present the expected or desired output
but only indicated if the computed output is correct or incorrect.
- The information provided helps the network in its learning process.
- A reward is given for correct answer computed and a penalty for a wrong
answer.
Note : The Supervised and Unsupervised learning methods are most popular
forms of learning compared to Reinforced learning.
26
27. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Learning methods
• Hebbian Learning
Hebb proposed a rule based on correlative weight adjustment.
In this rule, the input-output pattern pairs (Xi , Yi) are associated by
the weight matrix W, known as correlation matrix computed as
W = Xi YiT
where Yi
T
is the transpose of the associated output vector Yi
There are many variations of this rule proposed by the other
researchers (Kosko, Anderson, Lippman) .
27
Σ
i=1
n
28. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Learning methods
• Gradient descent Learning
This is based on the minimization of errors E defined in terms of weights
and the activation function of the network.
- Here, the activation function of the network is required to be
differentiable, because the updates of weight is dependent on
the gradient of the error E.
- If ∆ Wij is the weight update of the link connecting the i th and the j th
neuron of the two neighboring layers, then ∆ Wij is defined as
∆ Wij = η (∂ E / ∂ Wij )
where η is the learning rate parameters and (∂ E / ∂ Wij ) is error
gradient with reference to the weight Wij .
Note : The Hoffs Delta rule and Back-propagation learning rule are
the examples of Gradient descent learning.
28
29. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Learning methods
• Competitive Learning
- In this method, those neurons which respond strongly to the input
stimuli have their weights updated.
- When an input pattern is presented, all neurons in the layer compete,
and the winning neuron undergoes weight adjustment .
- This strategy is called "winner-takes-all".
• Stochastic Learning
- In this method the weights are adjusted in a probabilistic fashion.
- Example : Simulated annealing which is a learning mechanism
employed by Boltzmann and Cauchy machines.
29
30. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Systems
5. Taxonomy Of Neural Network Systems
In the previous sections, the Neural Network Architectures and the
Learning methods have been discussed. Here the popular neural network
systems are listed. The grouping of these systems in terms of architectures
and the learning methods are presented in the next slide.
• Neural Network Systems
– ADALINE (Adaptive Linear Neural Element)
– ART (Adaptive Resonance Theory)
– AM (Associative Memory)
– BAM (Bidirectional Associative Memory)
– Boltzmann machines
– BSB ( Brain-State-in-a-Box)
– Cauchy machines
– Hopfield Network
– LVQ (Learning Vector Quantization)
– Neoconition
– Perceptron
– RBF ( Radial Basis Function)
– RNN (Recurrent Neural Network)
– SOFM (Self-organizing Feature Map)
30
31. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Systems
• Classification of Neural Network
A taxonomy of neural network systems based on Architectural types
and the Learning methods is illustrated below.
Learning Methods
Gradient
descent
Hebbian Competitive Stochastic
Single-layer
feed-forward
ADALINE,
Hopfield,
Percepton,
AM,
Hopfield,
LVQ,
SOFM
-
Multi-layer
feed- forward
CCM,
MLFF,
RBF
Neocognition
Recurrent
Networks
RNN BAM,
BSB,
Hopfield,
ART Boltzmann and
Cauchy
machines
Table : Classification of Neural Network Systems with respect to
learning methods and Architecture types
31
32. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Single Layer learning
6. Single-Layer NN Systems
Here, a simple Perceptron Model and an ADALINE Network Model is presented.
6.1 Single layer Perceptron
Definition : An arrangement of one input layer of neurons feed forward
to one output layer of neurons is known as Single Layer Perceptron.
Fig. Simple Perceptron Model
1 if net j ≥ 0
y j = f (net j) = where net j = xi wij
0 if net j < 0
32
Σ
i=1
n
w21
w11
w12
wn2
wn1
w1m
w2m
wnm
w22
y1
y2
ym
x1
x2
xn
output yj
input xi weights wij
Single layer
Perceptron
33. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Single Layer learning
• Learning Algorithm : Training Perceptron
The training of Perceptron is a supervised learning algorithm where
weights are adjusted to minimize error when ever the output does
not match the desired output.
− If the output is correct then no adjustment of weights is done.
i.e. =
− If the output is 1 but should have been 0 then the weights are
decreased on the active input link
i.e. = − α . xi
− If the output is 0 but should have been 1 then the weights are
increased on the active input link
i.e. = + α . xi
Where
is the new adjusted weight, is the old weight
xi is the input and α is the learning rate parameter.
α small leads to slow and α large leads to fast learning.
33
W
i j
K+1
W
i j
K+1
W
i j
K
W
i j
K+1
W
i j
K
W
i j
K+1
W
i j
K
W
i j
K
34. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Single Layer learning
• Perceptron and Linearly Separable Task
Perceptron can not handle tasks which are not separable.
- Definition : Sets of points in 2-D space are linearly separable if the
sets can be separated by a straight line.
- Generalizing, a set of points in n-dimensional space are linearly
separable if there is a hyper plane of (n-1) dimensions separates
the sets.
Example
S1 S2 S1
S2
(a) Linearly separable patterns (b) Not Linearly separable patterns
Note : Perceptron cannot find weights for classification problems that
are not linearly separable.
34
35. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Single Layer learning
• XOR Problem :
Exclusive OR operation
Input x1 Input x2 Output
0 0 0
1 1 0
0 1 1
1 0 1
XOR truth table
X2
(0, 1) (1, 1)
(0, 0) X1
(0, 1)
Fig. Output of XOR in
X1 , x2 plane
Even parity is, even number of 1 bits in the input
Odd parity is, odd number of 1 bits in the input
- There is no way to draw a single straight line so that the circles are on
one side of the line and the dots on the other side.
- Perceptron is unable to find a line separating even parity input
patterns from odd parity input patterns.
35
•
°
•
°
Even parity •
Odd parity °
36. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Single Layer learning
• Perceptron Learning Algorithm
The algorithm is illustrated step-by-step.
■ Step 1 :
Create a peceptron with (n+1) input neurons x0 , x1 , . . . . . , . xn ,
where x0 = 1 is the bias input.
Let O be the output neuron.
■ Step 2 :
Initialize weight W = (w0 , w1 , . . . . . , . wn ) to random weights.
■ Step 3 :
Iterate through the input patterns Xj of the training set using the
weight set; ie compute the weighted sum of inputs net j = xi wi
for each input pattern j .
■ Step 4 :
Compute the output y j using the step function
1 if net j ≥ 0
y j = f (net j) = where net j = xi wij
0 if net j < 0
■ Step 5 :
Compare the computed output yj with the target output yj for
each input pattern j .
If all the input patterns have been classified correctly, then output
(read) the weights and exit.
■ Step 6 :
Otherwise, update the weights as given below :
If the computed outputs yj is 1 but should have been 0,
Then wi = wi - α xi , i= 0, 1, 2, . . . . , n
If the computed outputs yj is 0 but should have been 1,
Then wi = wi + α xi , i= 0, 1, 2, . . . . , n
where α is the learning parameter and is constant.
■ Step 7 :
goto step 3
■ END
36
Σ
i=1
n
Σ
i=1
n
37. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –ADALINE
6.2 ADAptive LINear Element (ADALINE)
An ADALINE consists of a single neuron of the McCulloch-Pitts type,
where its weights are determined by the normalized least mean
square (LMS) training law. The LMS learning rule is also referred to as
delta rule. It is a well-established supervised training method that
has been used over a wide range of diverse applications.
• Architecture of a simple ADALINE
The basic structure of an ADALINE is similar to a neuron with a
linear activation function and a feedback loop. During the training
phase of ADALINE, the input vector as well as the desired output
are presented to the network.
[The complete training mechanism has been explained in the next slide. ]
37
W1
Σ
W2
Wn
Σ
x1
x2
xn
Neuron
Error
Desired Output
Output
–
+
38. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –ADALINE
• ADALINE Training Mechanism
(Ref. Fig. in the previous slide - Architecture of a simple ADALINE)
■ The basic structure of an ADALINE is similar to a linear neuron
with an extra feedback loop.
■ During the training phase of ADALINE, the input vector
X = [x1 , x2 , . . . , xn]
T
as well as desired output are presented
to the network.
■ The weights are adaptively adjusted based on delta rule.
■ After the ADALINE is trained, an input vector presented to the
network with fixed weights will result in a scalar output.
■ Thus, the network performs an n dimensional mapping to a
scalar value.
■ The activation function is not used during the training phase.
Once the weights are properly adjusted, the response of the
trained unit can be tested by applying various inputs, which are
not in the training set. If the network produces consistent
responses to a high degree with the test inputs, it is said
that the network could generalize. The process of training and
generalization are two important attributes of this network.
Usage of ADLINE :
In practice, an ADALINE is used to
- Make binary decisions; the output is sent through a binary threshold.
- Realizations of logic gates such as AND, NOT and OR .
- Realize only those logic functions that are linearly separable.
38
39. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –Applications
7. Applications of Neural Network
Neural Network Applications can be grouped in following categories:
■ Clustering:
A clustering algorithm explores the similarity between patterns and
places similar patterns in a cluster. Best known applications include
data compression and data mining.
■ Classification/Pattern recognition:
The task of pattern recognition is to assign an input pattern
(like handwritten symbol) to one of many classes. This category
includes algorithmic implementations such as associative memory.
■ Function approximation :
The tasks of function approximation is to find an estimate of the
unknown function subject to noise. Various engineering and scientific
disciplines require function approximation.
■ Prediction Systems:
The task is to forecast some future values of a time-sequenced
data. Prediction has a significant impact on decision support systems.
Prediction differs from function approximation by considering time factor.
System may be dynamic and may produce different results for the
same input data based on system state (time).
39
40. R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
R
C
C
h
a
k
r
a
b
o
r
t
y
,
w
w
w
.
m
y
r
e
a
d
e
r
s
.
i
n
f
o
SC - Neural Network –References
8. References : Textbooks
1. "Neural Network, Fuzzy Logic, and Genetic Algorithms - Synthesis and
Applications", by S. Rajasekaran and G.A. Vijayalaksmi Pai, (2005), Prentice Hall,
Chapter 2, page 11-33.
2. "Soft Computing and Intelligent Systems Design - Theory, Tools and Applications",
by Fakhreddine karray and Clarence de Silva (2004), Addison Wesley, chapter 4,
page 223-248.
3. "Neural Networks: A Comprehensive Foundation", by Simon S. Haykin, (1999),
Prentice Hall, Chapter 1-7, page 1-363.
4. "Elements of Artificial Neural Networks", by Kishan Mehrotra, Chilukuri K. Mohan
and Sanjay Ranka, (1996), MIT Press, Chapter 1-5, page 1-214.
5. "Fundamentals of Neural Networks: Architecture, Algorithms and Applications", by
Laurene V. Fausett, (1993), Prentice Hall, Chapter1-4, page 1-214.
6. "Neural Network Design", by Martin T. Hagan, Howard B. Demuth and Mark
Hudson Beale, ( 1996) , PWS Publ. Company, Chapter 1-7, page 1-1 to 7-31.
7. "An Introduction to Neural Networks", by James A. Anderson, (1997), MIT Press,
Chapter 1- 12, page 1-401.
8. Related documents from open source, mainly internet. An exhaustive list is
being prepared for inclusion at a later date.
40