Unit-2 Notes
Unit-2 Notes
Maxnet
Maxnet network was developed by Lippmann in 1987. The Maxner serves as a sub net for
picking the node whose input is larger. All the nodes present in this subnet are fully
interconnected and there exist symmetrical weights in all these weighted interconnections.
Architecture of Maxnet
The architecrure of Maxnet is a fixed symmetrical weights are present over the weighted
interconnections. The weights between the neurons are inhibitory and fixed. The Maxnet with
this structure can be used as a subnet to select a particular node whose net input is the largest.
and
Step 3: Save the activations obtained for use in the next iteration. For j = 1 to m,
Step 4: Finally, test the stopping condition for convergence of the network. The following is
the stopping condition: If more than one node has a nonzero activation, continue; else stop.
Hamming Network
The Hamming network is a two-layer feedforward neural network for classification of binary
bipolar n-tuple input vectors using minimum Hamming distance denoted as DH(Lippmann,
1987). The first layer is the input layer for the n-tuple input vectors. The second layer (also
called the memory layer) stores p memory patterns. A p-class Hamming network has p output
neurons in this layer. The strongest response of a neuron is indicative of the minimum
Hamming distance between the stored pattern and the input vector.
Hamming Distance
Hamming distance of two vectors, x and y of dimension n
x.y = a - d
where: a is number of bits in aggreement in x & y(No.of Similaritie bits in x & y), and d is
number of bits different in x and y(No.of Dissimilaritie bits in x & y).
The value "a - d" is the Hamming distance existing between two vectors. Since, the total
number of components is n, we have,
n=a+d
i.e., d = n - a
On simplification, we get
x.y = a - (n - a)
x.y = 2a - n
2a = x.y + n
a = 12x.y + 12n
From the above equation, it is clearly understood that the weights can be set to one-half the
exemplar vector and bias can be set initially to n/2
Initialize the bias for storing the "m" exemplar vectors. For j = 1 to m,
Step 4: Maxnet is found to iterate for finding the exemplar that best matches the
input patterns.
Training Algorithm
Step 0: Initialize the weights with Random values and the learning rate
Step 1: Perform Steps 2-8 when stopping condition is false.
Step 2: Perform Steps 3-5 for each input vector x.
Step 3: Compute the square of the Euclidean distance, i.e., for each j = i to m,
Training Algorithm
Step 0: Initialize the reference vectors. This can be done using the following steps. From the
given set of training vectors, take the first "m" (number of clusters) training vectors and use
them as weight vectors, the remaining vectors can be used for training.Assign the initial
weights and classifications randomly. K-means clustering method.Set initial learning rate α
Step l: Perform Steps 2-6 if the stopping condition is false.
Step 2: Perform Steps 3-4 for each training input vector x
Step 3: Calculate the Euclidean distance; for i = 1 to n, j = 1 to m,
Counterpropagation network
Counterpropagation network (CPN) were proposed by Hecht Nielsen in 1987.They are
multilayer network based on the combinations of the input, output, and clustering layers. The
application of counterpropagation net are data compression, function approximation and
pattern association. The ccounterpropagation network is basically constructed from an instar-
outstar model. This model is three layer neural network that performs input-output data
mapping, producing an output vector y in response to input vector x, on the basis of
competitive learning. The three layer in an instar-outstar model are the input layer, the
hidden(competitive) layer and the output layer.
There are two stages involved in the training process of a counterpropagation net. The input
vector are clustered in the first stage. In the second stage of training, the weights from the
cluster layer units to the output units are tuned to obtain the desired response.
There are two types of counterpropagation network:
If Euclidean distance method is used, find the cluster unit Zj whose squared
distance from input vectors is the smallest
If there occurs a tie in case of selection of winner unit, the unit with the smallest
index is the winner. Take the winner unit index as J.
Step 8: Perform Steps 9-15 when stopping condition is false for phase-II training.
Step 9: Perform Steps 10-13 for each training input Pair x:y..
Step 10: Set X-input layer activations to vector X. Sec Y-outpur layer activations to vector
Y.
Step 11: Find the winning cluster unit (use formulas from Step 4). Take the winner unit index
as J.
Step 12: Update the weights entering into unit ZJ,
Adaptive Resonance Theory
The Adaptive Resonance Theory (ART) was incorporated as a hypothesis for human cognitive data
handling. The hypothesis has prompted neural models for pattern recognition and unsupervised
learning. ART system has been utilized to clarify different types of cognitive and brain data.
The Adaptive Resonance Theory addresses the stability-plasticity(stability can be defined as the
nature of memorizing the learning and plasticity refers to the fact that they are flexible to gain new
information) dilemma of a system that asks how learning can proceed in response to huge input
patterns and simultaneously not to lose the stability for irrelevant patterns. Other than that, the
stability-elasticity dilemma is concerned about how a system can adapt new data while keeping what
was learned before. For such a task, a feedback mechanism is included among the ART neural
network layers. In this neural network, the data in the form of processing elements output reflects
back and ahead among layers. If an appropriate pattern is build-up, the resonance is reached, then
adaption can occur during this period.
It can be defined as the formal analysis of how to overcome the learning instability accomplished by a
competitive learning model, let to the presentation of an expended hypothesis, called adaptive
resonance theory (ART). This formal investigation indicated that a specific type of top-down learned
feedback and matching mechanism could significantly overcome the instability issue. It was
understood that top-down attentional mechanisms, which had prior been found through an
investigation of connections among cognitive and reinforcement mechanisms, had similar
characteristics as these code-stabilizing mechanisms. In other words, once it was perceived how to
solve the instability issue formally, it also turned out to be certain that one did not need to develop any
quantitatively new mechanism to do so. One only needed to make sure to incorporate previously
discovered attentional mechanisms. These additional mechanisms empower code learning to self-
stabilize in response to an essentially arbitrary input system. Grossberg presented the basic principles
of the adaptive resonance theory. A category of ART called ART1 has been described as an
arrangement of ordinary differential equations by carpenter and Grossberg. These theorems can
predict both the order of search as the function of the learning history of the system and the input
patterns.
ART1 is an unsupervised learning model primarily designed for recognizing binary patterns. It
comprises an attentional subsystem, an orienting subsystem, a vigilance parameter, and a reset
module, as given in the figure given below. The vigilance parameter has a huge effect on the system.
High vigilance produces higher detailed memories. The ART1 attentional comprises of two
competitive networks, comparison field layer L1 and the recognition field layer L2, two control gains,
Gain1 and Gain2, and two short-term memory (STM) stages S1 and S2. Long term memory (LTM)
follows somewhere in the range of S1 and S2 multiply the signal in these pathways.
Gains control empowers L1 and L2 to recognize the current stages of the running cycle. STM
reset wave prevents active L2 cells when mismatches between bottom-up and top-down signals
happen at L1. The comparison layer gets the binary external input passing it to the recognition layer
liable for coordinating it to a classification category. This outcome is given back to the comparison
layer to find out when the category coordinates the input vector. If there is a match, then a new input
vector is read, and the cycle begins once again. If there is a mismatch, then the orienting system is in
charge of preventing the previous category from getting a new category match in the recognition
layer. The given two gains control the activity of the recognition and the comparison layer,
respectively. The reset wave specifically and enduringly prevents active L2 cell until the current is
stopped. The offset of the input pattern ends its processing L1 and triggers the offset of Gain2. Gain2
offset causes consistent decay of STM at L2 and thereby prepares L2 to encode the next input pattern
without bais.
It can be coordinated and utilized with different techniques to give more precise outcomes.
It can be used in different fields such as face recognition, embedded system, and robotics,
target recognition, medical diagnosis, signature verification, etc.
t shows stability and is not disturbed by a wide range of inputs provided to inputs.
It has got benefits over competitive learning. The competitive learning cant include new
clusters when considered necessary.
Application of ART:
ART stands for Adaptive Resonance Theory. ART neural networks used for fast, stable
learning and prediction have been applied in different areas. The application incorporates
target recognition, face recognition, medical diagnosis, signature verification, mobile control
robot.
Target recognition:
Fuzzy ARTMAP neural network can be used for automatic classification of targets depend on
their radar range profiles. Tests on synthetic data show the fuzzy ARTMAP can result in
substantial savings in memory requirements when related to k nearest neighbor(kNN)
classifiers. The utilization of multiwavelength profiles mainly improves the performance of
both kinds of classifiers.
Medical diagnosis:
Signature verification:
Automatic signature verification is a well known and active area of research with various
applications such as bank check confirmation, ATM access, etc. the training of the network is
finished using ART1 that uses global features as input vector and the verification and
recognition phase uses a two-step process. In the initial step, the input vector is coordinated
with the stored reference vector, which was used as a training set, and in the second step,
cluster formation takes place.
Nowadays, we perceive a wide range of robotic devices. It is still a field of research in their program
part, called artificial intelligence. The human brain is an interesting subject as a model for such an
intelligent system. Inspired by the structure of the human brain, an artificial neural emerges. Similar
to the brain, the artificial neural network contains numerous simple computational units, neurons that
are interconnected mutually to allow the transfer of the signal from the neurons to neurons. Artificial
neural networks are used to solve different issues with good outcomes compared to other decision
algorithms.