0% found this document useful (0 votes)
43 views

MLT unit 4 (1)

ml notes

Uploaded by

ANMOL RATAN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
43 views

MLT unit 4 (1)

ml notes

Uploaded by

ANMOL RATAN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 15

GREATER NOIDA INSTITUTE OF TECHNOLOGY, GREATERNOIDA

DEPARTMENT OF COMPUTER SCIENCE & ENGINEERING


3 RD YEAR 5TH SEM
MACHINE LEARNING TECHNIQUES
UNIT -IV

UNIT-4

Artificial Neural Network (ANN) with different layers:

A neural network contains the following three layers :


a. Input layer : The activity of the input units represents the raw information that can
feed into the network.
b. Hidden layer :
i. Hidden layer is used to determine the activity of each hidden unit.
ii. The activities of the input units and the weights depend on the connections between
the input and the hidden units.
iii. There may be one or more hidden layers.
c. Output layer : The behaviour of the output units depends on the activity of the hidden
units and the weights between the hidden and output units.

Advantages of Artificial Neural Networks (ANN) :


1. Problems in ANN are represented by attribute-value pairs.
2. ANNs are used for problems having the target function, output may be discrete-valued, real-
valued, or a vector of several real or discrete-valued attributes.
3. ANNs learning methods are quite robust to noise in the training data. The training examples
may contain errors, which do not affect the final output.
4. It is used where the fast evaluation of the learned target function required.
5. ANNs can bear long training times depending on factors such as the number of weights in
the network, the number of training examples considered, and the settings of various
learning algorithm parameters.
Disadvantages of Artificial Neural Networks (ANN) :
1. Hardware dependence :
a. Artificial neural networks require processors with parallel processing power, by their
structure.
b. For this reason, the realization of the equipment is dependent.
2. Unexplained functioning of the network :
a. This is the most important problem of ANN.
b. When ANN gives a probing solution, it does not give a clue as to why and how.
c. This reduces trust in the network.
3. Assurance of proper network structure :
a. There is no specific rule for determining the structure of artificial neural networks.
b. The appropriate network structure is achieved through experience and trial and error.
4. The difficulty of showing the problem to the network :
a. ANNs can work with numerical information.
b. Problems have to be translated into numerical values before being introduced to ANN.
c. The display mechanism to be determined will directly influence the performance of the
network.
d. This is dependent on the user’s ability.
Characteristics of Artificial Neural Network are :

1. It is neurally implemented mathematical model.


It contains large number of interconnected processing elements called neurons to do all the
operations
2. Information stored in the neurons is basically the weighted linkage of neurons.
3. The input signals arrive at the processing elements through connections and connecting
weights.
4. It has the ability to learn, recall and generalize from the given data by suitable assignment
and adjustment of weights.
5. The collective behaviour of the neurons describes its computational power, and no single
neuron carries specific information.
Application areas of artificial neural network are :
1. Speech recognition :
a. Speech occupies a prominent role in human-human interaction.
b. Therefore, it is natural for people to expect speech interfaces with computers.
c. In the present era, for communication with machines, humans still need sophisticated
languages which are difficult to learn and use.
d. To ease this communication barrier, a simple solution could be communication in a
spoken language that is possible for the machine to understand.
e. Hence, ANN is playing a major role in speech recognition.
2. Character recognition :
a. It is a problem which falls under the general area of Pattern Recognition.
b. Many neural networks have been developed for automatic recognition of
handwritten characters, either letters or digits.
3. Signature verification application :
a. Signatures are useful ways to authorize and authenticate a person in legal transactions.
b. Signature verification technique is a non-vision based technique.
c. For this application, the first approach is to extract the feature or rather the geometrical
feature set representing the signature.
d. With these feature sets, we have to train the neural networks using an efficient neural
network algorithm.
e. This trained neural network will classify the signature as being genuine or forged under
the verification stage.
4. Human face recognition :

Different types of neuron connection are :


1. Single-layer feed forward network :
a. In this type of network, we have only two layers i.e., input layer and output layer but
input layer does not count because no computation is performed in this layer.
b. Output layer is formed when different weights are applied on input nodes and the
cumulative effect per node is taken.
c. After this the neurons collectively give the output layer to compute the output signals.

2. Multilayer feed forward network :


a. This layer has hidden layer which is internal to the network and has no direct contact
with the external layer.
b. Existence of one or more hidden layers enables the network to be computationally
stronger.
c. There are no feedback connections in which outputs of the model are fed back into
itself.

Input layer Hidden layer Output layer

x1 w11 y1 v11 z1

w12 v12
w21 v21
x2 w22 y2 z2
wn1 vv1m
22
w1m vk1
wn2 v2m
vk2
w2m
xn wnm yk vmk zm

3. Single node with its own feedback :


a. When outputs can be directed back as inputs to the same layer or preceding layer nodes,
then it results in feedback networks.
b. Recurrent networks are feedback networks with closed loop. Fig. 4.5.1 shows a single
recurrent network having single neuron with feedback to itself.

Input Output

Feedback
4. Single-layer recurrent network :
a. This network is single layer network with feedback connection in which processing
element’s output can be directed back to itself or to other processing element or both.
b. Recurrent neural network is a class of artificial neural network where connections
between nodes form a directed graph along a sequence.
c. This allows it to exhibit dynamic temporal behaviour for a time sequence. Unlike feed
forward neural networks, RNNs can use their internal state (memory) to process
sequences of inputs.
5. Multilayer recurrent network :
a. In this type of network, processing element output can be directed to the processing
element in the same layer and in the preceding layer forming a multilayer recurrent
network.
b. They perform the same task for every element of a sequence, with the output being
depended on the previous computations. Inputs are not needed at each time step.
c. The main feature of a multilayer recurrent neural network is its hidden state, which
captures information about a sequence.
Gradient Descent:
Gradient descent is an optimization algorithm used to minimize some function by
iteratively moving in the direction of steepest descent as defined by the negative of the
gradient. A gradient is the slope of a function, the degree of change of a parameter with the
amount of change in another parameter. Mathematically, it can be described as the partial
derivatives of a set of parameters with respect to its inputs. The more the gradient, the steeper
the slope. Gradient Descent is a convex function.
Gradient Descent can be described as an iterative method which is used to find the values of
the parameters of a function that minimizes the cost function as much as possible. The
parameters are initially defined a particular value and from that, Gradient Descent run in an
iterative fashion to find the optimal values of the parameters, using calculus, to find the
minimum possible value of the given cost function.

Different types of gradient descent are:


1. Batch gradient descent :
a. This is a type of gradient descent which processes all the training examples for each
iteration of gradient descent.
b. When the number of training examples is large, then batch gradient descent is
computationally very expensive. So, it is not preferred.
Instead, we prefer to use stochastic gradient descent or mini-batch gradient descent

2. Stochastic gradient descent:


This is a type of gradient descent which processes single training example per iteration
a. Hence, the parameters are being updated even after one iteration in which only a single
example has been processed.
b. Hence, this is faster than batch gradient descent. When the number of training examples is
large, even then it processes only one example which can be additional overhead for
the system as the number of iterations will be large.
3. Mini-batch gradient descent:
a. This is a mixture of both stochastic and batch gradient descent.
The training set is divided into multiple groups called batches.
b.
Each batch has a number of training samples in it.
c.
d. At a time, a single batch is passed through the network which computes the loss of
every sample in the batch and uses their average to update the parameters of the neural
network.
Advantages of batch gradient descent :
1. Less oscillations and noisy steps taken towards the global minima of the loss function due to
updating the parameters by computing the average of all the training samples rather than the
value of a single sample.
2. It can benefit from the vectorization which increases the speed of processing all training
samples together.
3. It produces a more stable gradient descent convergence and stable error gradient than
stochastic gradient descent.
4. It is computationally efficient as all computer resources are not being used to process a
single sample rather are being used for all training samples.
Disadvantages of batch gradient descent :
1. Sometimes a stable error gradient can lead to a local minima and unlike stochastic gradient
descent no noisy steps are there to help to get out of the local minima.
2. The entire training set can be too large to process in the memory due to which additional
memory might be needed.
3. Depending on computer resources it can take too long for processing all the training samples
as a batch.
Advantages of stochastic gradient descent :
1. It is easier to fit into memory due to a single training sample being processed by the
network.
2. It is computationally fast as only one sample is processed at a time.
3. For larger datasets it can converge faster as it causes updates to the parameters more
frequently.
4. Due to frequent updates the steps taken towards the minima of the loss function have
oscillations which can help getting out of local minimums of the loss function (in case the
computed position turns out to be the local minimum).
Disadvantages of stochastic gradient descent :
1. Due to frequent updates the steps taken towards the minima are very noisy. This can often
lead the gradient descent into other directions.
2. Also, due to noisy steps it may take longer to achieve convergence to the minima of the loss
function.
3. Frequent updates are computationally expensive due to using all resources for processing
one training sample at a time.
It loses the advantage of vectorized operations as it deals with only a single example at a
time.

Delta rule :
1. The delta rule is specialized version of backpropagation’s learning rule that uses single layer
neural networks.
2. It calculates the error between calculated output and sample output data, and uses this to
create a modification to the weights, thus implementing a form of gradient descent.
Generalized delta learning rule (Error backpropagation learning) :
In generalized delta learning rule (error backpropagation learning). We are given the training
set :
{x1, y1), ..., (xk, yk)

where xk = [xk1, ... xnk] and yk  R, k = 1, ..., K.


Step 1 :  > 0, Emax > 0 are chosen.
Step 2 : Weights w are initialized at small random values, k = 1, and the running error E is set
to 0.
Step 3 : Input xk is presented, x : = xk, y : = yk, and output O is computed as :

1
O=

1  exp( WT O)

where Ot is the output vector of the hidden layer :


1
O=

l
l
1  exp( WT x)

Step 4 : Weights of the output unit are updated


W:=W+ o

where = (y – O)O(1 – O)
Step 5 : Weights of the hidden units are updated
wt = wt + WtOt(1 – Ot)x, l = 1, ..., L
Step 6 : Cumulative cycle error is computed by adding the present error to E
E := E + 1/2(y – O)2
Step 7 : If k < K then k := k + 1 and we continue the training by going back to step 2,
otherwise we go to step 8.
Step 8 : The training cycle is completed. For E < Emax terminate the training session. If E
> Emax then E : = 0, k := 1 and we initiate a new training cycle by going back to step 3.

Backpropagation Algorithm:
1. Backpropagation is an algorithm used in the training of feedforward neural networks for
supervised learning.
2. Backpropagation efficiently computes the gradient of the loss function with respect to the
weights of the network for a single input-output example.
3. This makes it feasible to use gradient methods for training multi-layer networks, updating
weights to minimize loss, we use gradient descent or variants such as stochastic gradient
descent.
4. The backpropagation algorithm works by computing the gradient of the loss function with
respect to each weight by the chain rule, iterating backwards one layer at a time from the
last layer to avoid redundant calculations of intermediate terms in the chain rule; this is an
example of dynamic programming.
5. The term backpropagation refers only to the algorithm for computing the gradient, but it is
often used loosely to refer to the entire learning algorithm, also including how the gradient is
used, such as by stochastic gradient descent.
6. Backpropagation generalizes the gradient computation in the delta rule, which is the single-
layer version of backpropagation, and is in turn generalized by automatic differentiation,
where backpropagation is a special case of reverse accumulation (reverse mode).

Selection of various parameters in BPN :


1. Number of hidden nodes :
a. The guiding criterion is to select the minimum nodes in the first and third
layer, so that the memory demand for storing the weights can be kept
minimum.
b. The number of separable regions in the input space M, is a function of the
number of hidden nodes H in BPN and H = M – 1.
c. When the number of hidden nodes is equal to the number of training patterns,
the learning could be fastest.
d. In such cases, BPN simply remembers training patterns losing all
generalization capabilities.
e. Hence, as far as generalization is concerned, the number of hidden nodes
should be small compared to the number of training patterns with help of
Vapnik Chervonenkis dimension (VCdim) of probability theory.
f. We can estimate the selection of number of hidden nodes for a given
number of training patterns as number of weights which is equal to I1 * I2
+ I2 * I3, where I1 and I3 denote input and output nodes and I2 denote
hidden nodes.
g. Assume the training samples T to be greater than VCdim. Now if we accept
the ratio 10 : 1
2. Momentum coefficient :
a. To reduce the training time we use the momentum factor because it
enhances the training process.
b. The influences of momentum on weight change is

[ W]n+1 = E [ W ]n
W
c. The momentum also overcomes the effect of local minima.
d. The use of momentum term will carry a weight change process through
one or local minima and get it into global minima.
3. Sigmoidal gain :
a. When the weights become large and force the neuron to operate in a region
where sigmoidal function is very flat, a better method of coping with
network paralysis is to adjust the sigmoidal gain.
b. By decreasing this scaling factor, we effectively spread out sigmoidal
function on wide range so that training proceeds faster.
4. Local minima :
a. One of the most practical solutions involves the introduction of a shock
which changes all weights by specific or random amounts.
b. If this fails, then the most practical solution is to rerandomize the weights
and start the training all over.
Perceptron:

1. The perceptron is the simplest form of a neural network used for classification of patterns
said to be linearly separable.
2. It consists of a single neuron with adjustable synaptic weights and bias.
3. The perceptron build around a single neuron is limited for performing pattern classification
with only two classes.
4. By expanding the output layer of perceptron to include more than one neuron, more than
two classes can be classified.
5. Suppose, a perceptron have synaptic weights denoted by w1, w2, w3, …..
wm.

6.The input applied to the perceptron are denoted by x1, x2, …… xm.
7.The externally applied bias is denoted by b.
Multilayer perceptron :
1. The perceptrons which are arranged in layers are called multilayer perceptron. This model
has three layers : an input layer, output layer and hidden layer.
2. For the perceptrons in the input layer, the linear transfer function used and for the perceptron
in the hidden layer and output layer, the sigmoidal or squashed-S function is used.
3. The input signal propagates through the network in a forward direction.
4. On a layer by layer basis, in the multilayer perceptron bias b(n) is treated as a synaptic weight
driven by fixed input equal to +1.
x(n) = [+1, x1(n), x2(n), xm(n)]T

where n denotes the iteration step in applying the algorithm. Correspondingly, we define the
weight vector as :
w(n) = [b(n), w1(n), w2(n), wm(n)]T
5. Accordingly, the linear combiner output is written in the compact form :
m
V(n) = wi(n)xi(n) = wT(n) × x(n)
Architecture of multilayer perceptron :
1. Fig. 4.15.1 shows architectural graph of multilayer perceptron with two hidden layer and an
output layer.
2. Signal flow through the network progresses in a forward direction, from the left to right
and on a layer-by-layer basis.
3. Two kinds of signals are identified in this network :

Functional signals : Functional signal is an input signal and propagates forward and
emerges at the output end of the network as an output signal.
Error signals: Error signal originates at an output neuron and propagates backward through
the network
Input signal
Output layer

Input layer First hidden Second hidden

Multilayer perceptrons have been applied successfully to solve some difficult and diverse
problems by training them in a supervised manner with highly popular algorithm known as
the error backpropagation algorithm.
Characteristics of multilayer perceptron :
1. In this model, each neuron in the network includes a non-linear activation function (non-
linearity is smooth). Most commonly used non-linear function is defined by :
1
y=
j 1 exp( vj )

where vj is the induced local field (i.e., the sum of all weights and bias) and y is the output
of neuron j.
2. The network contains hidden neurons that are not a part of input or output of the network.
Hidden layer of neurons enabled network to learn complex tasks.
3. The network exhibits a high degree of connectivity.

Self-Organizing Map (SOM) :


1. Self-Organizing Map (SOM) provides a data visualization technique which helps to understand
high dimensional data by reducing the dimensions of data to a map.
SOM also represents clustering concept by grouping similar data together.
2. A self-Organizing Map (SOM) or Self-Organizing Feature Map (SOFM) is a type of
Artificial Neural Network (ANN) that is trained using unsupervised learning to produce a
low-dimensional (typically two- dimensional), discretized representation of the input space of
the training samples, called a map, and is therefore a method to do dimensionality reduction.
Self-organizing maps differ from other artificial neural networks as they apply competitive
learning as opposed to error-correction learning (such as backpropagation with gradient
descent), and in the sense that they use a neighborhood function to preserve the topological
properties of the input space.
Following are the steps used in SOM algorithm :
1. Each node’s weights are initialized.
2. A vector is chosen at random from the set of training data.
3. Every node is examined to calculate which one’s weights are most like the input vector. The
winning node is commonly known as the Best Matching Unit (BMU).
4. Then the neighbourhood of the BMU is calculated. The amount of neighbors decreases
over time.
5. The winning weight is rewarded with becoming more like the sample vector. The
neighbours also become more like the sample vector. The closer a node is to the BMU, the
more its weights get altered and the farther away the neighbor is from the BMU, the less it
learns.
Repeat step 2 for N iterations.
Basics processes used in SOM algorithm are :
1. Initialization : All the connection weights are initialized with small random values.

Competition : For each input pattern, the neurons compute their respective values of a
discriminant function which provides the basis for competition. The particular neuron with
the smallest value of the discriminant function is declared the winner
2. Cooperation : The winning neuron determines the spatial location of a topological
neighbourhood of excited neurons, thereby providing the basis for cooperation among
neighbouring neurons.
3. Adaptation : The excited neurons decrease their individual values of the discriminant
function in relation to the input pattern through suitable adjustment of the associated connection
weights, such that the response of the winning neuron to the subsequent application of a
similar input pattern is enhanced.
Stages of SOM algorithm are :
1. Initialization : Choose random values for the initial weight vectors wj.
2. Sampling : Draw a sample training input vector x from the input space.
3. Matching : Find the winning neuron I(x) that has weight vector closest.
D

to the input vector, i.e., the minimum value of dj(x) = (x w )2 .


i ji
i 1

4. Updating : Apply the weight update equation


Dwji = h(t) Tj, I(x)(t) (xi – wji)

where Tj, I(x)(t) is a Gaussian neighbourhood and h(t) is the learning rate.
5. Continuation : Keep returning to step 2 until the feature map stops changing.

Deep learning:
Deep learning is the subfield of artificial intelligence that focuses on creating large neural
network models that are capable of making accurate data-driven decisions.
1. Deep learning is used where the data is complex and has large datasets.
2. Facebook uses deep learning to analyze text in online conversations. Google and Microsoft
all use deep learning for image search and machine translation.
3. All modern smart phones have deep learning systems running on them. For example, deep
learning is the standard technology for speech recognition, and also for face detection on
digital cameras.
4. In the healthcare sector, deep learning is used to process medical images (X-rays, CT, and
MRI scans) and diagnose health conditions.
5. Deep learning is also at the core of self-driving cars, where it is used for localization and
mapping, motion planning and steering, and environment perception, as well as tracking driver
state.
Different architecture of deep learning are :
1. Deep Neural Network : It is a neural network with a certain level of complexity (having
multiple hidden layers in between input and output layers). They are capable of modeling
and processing non-linear relationships.
2. Deep Belief Network (DBN) : It is a class of Deep Neural Network. It is multi-layer belief
networks. Steps for performing DBN are :
a. Learn a layer of features from visible units using Contrastive Divergence algorithm.
b. Treat activations of previously trained features as visible units and then learn features of
features.
c. Finally, the whole DBN is trained when the learning for the final hidden layer is
achieved.
3. Recurrent (perform same task for every element of a sequence) Neural Network :
Allows for parallel and sequential computation. Similar to the human brain (large feedback
network of connected neurons). They are able to remember important things about the input
they received and hence enable them to be more precise.
Advantages of deep learning:

1. Best in-class performance on problems.


2. Reduces need for feature engineering.
3. Eliminates unnecessary costs.
4. Identifies defects easily that are difficult to detect.

Disadvantages of deep learning :


1. Large amount of data required.
2. Computationally expensive to train.
3. No strong theoretical foundation.

Limitations of deep learning :


1. Learning through observations only.
2. The issue of biases.
Following are the application of deep learning:
1. Automatic text generation: Corpus of text is learned and from this model new text is
generated, word-by-word or character-by-character. Then this model is capable of learning
how to spell, punctuate, form sentences, or it may even capture the style.
2. Healthcare : Helps in diagnosing various diseases and treating it.
3. Automatic machine translation : Certain words, sentences or phrases in one language is
transformed into another language (Deep Learning is achieving top results in the areas of
text, images).
4. Image recognition : Recognizes and identifies peoples and objects in images as well as to
understand content and context. This area is already being used in Gaming, Retail, Tourism,
etc.
Convolutional networks:
Convolutional networks also known as Convolutional Neural Networks (CNNs) are a
specialized kind of neural network for processing data that has a known, grid-like topology.
Convolutional neural network indicates that the network employs a mathematical operation
called convolution.Convolution is a specialized kind of linear operation.Convolutional
networks are simply neural networks that use convolution in place of general matrix
multiplication in at least one of their layers.CNNs, (ConvNets), are quite similar to regular
neural networks.
They are still made up of neurons with weights that can be learned from data. Each neuron
receives some inputs and performs a dot product.They still have a loss function on the last
fully connected layer.They can still use a non-linearity function a regular neural network
receives input data as a single vector and passes through a series of hidden layers.
1. Every hidden layer consists of neurons, wherein every neuron is fully connected to all the
other neurons in the previous layer.
2. Within a single layer, each neuron is completely independent and they do not share any
connections.
The fully connected layer, (the output layer), contains class scores in the case of an image
classification problem. There are three main layers in a simple ConvNet.

Convolutional layers:
1. Convolutional layers are the major building blocks used in convolutional neural networks.
2. A convolution is the simple application of a filter to an input that results in an activation.
3. Repeated application of the same filter to an input results in a map of activations called a
feature map, indicating the locations and strength of a detected feature in an input, such as an
image.
4. The innovation of convolutional neural networks is the ability to automatically learn a
large number of filters in parallel specific to a training dataset under the constraints of a
specific predictive modeling problem, such as image classification.
5. The result is highly specific features that can be detected anywhere on input images.

Activation function :
1. An activation function is a function that is added into an artificial neural network in order to
help the network learn complex patterns in the data.
2. When comparing with a neuron-based model that is in our brains, the activation function is
at the end deciding what is to be fired to the next neuron.
3. That is exactly what an activation function does in an ANN as well.
4. It takes in the output signal from the previous cell and converts it into some form that can be
taken as input to the next cell.
Pooling layer :
1. A pooling layer is a new layer added after the convolutional layer. Specifically, after a non-
linearity (for example ReLU) has been applied to the feature maps output by a
convolutional layer, for example, the layers in a model may look as follows :
a. Input image
b. Convolutional layer
c. Non linearlity
d. Pooling layer
2. The addition of a pooling layer after the convolutional layer is a common pattern used for
ordering layers within a convolutional neural network that may be repeated one or more
times in a given model.
3. The pooling layer operates upon each feature map separately to create a new set of the same
number of pooled feature maps.
Fully connected layer :
1. Fully connected layers are an essential component of Convolutional Neural Networks
(CNNs), which have been proven very successful in recognizing and classifying images for
computer vision.
2. The CNN process begins with convolution and pooling, breaking down the image into
features, and analyzing them independently.
3. The result of this process feeds into a fully connected neural network structure that drives
the final classification decision.

1D convolutional neural network :


Convolutional Neural Network (CNN) models were developed for image classification, in
which the model accepts a two-dimensional input representing an image's pixels and color
channels, in a process called feature learning.
1. This same process can be applied to one-dimensional sequences of data.
2. The model extracts features from sequences data and maps the internal features of the
sequence.
3. A 1D CNN is very effective for deriving features from a fixed-length segment of the
overall dataset, where it is not so important where the feature is located in the segment.
4. 1D Convolutional Neural Networks work well for :
a. Analysis of a time series of sensor data.
b. Analysis of signal data over a fixed-length period, for example, an audio recording.
c. Natural Language Processing (NLP), although Recurrent Neural Networks which
leverage Long Short Term Memory (LSTM) cells are more promising than CNN as
they take into account the proximity of words to create trainable patterns.
2D convolutional neural network :
1. In a 2D convolutional network, each pixel within the image is represented by its x and y position
as well as the depth, representing image channels (red, green, and blue).
2. It moves over the images both horizontally and vertically.

Training Network:

1.Once a network has been structured for a particular application, that network is ready to be
trained.
2.To start this process the initial weights are chosen randomly. Then, the training, or learning
begins.
3. There are two approaches to training :
• In supervised training, both the inputs and the outputs are provided. The network then
processes the inputs and compares its resulting outputs against the desired outputs.
• Errors are then propagated back through the system, causing the system to adjust the
weights which control the network. This process occurs over and over as the
weights are continually tweaked.
• The set of data which enables the training is called the “training set.” During the
training of a network the same set of data is processed many times as the connection
weights are ever refined.
• The other type of training is called unsupervised training. In unsupervised training, the
network is provided with inputs but not with desired outputs.
• The system itself must then decide what features it will use to group the input data. This
is often referred to as self-organization or adaption.
Case Study:
1. Diabetic Retinopathy (DR) is one of the major causes of blindness in the western world.
Increasing life expectancy, indulgent lifestyles and other contributing factors mean the number
of people with diabetes is projected to continue rising.
2. Regular screening of diabetic patients for DR has been shown to be a cost-effective and
important aspect of their care.
3. The accuracy and timing of this care is of significant importance to both the cost and
effectiveness of treatment.
4. If detected early enough, effective treatment of DR is available; making this a vital process.
5. Classification of DR involves the weighting of numerous features and the location of such
features. This is highly time consuming for clinicians.
6. Computers are able to obtain much quicker classifications once trained, giving the ability to
aid clinicians in real-time classification.
7. The efficacy of automated grading for DR has been an active area of research in computer
imaging with encouraging conclusions.
8. Significant work has been done on detecting the features of DR using automated methods
such a support vector machines and k-NN classifiers.
9. The majority of these classification techniques arc on two class classification for DR or
no DR.

Self Driven Car using AI:

1. The rapid development of the Internet economy and Artificial Intelligence (AI) has promoted
the progress of self-driving cars.
2. The market demand and economic value of self-driving cars are increasingly prominent.
At present, more and more enterprises and scientific research institutions have invested in
this field. Google, Tesla, Apple, Nissan, Audi, General Motors, BMW, Ford, Honda,
Toyota, Mercedes, and Volkswagen have participated in the research and development of
self-driving cars.
3. Google is an Internet company, which is one of the leaders in self- driving cars, based on its
solid foundation in artificial intelligence.
4. In June 2015, two Google self-driving cars were tested on the road. So far, Google vehicles
have accumulated more than 3.2 million km of tests, becoming the closest to the actual use.
5. Another company that has made great progress in the field of self- driving cars is Tesla.
Tesla was the first company to devote self-driving technology to production.
6. Followed by the Tesla models series, its “auto-pilot” technology has made major
breakthroughs in recent years.
7. Although the Tesla's autopilot technology is only regarded as Level 2 stage by the National
Highway Traffic Safety Administration (NHTSA), Tesla shows us that the car has basically
realized automatic driving under certain conditions

You might also like