0% found this document useful (0 votes)
1 views

Introduction to Deep Learning

Deep learning is a subset of machine learning that uses hierarchical representations to learn complex data patterns, with applications in image recognition, natural language processing, and finance. Key types of deep learning algorithms include Convolutional Neural Networks (CNNs), Deep Reinforcement Learning, and Recurrent Neural Networks (RNNs), each suited for specific tasks. Advantages of deep learning over traditional methods include automatic feature learning, improved accuracy, and the ability to handle complex data relationships.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
1 views

Introduction to Deep Learning

Deep learning is a subset of machine learning that uses hierarchical representations to learn complex data patterns, with applications in image recognition, natural language processing, and finance. Key types of deep learning algorithms include Convolutional Neural Networks (CNNs), Deep Reinforcement Learning, and Recurrent Neural Networks (RNNs), each suited for specific tasks. Advantages of deep learning over traditional methods include automatic feature learning, improved accuracy, and the ability to handle complex data relationships.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Introduction to Deep Learning :

Practically, Deep Learning is a subset of Machine Learning that achieves


great power and flexibility by learning to represent the world as a nested
hierarchy of concepts, with each concept defined in relation to simpler
concepts, and more abstract representations computed in terms of less
abstract ones.

Elaborately, a deep learning technique learns categories incrementally


through its hidden layer architecture, defining low-level categories like
letters first then little higher level categories like words and then higher
level categories like sentences. In the example of image recognition it
means identifying light/dark areas before categorizing lines and then
shapes to allow face recognition. Each neuron or node in the network
represents one aspect of the whole and together they provide a full
representation of the image. Each node or hidden layer is given a weight
that represents the strength of its relationship with the output and as the
model develops the weights are adjusted.

Deep Learning Architecture


Types of Deep Learning :

Some common types of deep learning algorithms include :

Convolutional neural networks(CNNs) :


CNNs are used for image recognition and processing. They are particularly
good at identifying objects in images, even when those objects are partially
obscured or distorted.

Deep reinforcement learning :


Deep reinforcement learning is used for robotics and game playing. It is a
type of machine learning that allows an agent to learn how to behave in an
environment by interacting with it and receiving rewards or punishments.
Recurrent neural networks(RNNs) :

RNNs are used for natural language processing and speech recognition. They
are particularly good at understanding the context of a sentence or phrase,
and they can be used to generate text or translate languages.

Applications of Deep Learning :

● Deep learning can be used in a wide variety of applications, including:


● Image recognition: To identify objects and features in images, such
as people, animals, places, etc.
● Natural language processing: To help understand the meaning of text,
such as in customer service chatbots and spam filters.
● Finance: To help analyze financial data and make predictions about
market trends
● Text to image: Convert text into images, such as in the Google
Translate app
Advantages of Deep Learning over traditional Machine
Learning algorithms:
1. One of the biggest advantages of deep learning is its ability to learn
complex representations of data automatically. In traditional machine
learning, experts in the domain must manually engineer features to
represent the input data. This process can be time-consuming and
may require specialized knowledge of the domain. Deep learning
eliminates this need by allowing the model to learn hierarchical
representations of the data automatically. This ability to learn
complex representations can lead to better performance on tasks such
as image and speech recognition.
2. Machine learning systems can be set up and operate quickly but may
be limited in the power of their results. Deep learning systems take
more time to set up but can generate results instantaneously
(although the quality is likely to improve over time as more data
becomes available).
3. Deep learning models have shown impressive results on many tasks,
achieving state-of-the-art accuracy levels on tasks such as image
recognition, speech recognition, and natural language processing.
This improved accuracy is due in part to the ability of deep neural
networks to capture more complex relationships between the input
data and the output predictions. Deep learning models can often
outperform traditional machine learning models, even when trained
on smaller datasets.
4. In traditional machine learning, feature engineering is often a critical
part of the modeling process. Experts must manually engineer
features that are relevant to the problem at hand. Deep learning
eliminates the need for feature engineering by allowing the model to
learn features automatically. This can save time and effort in
developing machine learning models, and can also lead to more
accurate models since the features learned are specific to the data at
hand.
Basic Concepts of Deep Learning :

Artificial Neural Networks :

Artificial Neural Networks contain artificial neurons which are called units.
These units are arranged in a series of layers that together constitute the
whole Artificial Neural Network in a system. A layer can have only a dozen
units or millions of units as this depends on how the complex neural
networks will be required to learn the hidden patterns in the dataset.
Commonly, Artificial Neural Networks have an input layer, an output layer
as well as hidden layers. The input layer receives data from the outside
world which the neural network needs to analyze or learn about. Then this
data passes through one or multiple hidden layers that transform the input
into data that is valuable for the output layer. Finally, the output layer
provides an output in the form of a response of the Artificial Neural
Networks to input data provided.

Neural Networks Architecture


In the majority of neural networks, units are interconnected from one layer
to another. Each of these connections has weights that determine the
influence of one unit on another unit. As the data transfers from one unit to
another, the neural network learns more and more about the data which
eventually results in an output from the output layer.

The structures and operations of human neurons serve as the basis for
artificial neural networks. It is also known as neural networks or neural
nets. The input layer of an artificial neural network is the first layer, and it
receives input from external sources and releases it to the hidden layer,
which is the second layer. In the hidden layer, each neuron receives input
from the previous layer neurons, computes the weighted sum, and sends it
to the neurons in the next layer. These connections are weighted means
effects of the inputs from the previous layer are optimized more or less by
assigning different-different weights to each input and it is adjusted during
the training process by optimizing these weights for improved model
performance.

Activation Function :

An Activation Function decides whether a neuron should be activated or


not. This means that it will decide whether the neuron’s input to the
network is important or not in the process of prediction using simpler
mathematical operations.

The primary role of the Activation Function is to transform the summed


weighted input from the node into an output value to be fed to the next
hidden layer or as output.
Need of Activation function in Neural Network :

The purpose of an activation function is to add non-linearity to the neural


network.Activation functions introduce an additional step at each layer
during the forward propagation.

What if we do not have use an Activation function:

In that case, every neuron will only be performing a linear transformation


on the inputs using the weights and biases. It’s because it doesn’t matter
how many hidden layers we attach in the neural network; all layers will
behave in the same way because the composition of two linear functions is a
linear function itself.

Although the neural network becomes simpler, learning any complex task is
impossible, and our model would be just a linear regression model.

Backpropagation :

In machine learning, backpropagation is an effective algorithm used to


train artificial neural networks, especially in feed-forward neural networks.
Backpropagation is an iterative algorithm that helps to minimize the cost
function by determining which weights and biases should be adjusted.
During every epoch, the model learns by adapting the weights and biases to
minimize the loss by moving down toward the gradient of the error. Thus, it
involves the two most popular optimization algorithms, such as gradient
descent or stochastic gradient descent.

Computing the gradient in the backpropagation algorithm helps to


minimize the cost function and it can be implemented by using the
mathematical rule called chain rule from calculus to navigate through
complex layers of the neural network.

A simple illustration of how the backpropagation works by adjustments of weights ]

You might also like