AI_slides1
AI_slides1
Fast progression of AI
Need of Data Lebling : Supervised or Unsupervised Way
Supervised Learning:
1.Create a set of labeled data, i.e. "correct" data with both input and output information:
pictures of cars and trucks, each with the corresponding class names.
The Machine Learning algorithm begins to "see" patterns between input (image) and output
(class). The algorithm might learn complex relationships as "the distance between wheels is
larger for trucks" – note that in reality, it is usually hard to interpret algorithms in such a way.
3. Test the model on unseen data and measure how accurately it predicts the class.
The term supervised learning stems from the fact that, in the beginning, we gave the algorithm
a data set in which the “correct answers” were given. This is the key difference between
unsupervised learning.
Ways of Data Lebling
Ways of Data Lebling
There are two main tasks of supervised learning:
In reinforcement learning, the algorithm (in this context also often referred to as an agent) learns
through trial-and-error using feedback from its actions. Rewards and punishment operate as
signals for desired and undesired behavior.
Examples of Supervised and
Unsupervised
Learning Algorithms
Decision Trees
• Is another classification method.
• A decision tree is a set of simple rules, such as "if the
sepal length is less than 5.45, classify the specimen as
setosa."
• Decision trees are also nonparametric because they do
not require any assumptions about the distribution of
the variables in each class.
Types of Clustering
• Types of clustering:
– HIERARCHICAL: finds successive clusters using previously
established clusters
• agglomerative (bottom‐up): start with each element in a separate cluster
and merge them accordingly to a given property
• divisive (top‐down)
– PARTITIONAL: usually determines all clusters at once
Distances
• Determine the similarity between two clusters and
the shape of the clusters.
In case of strings…
• The Hamming distance between two strings of equal length is
the number of positions at which the corresponding symbols
are different.
– measures the minimum number of substitutions required to
change one string into the other
• The Levenshtein (edit) distance is a metric for measuring the
amount of difference between two sequences.
– is defined as the minimum number of edits needed to transform
one string into the other.
RANGE (Min‐Max Normalization): subtracts the minimum value of an attribute from each value
of the attribute and then divides the difference by the range of the attribute. It has the
advantage of preserving exactly all relationship in the data, without adding any bias.
SOFTMAX: is a way of reducing the influence of extreme values or outliers in the data without
removing them from the data set. It is useful when you have outlier data that you wish to
include in the data set while still preserving the significance of data within a standard deviation
of the mean.
KMeans: how it works
Kmeans: Pro and Cons
Summary
• Finding the optimal approach
• Supervised Models
– Neural Networks
– Multi Layer Perceptron
– Decision Trees
• Unsupervised Models
– Different Types of Clustering
– Distances and Normalization
– Kmeans
• Combining different models
– Committee Machines
– Introducing a Priori Knowledge
– Sleeping Expert Framework
Deep Learning vs Classical Machine Learning
Deep Learning vs Classical Machine Learning
Deep Learning vs Classical Machine Learning
Deep Learning vs Classical Machine Learning
▪ Traditional Machine Learning algorithms have simpler structure, such as linear regression or a decision
tree,
▪ Deep Learning is based on an artificial neural network. This multi-layered ANN is, like a human brain,
complex and intertwined.
▪ DL algorithms require much less human intervention. Manually feature engineering and a classifier
selection is needed to sort images, check whether the output is as required, and adjust the algorithm if
this is not the case
▪ As a deep learning algorithm, however, the features are extracted automatically, and the algorithm
learns from its own errors.
▪ Deep Learning requires much more data than a traditional Machine Learning algorithm to function
properly. Due to the complex multi-layer structure, a deep learning system needs a large dataset to
eliminate fluctuations and make high-quality interpretations.
Sources of Data
Increasing numbers of smartphones and internet devices
Expert System
Predictive AI
Generative AI
Predictive AI
Generative AI
Selection of Model
Considerations for Opting Computer Vision
Equipment Intelligence
Energy providers are continually seeking to improve the
management of their transmission network through efficient
network
investment opportunities