This document provides an overview of machine learning and neural network techniques. It defines machine learning as the field that focuses on algorithms that can learn. The document discusses several key components of a machine learning model, including what is being learned (the domain) and from what information the learner is learning. It then summarizes several common machine learning algorithms like k-NN, Naive Bayes classifiers, decision trees, reinforcement learning, and the Rocchio algorithm for relevance feedback in information retrieval. For each technique, it provides a brief definition and examples of applications.
This document provides an overview of machine learning techniques for classification and anomaly detection. It begins with an introduction to machine learning and common tasks like classification, clustering, and anomaly detection. Basic classification techniques are then discussed, including probabilistic classifiers like Naive Bayes, decision trees, instance-based learning like k-nearest neighbors, and linear classifiers like logistic regression. The document provides examples and comparisons of these different methods. It concludes by discussing anomaly detection and how it differs from classification problems, noting challenges like having few positive examples of anomalies.
Methodological Study Of Opinion Mining And Sentiment Analysis Techniques ijsc
Decision making both on individual and organizational level is always accompanied by the search of other’s opinion on the same. With tremendous establishment of opinion rich resources like, reviews, forum discussions, blogs, micro-blogs, Twitter etc provide a rich anthology of sentiments. This user generated content can serve as a benefaction to market if the semantic orientations are deliberated. Opinion mining and sentiment analysis are the formalization for studying and construing opinions and sentiments. The digital ecosystem has itself paved way for use of huge volume of opinionated data recorded. This paper is an attempt to review and evaluate the various techniques used for opinion and sentiment analysis.
Methodological study of opinion mining and sentiment analysis techniquesijsc
Decision making both on individual and organizational level is always accompanied by the search of
other’s opinion on the same. With tremendous establishment of opinion rich resources like, reviews, forum
discussions, blogs, micro-blogs, Twitter etc provide a rich anthology of sentiments. This user generated
content can serve as a benefaction to market if the semantic orientations are deliberated. Opinion mining
and sentiment analysis are the formalization for studying and construing opinions and sentiments. The
digital ecosystem has itself paved way for use of huge volume of opinionated data recorded. This paper is
an attempt to review and evaluate the various techniques used for opinion and sentiment analysis.
The document provides an overview of neural networks. It begins by discussing biological inspiration from the human brain, including key facts about neurons and synapses. It then defines artificial neurons and various components like dendrites, axons, and synapses. The document explores different types of neural networks including feedforward, recurrent, self-organizing maps and time delay neural networks. It also covers common neural network architectures, learning algorithms, activation functions, and applications of neural networks.
Machine Learning and Artificial Neural Networks.pptAnshika865276
Machine learning and neural networks are discussed. Machine learning investigates how knowledge is acquired through experience. A machine learning model includes what is learned (the domain), who is learning (the computer program), and the information source. Techniques discussed include k-nearest neighbors algorithm, Winnow algorithm, naive Bayes classifier, decision trees, and reinforcement learning. Reinforcement learning involves an agent interacting with an environment to optimize outcomes through trial and error.
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document provides an overview of the CS303 Computer Algorithms course taught by Dr. Yanxia Jia. It discusses the importance of algorithms, provides examples of classic algorithm problems like sorting and searching, and summarizes common algorithm design techniques and data structures used, including arrays, linked lists, stacks, queues, heaps, graphs and trees.
Classifiers are algorithms that map input data to categories in order to build models for predicting unknown data. There are several types of classifiers that can be used including logistic regression, decision trees, random forests, support vector machines, Naive Bayes, and neural networks. Each uses different techniques such as splitting data, averaging predictions, or maximizing margins to classify data. The best classifier depends on the problem and achieving high accuracy, sensitivity, and specificity.
The document discusses various clustering algorithms and concepts:
1) K-means clustering groups data by minimizing distances between points and cluster centers, but it is sensitive to initialization and may find local optima.
2) K-medians clustering is similar but uses point medians instead of means as cluster representatives.
3) K-center clustering aims to minimize maximum distances between points and clusters, and can be approximated with a farthest-first traversal algorithm.
This document discusses object detection using Adaboost and various techniques. It begins with an overview of the Adaboost algorithm and provides a toy example to illustrate how it works. Next, it describes how Viola and Jones used Adaboost with Haar-like features and an integral image representation for rapid face detection in images. It achieved high detection rates with very low false positives. The document also discusses how Schneiderman and Kanade used a parts-based representation with localized wavelet coefficients as features for object detection and used statistical independence of parts to obtain likelihoods for classification.
Anomaly detection using deep one class classifier홍배 김
The document discusses anomaly detection techniques using deep one-class classifiers and generative adversarial networks (GANs). It proposes using an autoencoder to extract features from normal images, training a GAN on those features to model the distribution, and using a one-class support vector machine (SVM) to determine if new images are within the normal distribution. The method detects and localizes anomalies by generating a binary mask for abnormal regions. It also discusses Gaussian mixture models and the expectation-maximization algorithm for modeling multiple distributions in data.
This document provides an overview of various techniques for text categorization, including decision trees, maximum entropy modeling, perceptrons, and K-nearest neighbor classification. It discusses the data representation, model class, and training procedure for each technique. Key aspects covered include feature selection, parameter estimation, convergence criteria, and the advantages/limitations of each approach.
MS CS - Selecting Machine Learning AlgorithmKaniska Mandal
ML Algorithms usually solve an optimization problem such that we need to find parameters for a given model that minimizes
— Loss function (prediction error)
— Model simplicity (regularization)
Unit-1 Introduction and Mathematical Preliminaries.pptxavinashBajpayee1
This document provides an introduction to pattern recognition and classification. It discusses key concepts such as patterns, features, classes, supervised vs. unsupervised learning, and classification vs. clustering. Examples of pattern recognition applications are given such as handwriting recognition, license plate recognition, and medical imaging. The main phases of developing a pattern recognition system are outlined as data collection, feature choice, model choice, training, evaluation, and considering computational complexity. Finally, some relevant basics of linear algebra are reviewed.
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MININGcsandit
Although opinion mining is in a nascent stage of development but still the ground is set for dense growth of researches in the field. One of the important activities of opinion mining is to extract opinions of people based on characteristics of the object under study. Feature extraction in opinion mining can be done by various ways like that of clustering, support vector machines
etc. This paper is an attempt to appraise the various techniques of feature extraction. The first part discusses various techniques and second part makes a detailed appraisal of the major techniques used for feature extraction
Analytical study of feature extraction techniques in opinion miningcsandit
Although opinion mining is in a nascent stage of development but still the ground is set for
dense growth of researches in the field. One of the important activities of opinion mining is to
extract opinions of people based on characteristics of the object under study. Feature extraction
in opinion mining can be done by various ways like that of clustering, support vector machines
etc. This paper is an attempt to appraise the various techniques of feature extraction. The first
part discusses various techniques and second part makes a detailed appraisal of the major
techniques used for feature extraction
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...cscpconf
Although opinion mining is in a nascent stage of development but still the ground is set for dense growth of researches in the field. One of the important activities of opinion mining is to
extract opinions of people based on characteristics of the object under study. Feature extraction in opinion mining can be done by various ways like that of clustering, support vector machines
etc. This paper is an attempt to appraise the various techniques of feature extraction. The first part discusses various techniques and second part makes a detailed appraisal of the major techniques used for feature extraction.
This document provides an overview of machine learning techniques for text mining and information extraction, including supervised, unsupervised, and weakly supervised learning algorithms. It discusses support vector machines, naive Bayes models, maximum entropy models, and feature selection methods. Key machine learning approaches covered are support vector machines, naive Bayes classifiers, maximum entropy models, and the use of kernels and feature extraction for text classification tasks.
https://telecombcn-dl.github.io/2017-dlai/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks or Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles of deep learning from both an algorithmic and computational perspectives.
The document discusses various machine learning techniques including k-nearest neighbors, which classifies new data based on its similarity to existing examples; Naive Bayes classifiers, which use Bayes' theorem to classify items based on the presence or absence of features; and decision trees, which classify items by sorting them based on the results of tests on their features and dividing them into branches. Reinforcement learning is also covered, where an agent learns through trial-and-error interactions with an environment by receiving rewards or penalties for actions.
Islamic University Pattern Recognition & Neural Network 2019 Rakibul Hasan Pranto
The document discusses various topics related to pattern recognition including:
1. Pattern recognition is the automated recognition of patterns and regularities in data through techniques like machine learning. It has applications in areas like optical character recognition, diagnosis systems, and security.
2. There are two main approaches to pattern recognition - sub-symbolic and symbolic. Sub-symbolic uses connectionist models like neural networks while symbolic uses formal structures like strings and automata to represent patterns.
3. A pattern recognition system consists of steps like data acquisition, pre-processing, feature extraction, model learning, classification, and post-processing to classify patterns. Bayesian decision making and Bayes' theorem are statistical techniques used in classification.
Yulia Honcharenko "Application of metric learning for logo recognition"Fwdays
Typical approaches of solving classification problems require the collection of a dataset for each new class and retraining of the model. Metric learning allows you to train a model once and then easily add new classes with 5-10 reference images.
So we’ll talk about metric learning based on YouScan experience: task, data, different losses and approaches, metrics we used, pitfalls and peculiarities, things that worked and didn’t.
The document discusses different clustering algorithms, including k-means and EM clustering. K-means aims to partition items into k clusters such that each item belongs to the cluster with the nearest mean. It works iteratively to assign items to centroids and recompute centroids until the clusters no longer change. EM clustering generalizes k-means by computing probabilities of cluster membership based on probability distributions, with the goal of maximizing the overall probability of items given the clusters. Both algorithms are used to group similar items in applications like market segmentation.
Machine Learning and Artificial Neural Networks.pptAnshika865276
Machine learning and neural networks are discussed. Machine learning investigates how knowledge is acquired through experience. A machine learning model includes what is learned (the domain), who is learning (the computer program), and the information source. Techniques discussed include k-nearest neighbors algorithm, Winnow algorithm, naive Bayes classifier, decision trees, and reinforcement learning. Reinforcement learning involves an agent interacting with an environment to optimize outcomes through trial and error.
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document provides an overview of knowledge representation techniques and object recognition. It discusses syntax and semantics in representation, as well as descriptions, features, grammars, languages, predicate logic, production rules, fuzzy logic, semantic nets, and frames. It then covers statistical and cluster-based pattern recognition methods, feedforward and backpropagation neural networks, unsupervised learning including Kohonen feature maps, and Hopfield neural networks. The goal is to represent knowledge in a way that enables object classification and decision-making.
This document provides an overview of the CS303 Computer Algorithms course taught by Dr. Yanxia Jia. It discusses the importance of algorithms, provides examples of classic algorithm problems like sorting and searching, and summarizes common algorithm design techniques and data structures used, including arrays, linked lists, stacks, queues, heaps, graphs and trees.
Classifiers are algorithms that map input data to categories in order to build models for predicting unknown data. There are several types of classifiers that can be used including logistic regression, decision trees, random forests, support vector machines, Naive Bayes, and neural networks. Each uses different techniques such as splitting data, averaging predictions, or maximizing margins to classify data. The best classifier depends on the problem and achieving high accuracy, sensitivity, and specificity.
The document discusses various clustering algorithms and concepts:
1) K-means clustering groups data by minimizing distances between points and cluster centers, but it is sensitive to initialization and may find local optima.
2) K-medians clustering is similar but uses point medians instead of means as cluster representatives.
3) K-center clustering aims to minimize maximum distances between points and clusters, and can be approximated with a farthest-first traversal algorithm.
This document discusses object detection using Adaboost and various techniques. It begins with an overview of the Adaboost algorithm and provides a toy example to illustrate how it works. Next, it describes how Viola and Jones used Adaboost with Haar-like features and an integral image representation for rapid face detection in images. It achieved high detection rates with very low false positives. The document also discusses how Schneiderman and Kanade used a parts-based representation with localized wavelet coefficients as features for object detection and used statistical independence of parts to obtain likelihoods for classification.
Anomaly detection using deep one class classifier홍배 김
The document discusses anomaly detection techniques using deep one-class classifiers and generative adversarial networks (GANs). It proposes using an autoencoder to extract features from normal images, training a GAN on those features to model the distribution, and using a one-class support vector machine (SVM) to determine if new images are within the normal distribution. The method detects and localizes anomalies by generating a binary mask for abnormal regions. It also discusses Gaussian mixture models and the expectation-maximization algorithm for modeling multiple distributions in data.
This document provides an overview of various techniques for text categorization, including decision trees, maximum entropy modeling, perceptrons, and K-nearest neighbor classification. It discusses the data representation, model class, and training procedure for each technique. Key aspects covered include feature selection, parameter estimation, convergence criteria, and the advantages/limitations of each approach.
MS CS - Selecting Machine Learning AlgorithmKaniska Mandal
ML Algorithms usually solve an optimization problem such that we need to find parameters for a given model that minimizes
— Loss function (prediction error)
— Model simplicity (regularization)
Unit-1 Introduction and Mathematical Preliminaries.pptxavinashBajpayee1
This document provides an introduction to pattern recognition and classification. It discusses key concepts such as patterns, features, classes, supervised vs. unsupervised learning, and classification vs. clustering. Examples of pattern recognition applications are given such as handwriting recognition, license plate recognition, and medical imaging. The main phases of developing a pattern recognition system are outlined as data collection, feature choice, model choice, training, evaluation, and considering computational complexity. Finally, some relevant basics of linear algebra are reviewed.
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MININGcsandit
Although opinion mining is in a nascent stage of development but still the ground is set for dense growth of researches in the field. One of the important activities of opinion mining is to extract opinions of people based on characteristics of the object under study. Feature extraction in opinion mining can be done by various ways like that of clustering, support vector machines
etc. This paper is an attempt to appraise the various techniques of feature extraction. The first part discusses various techniques and second part makes a detailed appraisal of the major techniques used for feature extraction
Analytical study of feature extraction techniques in opinion miningcsandit
Although opinion mining is in a nascent stage of development but still the ground is set for
dense growth of researches in the field. One of the important activities of opinion mining is to
extract opinions of people based on characteristics of the object under study. Feature extraction
in opinion mining can be done by various ways like that of clustering, support vector machines
etc. This paper is an attempt to appraise the various techniques of feature extraction. The first
part discusses various techniques and second part makes a detailed appraisal of the major
techniques used for feature extraction
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...cscpconf
Although opinion mining is in a nascent stage of development but still the ground is set for dense growth of researches in the field. One of the important activities of opinion mining is to
extract opinions of people based on characteristics of the object under study. Feature extraction in opinion mining can be done by various ways like that of clustering, support vector machines
etc. This paper is an attempt to appraise the various techniques of feature extraction. The first part discusses various techniques and second part makes a detailed appraisal of the major techniques used for feature extraction.
This document provides an overview of machine learning techniques for text mining and information extraction, including supervised, unsupervised, and weakly supervised learning algorithms. It discusses support vector machines, naive Bayes models, maximum entropy models, and feature selection methods. Key machine learning approaches covered are support vector machines, naive Bayes classifiers, maximum entropy models, and the use of kernels and feature extraction for text classification tasks.
https://telecombcn-dl.github.io/2017-dlai/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks or Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles of deep learning from both an algorithmic and computational perspectives.
The document discusses various machine learning techniques including k-nearest neighbors, which classifies new data based on its similarity to existing examples; Naive Bayes classifiers, which use Bayes' theorem to classify items based on the presence or absence of features; and decision trees, which classify items by sorting them based on the results of tests on their features and dividing them into branches. Reinforcement learning is also covered, where an agent learns through trial-and-error interactions with an environment by receiving rewards or penalties for actions.
Islamic University Pattern Recognition & Neural Network 2019 Rakibul Hasan Pranto
The document discusses various topics related to pattern recognition including:
1. Pattern recognition is the automated recognition of patterns and regularities in data through techniques like machine learning. It has applications in areas like optical character recognition, diagnosis systems, and security.
2. There are two main approaches to pattern recognition - sub-symbolic and symbolic. Sub-symbolic uses connectionist models like neural networks while symbolic uses formal structures like strings and automata to represent patterns.
3. A pattern recognition system consists of steps like data acquisition, pre-processing, feature extraction, model learning, classification, and post-processing to classify patterns. Bayesian decision making and Bayes' theorem are statistical techniques used in classification.
Yulia Honcharenko "Application of metric learning for logo recognition"Fwdays
Typical approaches of solving classification problems require the collection of a dataset for each new class and retraining of the model. Metric learning allows you to train a model once and then easily add new classes with 5-10 reference images.
So we’ll talk about metric learning based on YouScan experience: task, data, different losses and approaches, metrics we used, pitfalls and peculiarities, things that worked and didn’t.
The document discusses different clustering algorithms, including k-means and EM clustering. K-means aims to partition items into k clusters such that each item belongs to the cluster with the nearest mean. It works iteratively to assign items to centroids and recompute centroids until the clusters no longer change. EM clustering generalizes k-means by computing probabilities of cluster membership based on probability distributions, with the goal of maximizing the overall probability of items given the clusters. Both algorithms are used to group similar items in applications like market segmentation.
Welcome to the May 2025 edition of WIPAC Monthly celebrating the 14th anniversary of the WIPAC Group and WIPAC monthly.
In this edition along with the usual news from around the industry we have three great articles for your contemplation
Firstly from Michael Dooley we have a feature article about ammonia ion selective electrodes and their online applications
Secondly we have an article from myself which highlights the increasing amount of wastewater monitoring and asks "what is the overall" strategy or are we installing monitoring for the sake of monitoring
Lastly we have an article on data as a service for resilient utility operations and how it can be used effectively.
Efficient Algorithms for Isogeny Computation on Hyperelliptic Curves: Their A...IJCNCJournal
We present efficient algorithms for computing isogenies between hyperelliptic curves, leveraging higher genus curves to enhance cryptographic protocols in the post-quantum context. Our algorithms reduce the computational complexity of isogeny computations from O(g4) to O(g3) operations for genus 2 curves, achieving significant efficiency gains over traditional elliptic curve methods. Detailed pseudocode and comprehensive complexity analyses demonstrate these improvements both theoretically and empirically. Additionally, we provide a thorough security analysis, including proofs of resistance to quantum attacks such as Shor's and Grover's algorithms. Our findings establish hyperelliptic isogeny-based cryptography as a promising candidate for secure and efficient post-quantum cryptographic systems.
PRIZ Academy - Functional Modeling In Action with PRIZ.pdfPRIZ Guru
This PRIZ Academy deck walks you step-by-step through Functional Modeling in Action, showing how Subject-Action-Object (SAO) analysis pinpoints critical functions, ranks harmful interactions, and guides fast, focused improvements. You’ll see:
Core SAO concepts and scoring logic
A wafer-breakage case study that turns theory into practice
A live PRIZ Platform demo that builds the model in minutes
Ideal for engineers, QA managers, and innovation leads who need clearer system insight and faster root-cause fixes. Dive in, map functions, and start improving what really matters.
Jacob Murphy Australia - Excels In Optimizing Software ApplicationsJacob Murphy Australia
In the world of technology, Jacob Murphy Australia stands out as a Junior Software Engineer with a passion for innovation. Holding a Bachelor of Science in Computer Science from Columbia University, Jacob's forte lies in software engineering and object-oriented programming. As a Freelance Software Engineer, he excels in optimizing software applications to deliver exceptional user experiences and operational efficiency. Jacob thrives in collaborative environments, actively engaging in design and code reviews to ensure top-notch solutions. With a diverse skill set encompassing Java, C++, Python, and Agile methodologies, Jacob is poised to be a valuable asset to any software development team.
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)ijflsjournal087
Call for Papers..!!!
6th International Conference on Big Data, Machine Learning and IoT (BMLI 2025)
June 21 ~ 22, 2025, Sydney, Australia
Webpage URL : https://inwes2025.org/bmli/index
Here's where you can reach us : [email protected] (or) [email protected]
Paper Submission URL : https://inwes2025.org/submission/index.php
Interfacing PMW3901 Optical Flow Sensor with ESP32CircuitDigest
Learn how to connect a PMW3901 Optical Flow Sensor with an ESP32 to measure surface motion and movement without GPS! This project explains how to set up the sensor using SPI communication, helping create advanced robotics like autonomous drones and smart robots.
YJIT can make Ruby code run faster, but this is a balancing act, because the JIT compiler itself must consume both memory and CPU cycles to compile and optimize your code while it is running. Furthermore, in large-scale production environments such as those of GitHub, Shopify and Stripe, we end up in a situation where YJIT is compiling the same code over and over again on a very large number of servers, which seems very inefficient.
In this presentation, we will go over the design of ZJIT, a next generation Ruby JIT which aims to save and reuse compiled code between executions. We hope that this will help us eliminate duplicated work while also allowing the compiler to spend more time optimizing code so that we can get better performance.
1. 1
Pattern Recognition
Pattern recognition is:
1. The name of the journal of the Pattern Recognition
Society.
2. A research area in which patterns in data are
found, recognized, discovered, …whatever.
3. A catchall phrase that includes
• classification
• clustering
• data mining
• ….
2. 2
Two Schools of Thought
1. Statistical Pattern Recognition
The data is reduced to vectors of numbers
and statistical techniques are used for
the tasks to be performed.
2. Structural Pattern Recognition
The data is converted to a discrete structure
(such as a grammar or a graph) and the
techniques are related to computer science
subjects (such as parsing and graph matching).
3. 3
In this course
1. How should objects to be classified be
represented?
2. What algorithms can be used for recognition
(or matching)?
3. How should learning (training) be done?
4. 4
Classification in Statistical PR
• A class is a set of objects having some important
properties in common
• A feature extractor is a program that inputs the
data (image) and extracts features that can be
used in classification.
• A classifier is a program that inputs the feature
vector and assigns it to one of a set of designated
classes or to the “reject” class.
With what kinds of classes do you work?
5. 5
Feature Vector Representation
X=[x1, x2, … , xn],
each xj a real number
xj may be an object
measurement
xj may be count of
object parts
Example: object rep.
[#holes, #strokes,
moments, …]
7. 7
Some Terminology
Classes: set of m known categories of objects
(a) might have a known description for each
(b) might have a set of samples for each
Reject Class:
a generic class for objects not in any of
the designated known classes
Classifier:
Assigns object to a class based on features
8. 8
Discriminant functions
Functions f(x, K)
perform some
computation on
feature vector x
Knowledge K
from training or
programming is
used
Final stage
determines class
9. 9
Classification using nearest class
mean
Compute the
Euclidean distance
between feature vector
X and the mean of
each class.
Choose closest class,
if close enough (reject
otherwise)
10. 10
Nearest mean might yield poor
results with complex structure
Class 2 has two
modes; where is
its mean?
But if modes are
detected, two
subclass mean
vectors can be
used
12. 12
Nearest Neighbor Classification
• Keep all the training samples in some efficient
look-up structure.
• Find the nearest neighbor of the feature vector
to be classified and assign the class of the neighbor.
• Can be extended to K nearest neighbors.
13. 13
Receiver Operating Curve ROC
Plots correct
detection rate
versus false
alarm rate
Generally, false
alarms go up
with attempts to
detect higher
percentages of
known objects
16. 16
Classifiers often used in CV
• Decision Tree Classifiers
• Artificial Neural Net Classifiers
• Bayesian Classifiers and Bayesian Networks
(Graphical Models)
• Support Vector Machines
18. 18
Decision Tree Characteristics
1. Training
How do you construct one from training data?
Entropy-based Methods
2. Strengths
Easy to Understand
3. Weaknesses
Overtraining
19. 19
Entropy-Based Automatic
Decision Tree Construction
Node 1
What feature
should be used?
What values?
Training Set S
x1=(f11,f12,…f1m)
x2=(f21,f22, f2m)
.
.
xn=(fn1,f22, f2m)
Quinlan suggested information gain in his ID3 system
and later the gain ratio, both based on entropy.
20. 20
Entropy
Given a set of training vectors S, if there are c classes,
Entropy(S) = -pi log (pi)
Where pi is the proportion of category i examples in S.
i=1
c
2
If all examples belong to the same category, the entropy
is 0.
If the examples are equally mixed (1/c examples of each
class), the entropy is a maximum at 1.0.
e.g. for c=2, -.5 log .5 - .5 log .5 = -.5(-1) -.5(-1) = 1
2 2
21. 21
Information Gain
The information gain of an attribute A is the expected
reduction in entropy caused by partitioning on this attribute.
Gain(S,A) = Entropy(S) - ----- Entropy(Sv)
v Values(A)
|Sv|
|S|
where Sv is the subset of S for which attribute A has
value v.
Choose the attribute A that gives the maximum
information gain.
22. 22
Information Gain (cont)
Attribute A
v1 vk
v2
Set S
Set S
repeat
recursively
Information gain has the disadvantage that it prefers
attributes with large number of values that split the
data into small, pure subsets.
S={sS | value(A)=v1}
23. 23
Gain Ratio
Gain ratio is an alternative metric from Quinlan’s 1986
paper and used in the popular C4.5 package (free!).
GainRatio(S,A) = ------------------
Gain(S,a)
SplitInfo(S,A)
SplitInfo(S,A) = - ----- log ------
|Si|
|S|
|Si|
|S|
where Si is the subset of S in which attribute A has its ith value.
2
i=1
ni
SplitInfo measures the amount of information provided
by an attribute that is not specific to the category.
24. 24
Information Content
Note:
A related method of decision tree construction using
a measure called Information Content is given in the
text, with full numeric example of its use.
25. 25
Artificial Neural Nets
Artificial Neural Nets (ANNs) are networks of
artificial neuron nodes, each of which computes
a simple function.
An ANN has an input layer, an output layer, and
“hidden” layers of nodes.
.
.
.
.
.
.
Inputs
Outputs
27. 27
Neural Net Learning
That’s beyond the scope of this text; only
simple feed-forward learning is covered.
The most common method is called back propagation.
We’ve been using a free package called NevProp.
What do you use?
28. 28
Support Vector Machines (SVM)
Support vector machines are learning algorithms
that try to find a hyperplane that separates
the differently classified data the most.
They are based on two key ideas:
• Maximum margin hyperplanes
• A kernel ‘trick’.
31. 31
The kernel trick
The SVM algorithm implicitly maps the original
data to a feature space of possibly infinite dimension
in which data (which is not separable in the
original space) becomes separable in the feature space.
0 0
0 0
0
1
1 1
Original space Rk
0
0
0
0
0
1
1
1
Feature space Rn
1
1
Kernel
trick
32. 32
Our Current Application
• Sal Ruiz is using support vector machines in his
work on 3D object recognition.
• He is training classifiers on data representing deformations
of a 3D model of a class of objects.
• The classifiers are starting to learn what kinds of
surface patches are related to key parts of the model
(ie. A snowman’s face)