Svm
Svm
MACHINE
Prof. Subodh Kumar Mohanty
The idea of support vectors and its
importance
Introduction
• The support vector machine is currently considered to be the best
off-the-shelf learning algorithm and has been applied successfully in various
domains.
• Support vector machines were originally designed for binary classification.
• Then, it is extended to solve multi-class and regression problems.
• But, it is widely used in classification objectives.
• The objective of the support vector machine algorithm is to find a hyperplane
in an N-dimensional space (N — the number of features) that distinctly
classifies the data points.
• To separate the two classes of data points, there are many possible
hyperplanes that could be chosen. Our objective is to find a plane that
has the maximum margin, i.e the maximum distance between data
points of both classes.
Possible
hyperplanes
• Maximizing the margin distance provides some reinforcement so that
future data points can be classified with more confidence.
Possible
hyperplanes
• Maximizing the margin distance provides some reinforcement so that
future data points can be classified with more confidence.
• The goal is to choose a hyperplane with the greatest possible margin
between the hyperplane and any point within the training set, giving a
greater chance of new data being classified correctly.
• Hyperplanes are decision boundaries that help classify the data points.
• Data points falling on either side of the hyperplane can be attributed to
different classes.
• Also, the dimension of the hyperplane depends upon the number of
features.
• It becomes difficult to imagine when the number of features exceeds 3.
• Support vectors are data points that are closer to the hyperplane and
influence the position and orientation of the hyperplane.
• Using these support vectors, we maximize the margin of the classifier.
• Deleting the support vectors will change the position of the hyperplane.
• These are the points that help us build our SVM (that works for a
new/test data).
Cons
• Isn’t suited to larger datasets as the training time with SVMs can be high
• Less effective on noisier datasets with overlapping classes
Applications
• SVM is used for text classification tasks such as category assignment,
detecting spam and sentiment analysis.
• It is also commonly used for image recognition challenges,
performing particularly well in aspect-based recognition and
color-based classification.
• SVM also plays a vital role in many areas of handwritten digit
recognition, such as postal automation services.
Derivation of Support Vector Equation
Comparison with logistic regression
Training set:
m examples
Sigmoid function