0% found this document useful (0 votes)
20 views15 pages

Unit6 - 6 Support Vector Machine

The document discusses various classification techniques including decision trees, Bayesian classification, artificial neural networks, and support vector machines (SVM). SVM is highlighted as a powerful supervised learning algorithm that creates a hyperplane to separate classes, with advantages in high-dimensional spaces but limitations in handling large datasets and noise. It also covers evaluation metrics for classification accuracy and issues like overfitting and underfitting.

Uploaded by

nikhillamsal1
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views15 pages

Unit6 - 6 Support Vector Machine

The document discusses various classification techniques including decision trees, Bayesian classification, artificial neural networks, and support vector machines (SVM). SVM is highlighted as a powerful supervised learning algorithm that creates a hyperplane to separate classes, with advantages in high-dimensional spaces but limitations in handling large datasets and noise. It also covers evaluation metrics for classification accuracy and issues like overfitting and underfitting.

Uploaded by

nikhillamsal1
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 15

6.

1 Definition (Classification, Prediction), Learning and


testing of classification,
6.2 Classification by decision tree induction, ID3 as attribute
selection algorithm
6.3 Bayesian classification, Laplace smoothing
6.4 ANN: Classification by backpropagation
6.5 Rule based classifier (Decision tree to rules, rule
coverage and accuracy, efficient of rule simplification)
6.6 Support vector machine,
6.7 Evaluating accuracy (precision, recall, f-measure)
• Issues in classification, Overfitting and underfitting

• K- fold cross validation, Comparing two classifier


(McNemar’s test)
6.6 Support vector machine
◼ Support Vector Machine or SVM is one of the most
popular Supervised Learning algorithms, which is used for
Classification as well as Regression problems.
◼ The goal of the SVM algorithm is to create the best line or
decision boundary that can segregate n-dimensional
space into classes so that we can easily put the new data
point in the correct category in the future.
◼ This best decision boundary is called a hyperplane.
◼ SVM chooses the extreme points/vectors that help in
creating the hyperplane.
◼ These extreme cases are called support vectors, and
hence algorithm is termed a Support Vector Machine.

9
◼ A Support Vector Machine (SVM) performs classification
by finding the hyperplane (classifier) that maximizes the
margin between the two classes subject to the constraint
that all the training tuples should be correctly classified.
◼ The hyperplane is defined by using the data points that are
closest to the boundary.
◼ These points are called support vectors and the decision
boundary itself is called support vector machine.
◼ The main advantage of the SVM classifier is that it
minimizes the training set error and the test set error.

10
◼ To obtain an SVM classifier with the best
generalization performance, appropriate training
is required. The most commonly used and popular
algorithm for training SVM is the sequential
minimal optimization (SMO) algorithm. The main
advantage of the SMO algorithm is that it works
analytically on a fixed size working set by
decomposing the large training data set. So, that it
can work fine even for large data sets and it also
gives superb performances in almost all kinds of
training data sets.

11
12
13
Advantages of SVM

◼ 1. SVM works relatively well when there is a clear


margin of separation between classes.
◼ 2 SVM is more effective in high-dimensional
spaces.
◼ 3. SVM is effective in cases where the number of
dimensions is greater than the number of
samples.
◼ 4. SVM is relatively memory efficient.
◼ 5. SVM can be used for both regression and
classification problems.
◼ 6. SVM can work well with image data as well.
14
Disadvantages of SVM
◼ 1. SVM algorithm is not suitable for large data sets.
◼ 2 SVM does not perform very well when the data set has
more noise i.e. target classes are overlapping.
◼ 3. In cases where the number of features for each data
point exceeds the number of training data samples, the
SVM will underperform.
◼ 4. As the support vector classifier works by putting data
points, above and below the classifying hyperplane there
is no probabilistic explanation for the classification.
◼ 5. It is difficult to understand and interpret the SVM model
compared to decision tree as SVM is more complex

15

You might also like