# Learn How to Use Support Vector Machines (SVM) for Data Science

Sunil Ray 01 Jul, 2024

## Introduction

Mastering machine learning algorithms isn’t a myth at all. Most beginners start by learning regression. It is simple to learn and use, but does that solve our purpose? Of course not! Because there is a lot more in ML beyond logistic regression and regression problems! For instance, have you heard of support vector regression and support vector machines algorithm or SVM?

Think of machine learning algorithms as an armory packed with axes, swords, blades, bows, daggers, etc. You have various tools, but you ought to learn to use them at the right time. As an analogy, think of ‘Regression’ as a sword capable of slicing and dicing data efficiently but incapable of dealing with highly complex data. That is where ‘Support Vector Machines’ acts like a sharp knife – it works on smaller datasets, but on complex ones, it can be much stronger and more powerful in building machine learning models.

Learning Objectives

• Understand support vector machine algorithm (SVM), a popular machine learning algorithm or classification.
• Learn to implement SVM models in R and Python.
• Know the pros and cons of Support Vector Machines (SVM) and their different applications in machine learning (artificial intelligence).

## What is a Support Vector Machine (SVM)?

“Support Vector Machine” (SVM) is a supervised learning machine learning algorithm that can be used for both classification or regression challenges. However, it is mostly used in classification problems, such as text classification. In the SVM algorithm, we plot each data item as a point in n-dimensional space (where n is the number of features you have), with the value of each feature being the value of a particular coordinate. Then, we perform classification by finding the optimal hyper-plane that differentiates the two classes very well (look at the below snapshot).

Support Vectors are simply the coordinates of individual observation, and a hyper-plane is a form of SVM visualization. The SVM classifier is a frontier that best segregates the two classes (hyper-plane/line).

## How does a Support Vector Machine / SVM Work?

Above, we got accustomed to the process of segregating the two classes with a hyper-plane. Now the burning question is, “How can we identify the right hyper-plane?”. Don’t worry; it’s not as hard as you think! Let’s understand:

#### Identify the Right Hyper-plane

• Here, we have three hyper-planes (A, B, and C). Now, identify the right hyper-plane to classify stars and circles.
• You need to remember a thumb rule to identify the right hyper-plane: “Select the hyper-plane which segregates the two classes better.” In this scenario, hyper-plane “B” has excellently performed this job.

#### Another Example of Identifying the Right Hyper-plane

• Here, we have three hyper-planes (A, B, and C), and all segregate the classes well. Now, How can we identify the right hyper-plane?
• Here, maximizing the distances between the nearest data point (either class) and the hyper-plane will help us to decide the right hyper-plane. This distance is called a Margin. Let’s look at the below snapshot:
• Above, you can see that the margin for hyper-plane C is high as compared to both A and B. Hence, we name the right hyper-plane as C. Another lightning reason for selecting the hyper-plane with a higher margin is robustness. If we select a hyper-plane having a low margin, then there is a high chance of misclassification.

#### Another Example of Identifing the right hyper-plane

• Hint: Use the rules as discussed in the previous section to identify the right hyper-plane.
• Some of you may have selected hyper-plane B as it has a higher margin compared to A. But, here is the catch, SVM selects the hyper-plane which classifies the classes accurately prior to maximizing the margin. Here, hyper-plane B has a classification error, and A has classified all correctly. Therefore, the right hyper-plane is A.

#### Can we classify two classes?

• Below, I am unable to segregate the two classes using a straight line, as one of the stars lies in the territory of the other (circle) class as an outlier.
• As I have already mentioned, one star at the other end is like an outlier for the star class. The SVM algorithm has a feature to ignore outliers and find the hyper-plane that has the maximum margin. Hence, we can say SVM classification is robust to outliers.

#### Find the Hyper-plane to Segregate to Classes

• In the scenario below, we can’t have a linear hyper-plane between the two classes, so how does SVM classify these two classes? Till now, we have only looked at the linear hyper-plane.
• SVM can solve this problem. Easily! Specifically, it solves this problem by introducing additional features. Here, we will add a new feature, ( z = x^2 + y^2 ). Now, let’s plot the data points on the x and z axes:

In the above plot, points to consider are:

• All values for z would always be positive because z is the squared sum of both x and y
• In the original plot, red circles appear close to the origin of the x and y axes, leading to a lower value of z. The star is relatively away from the original results due to the higher value of z.

In the SVM classifier, having a linear hyper-plane between these two classes is easy. But, another burning question that arises is if we should we need to add this feature manually to have a hyper-plane. No, the SVM  algorithm has a technique called the kernel trick. The SVM kernel transforms low-dimensional input space to a higher dimensional space, making non-separable problems separable, useful for non-linear data separation by applying complex data transformations based on labels.

When we look at the hyper-plane in the original input space, it looks like a circle:

Now, let’s look at the methods to apply the SVM classifier algorithm in a data science challenge.

You can also learn about the working of a Support Vector Machine in data mining video format from this Machine Learning certification course.

## Hyperplane and Support Vectors in SVM

#### Hyperplane

In an SVM, a hyperplane is a decision boundary that separates different classes of data points. For instance, in a two-dimensional space, the hyperplane is a line; in a three-dimensional space, it is a plane. The goal of the SVM is to find the optimal hyperplane that maximizes the margin between the classes. The margin is defined as the distance between the hyperplane and the nearest data points from either class.

#### Support Vectors

Support vectors are the data points that are closest to the hyperplane. These points are critical because they determine the position and orientation of the hyperplane. If you remove a support vector, it can change the hyperplane’s position.

## Types of Support Vector Machine

#### Linear SVM

Linear SVM is used when the data is linearly separable, which means that the classes can be separated with a straight line (in 2D) or a flat plane (in 3D). The SVM algorithm finds the hyperplane that best divides the data into classes.

#### Non-Linear SVM

Non-Linear SVM is used when the data is not linearly separable. In such cases, SVM employs kernel functions to transform the data into a higher-dimensional space where a linear separation is possible. The algorithm then finds the optimal hyperplane in this new space.

Kernels are functions that take low-dimensional input space and transform it into a higher-dimensional space. SVM can create complex decision boundaries by using kernel functions. Here are some popular kernel functions:

#### Linear Kernel

Used when the data is linearly separable.

#### Polynomial Kernel

Where c is a constant, and d is the degree of the polynomial. This kernel is useful for classifying data with polynomial relationships.

#### Radial Basis Function (RBF) Kernel / Gaussian Kernel

Where γ is a parameter that defines the influence of a single training example. This is one of the most popular kernels for non-linear data.

#### Sigmoid Kernel

Where α and care kernel parameters. It behaves like a neural network’s activation function.

## How to Implement SVM in Python and R?

In Python, scikit-learn is a widely used library for implementing machine learning algorithms. SVM algorithm is also available in the scikit-learn library, and we follow the same structure for using it(Import library, object creation, fitting model, and prediction).

Now, let us have a look at a real-life problem statement and dataset to understand how to apply SVM for classification.

### Problem Statement

Dream Housing Finance company deals in all home loans. They have a presence across all urban, semi-urban, and rural areas. A customer first applies for a home loan; after that, the company validates the customer’s eligibility for a loan.

The company wants to automate the loan eligibility process (real-time) based on customer details provided while filling out an online application form. These details are Gender, Marital Status, Education, Number of Dependents, Income, Loan Amount, Credit History, and others. To automate this process, they have given a problem of identifying the customers’ segments that are eligible for loan amounts so that they can specifically target these customers. Here they have provided a partial data set.

Use the coding window below to predict the loan eligibility on the test set(new data). Try changing the hyperparameters for the linear SVM to improve the accuracy.

### Support Vector Machine (SVM) Code in R

The e1071 package in R is used to create Support Vector Machine in data mining with ease. It has helper functions as well as code for the Naive Bayes Classifier. The creation of a support vector machine algorithm in R and Python follows similar approaches; let’s take a look now at the following code:

``````#Import Library
require(e1071) #Contains the SVM
# there are various options associated with SVM training; like changing kernel, gamma and C value.

# create model
model <- svm(Target~Predictor1+Predictor2+Predictor3,data=Train,kernel='linear',gamma=0.2,cost=100)

#Predict Output
preds <- predict(model,Test)
table(preds)``````

### How to Tune the Parameters of SVM?

Tuning the parameters’ values for machine learning algorithms effectively improves model performance. Therefore, let’s look at the list of parameters available with SVM.

``sklearn.svm.SVC(C=1.0, kernel='rbf', degree=3, gamma=0.0, coef0=0.0, shrinking=True, probability=False,tol=0.001, cache_size=200, class_weight=None, verbose=False, max_iter=-1, random_state=None)``

I am going to discuss some important parameters having a higher impact on model performance, “kernel,” “gamma,” and “C.”

kernel: We have already discussed it. Here, we have various options available with kernel like “linear,” “rbf”, ”poly”, and others (default value is “rbf”).  Here “rbf”(radial basis function) and “poly”(polynomial kernel) are useful for non-linear hyper-plane. It’s called nonlinear svm. Let’s look at the example where we’ve used linear kernel on two features of the iris data set to classify their class.

### Support Vector Machine (SVM) Code in Python

#### Have a Linear SVM kernel

``````import numpy as np
import matplotlib.pyplot as plt
from sklearn import svm, datasets

# import some data to play with
X = iris.data[:, :2] # we only take the first two features. We could
# avoid this ugly slicing by using a two-dim dataset
y = iris.target

# we create an instance of SVM and fit out data. We do not scale our
# data since we want to plot the support vectors
C = 1.0 # SVM regularization parameter
svc = svm.SVC(kernel='linear', C=1,gamma=0).fit(X, y)

# create a mesh to plot in
x_min, x_max = X[:, 0].min() - 1, X[:, 0].max() + 1
y_min, y_max = X[:, 1].min() - 1, X[:, 1].max() + 1
h = (x_max / x_min)/100
xx, yy = np.meshgrid(np.arange(x_min, x_max, h),
np.arange(y_min, y_max, h))

plt.subplot(1, 1, 1)
Z = svc.predict(np.c_[xx.ravel(), yy.ravel()])
Z = Z.reshape(xx.shape)
plt.contourf(xx, yy, Z, cmap=plt.cm.Paired, alpha=0.8)``````

#### Use SVM rbf kernel

``````plt.scatter(X[:, 0], X[:, 1], c=y, cmap=plt.cm.Paired)
plt.xlabel('Sepal length')
plt.ylabel('Sepal width')
plt.xlim(xx.min(), xx.max())
plt.title('SVC with linear kernel')``````
``plt.show()``

#### Use SVM rbf kernel

Change the kernel function type to rbf in the below line and look at the impact.

``svc = svm.SVC(kernel='rbf', C=1,gamma=0).fit(X, y)``

I would suggest you go for a linear SVM kernel if you have a large number of features (>1000) because it is more likely that the data is linearly separable in high dimensional space. Also, you can use RBF but do not forget to cross-validate for its parameters to avoid over-fitting.

gamma: Kernel coefficient for ‘rbf’, ‘poly’, and ‘sigmoid.’ The higher value of gamma will try to fit them exactly as per the training data set, i.e., generalization error and cause over-fitting problem.

#### Let’s differentiate if we have gamma different gamma values like 0, 10, or 100.

svc = svm.SVC(kernel=’rbf’, C=1,gamma=0).fit(X, y)

C: Penalty parameter C of the error term. It also controls the trade-off between smooth decision boundaries and classifying the training points correctly.

We should always look at the cross-validation score to effectively combine these parameters and avoid over-fitting.

In R, SVMs algorithm can be tuned in a similar fashion as they are in Python. Mentioned below are the respective parameters for the e1071 package:

• The kernel parameter can be tuned to take “Linear”, ”Poly”, ”rbf”, etc.
• The gamma value can be tuned by setting the “Gamma” parameter.
• The C value in Python is tuned by the “Cost” parameter in R.

## Pros and Cons of SVM

Pros:

• It works really well with a clear margin of separation.
• It is effective in high-dimensional spaces.
• It is effective in cases where the number of dimensions is greater than the number of samples.
• It uses a subset of the training set in the decision function (called support vectors), so it is also memory efficient.

Cons:

• It doesn’t perform well when we have a large data set because the required training time is higher.
• It also doesn’t perform very well when the data set has more noise, i.e., target classes are overlapping.
• The SVM algorithm doesn’t directly provide probability estimates; it calculates them using an expensive five-fold cross-validation. The related SVC method of the Python scikit-learn library includes this feature.

## SVM Practice Problem

Find the right additional feature to have a hyper-plane for segregating the classes in the below snapshot:

## Conclusion

In this article, we looked at the machine learning algorithm, Support Vector Machine, in detail. We discussed the concept of its working, the process of its implementation in python and R, and the tricks to make the model more efficient by tuning its parameters. Towards the end, we also pointed out the pros and cons of the algorithm. I suggest you try solving the problem above to practice your SVM algorithm skills and also try to analyze the power of this model by tuning the parameters.

Key Takeaways

• Support Vector Machine in data mining strongly and powerfully builds machine learning models with small data sets.
• You can effectively improve your model’s performance by tuning the SVM hyperparameters in Python.
• The algorithm works best when there are more dimensions than samples, and I do not recommend using it for noisy, large, or complex data sets.

Q1. What is support vector machines with examples?

A. Support vector machines (SVM) are supervised learning models used for classification and regression tasks. For instance, they can classify emails as spam or not spam. Additionally, they can be used to identify handwritten digits in image recognition.

Q2. What is the principle of SVM?

A. The principle of SVM involves finding the hyperplane that best separates different classes of data. Essentially, it maximizes the margin between the closest points of the classes, thereby ensuring robust classification.

Q3. What is the function of SVM?

A. The function of SVM is to classify data by finding the optimal hyperplane that separates different classes. Consequently, it works well for both linear and non-linear classification problems by transforming data using kernel functions.

Q4. Why is it called a support vector machine?

A. It is called a support vector machine because it relies on support vectors, which are the data points closest to the hyperplane. These support vectors are critical as they define the position and orientation of the hyperplane, thus influencing the model’s accuracy.

Sunil Ray 01 Jul, 2024

I am a Business Analytics and Intelligence professional with deep experience in the Indian Insurance industry. I have worked for various multi-national Insurance companies in last 7 years.

nishant 07 Oct, 2015

hi, gr8 articles..explaining the nuances of SVM...hope u can reproduce the same with R.....it would be gr8 help to all R junkies like me

ASHISH 07 Oct, 2015

NEW VARIABLE (Z) = SQRT(X) + SQRT (Y)

Mahmood A. Sheikh 07 Oct, 2015

Kernel

Sanjay 07 Oct, 2015

Nicely Explained . The hyperplane to separate the classes for the above problem can be imagined as 3-D Parabola. z=ax^2 + by^2 + c

FrankSauvage 12 Oct, 2015

Thanks a lot for this great hands-on article!

Harsha 08 Nov, 2015

Really impressive content. Simple and effective. It could be more efficient if you can describe each of the parameters and practical application where you faced non-trivial problem examples.

Aman Srivastava 26 Nov, 2015

kernel

How does the python code look like if we are using LSSVM instead of SVM?

Janpreet Singh 04 Mar, 2016

Polynomial kernel function?! for exzmple : Z= A(x^2) + B(y^2) + Cx + Dy + E

Krishna Kalaparti 18 Apr, 2016

Hi Sunil. Great Article. However, there's an issue in the code you've provided. When i compiled the code, i got the following error: Name error: name 'h' is not defined. I've faced this error at line 16, which is: "xx, yy = np.meshgrid(np.arange(x_min, x_min, h), ...). Could you look into it and let me know how to fix it?

Shikha 28 May, 2016

great explanation :) I think new variable Z should be x^2 + y.

VEERAMANI NATARAJAN 03 Jun, 2016

Nice Articlel

Carlos 14 Jun, 2016

The solution is analogue to scenario-5 if you replace y by y-k

K.Krithiga Lakshmi 15 Jun, 2016

Your SVM explanation and kernel definition is very simple, and easy to understand. Kudos for that effort.

pfcohen 19 Jun, 2016

Most intuitive explanation of multidimensional svm I have seen. Thank you!

yc 27 Jun, 2016

what is 'h' in the code of SVM . xx, yy = np.meshgrid(np.arange(x_min, x_max, h), np.arange(y_min, y_max, h))

Iresh 14 Jul, 2016

z = (x^2 - y) z > 0, red circles

LI JENG HUANG 05 Aug, 2016

very neat explaination to SVC. For the proposed problem, my answers are: (1) z = a* x^2 + b y + c, a parabola. (2) z = a (x-0)^2 + b (y- y0)^2 - R^2, a circle or an ellipse enclosing red stars.

Hari 18 Aug, 2016

Great article.. I think the below formula would give a new variable that help to separate the points in hyper plane z = y - |x|

THANKS FOR EASY EXPLANATION

Raghu 31 Aug, 2016

Useful article for Machine learners.. Why can't you discuss about effect of kernel functions.

Yamani 22 Sep, 2016

The explanation is really impressive. Can you also provide some information about how to determine the theoretical limits for the parameter's optimal accuracy.

harshel jain 23 Sep, 2016

how can we use SVM for regression? can someone please explain..

Diana 04 Oct, 2016

That was a really good explanation! thanks a lot. I read many explanations about SVM but this one help me to understand the basics which I really needed it.

Manjunath GS 27 Oct, 2016

Diptesh 28 Oct, 2016

This is very useful for understanding easily.

Dan 14 Nov, 2016

just substitude x with |x|

Min 23 Nov, 2016

Same goes with Diana. This really help me a lot to figure out things from basic. I hope you would also share any computation example using R provided with simple dataset, so that anyone can practice with their own after referring to your article. I have a question, if i have time-series dataset containing mixed linear and nonlinear data, (for example oxygen saturation data ; SaO2), by using svm to do classification for diseased vs health subjects, do i have to separate those data into linear and non-linear fisrt, or can svm just performed the analysis without considering the differences between the linearity of those data? Thanks a lot!

anu 29 Nov, 2016

z

Renny Varghese 03 Dec, 2016

Could you please explain how SVM works for multiple classes? How would it work for 9 classes? I used a function called multisvm here: http://www.mathworks.com/matlabcentral/fileexchange/39352-multi-class-svm but I'm not sure how it's working behind the scenes. Everything I've read online is rather confusing.

lubna 06 Dec, 2016

NEW VARIABLE (Z) = SQRT(X) + SQRT (Y)

Haftom A. 07 Dec, 2016

Thank you so much!! That is really good explanation! I read many explanations about SVM but this one help me to understand the basics which I really needed it. keep it up!!

Frank 06 Jan, 2017

Thanks for the great article. There are even cool shirts for anyone who became SVM fan ;) http://www.redbubble.com/de/people/perceptron/works/24728522-support-vector-machines?grid_pos=2&p=t-shirt&style=mens

bilashi 10 Jan, 2017

great explanation!! Thanks for posting it.

arun 19 Jan, 2017

I think this is |X|

It is very nicely written and understandable. Thanks a lot...

Walter 06 Feb, 2017

z=ax^2 + by^2

nice explanations with scenarios and margin values

lishanth 01 Mar, 2017

wow!!! excellent explanation.. only now i understood the concepts clearly thanks a lot..

anwar 01 Mar, 2017

(Z) = SQRT(X) + SQRT (Y)

Kresla Matty 20 Mar, 2017

thanks, and well done for the good article

Jonathan benitez 16 Apr, 2017

Aishwarya Jangam 20 Apr, 2017

Great Explanation..Thanks..

Hams 17 May, 2017

simple and refreshed the core concepts in just 5 mins! kudos Mr.Sunil

Shashi 17 May, 2017

Best starters material for SVM, really appreciate the simple and comprehensive writing style. Expecting more such articles from you

Ravindar 20 May, 2017

Z= square(x)

Narasimha 25 May, 2017

Hey Sunil, Nice job of explaining it concisely and intuitively! Easy to follow and covers many aspects in a short space. Thanks!

John Doe 30 May, 2017

Very well written - concise, clear, well-organized. Thank you.

Camille 11 Jun, 2017

Bonjour, je souhaite appliquer cet algorithme à mes données puis afficher le résultat. Je ne maîtrise pas très bien python, j'ai donc recopié le code suivant: def pro_svc(X): # we create an instance of SVM and fit out data. We do not scale our # data since we want to plot the support vectors C = 1.0 # SVM regularization parameter svc = svm.SVC(kernel='linear', C=1,gamma=0).fit(X, y) # create a mesh to plot in x_min, x_max = X[:, 0].min() - 1, X[:, 0].max() + 1 y_min, y_max = X[:, 1].min() - 1, X[:, 1].max() + 1 h = (x_max / x_min)/100 xx, yy = np.meshgrid(np.arange(x_min, x_max, h), np.arange(y_min, y_max, h)) #repartition d'un nobre de points entre un start et un end plt.subplot(1, 1, 1) Z = svc.predict(np.c_[xx.ravel(), yy.ravel()]) Z = Z.reshape(xx.shape) plt.contourf(xx, yy, Z, cmap=plt.cm.Paired, alpha=0.8) plt.scatter(X[:, 0], X[:, 1], c=y, cmap=plt.cm.Paired) plt.xlabel('Sepal length') plt.ylabel('Sepal width') plt.xlim(xx.min(), xx.max()) plt.title('SVC with linear kernel') plt.show() Lorque j'appelle ce code sur me données (matrices numpy 60 lignes, 2 colonnes) l'erreur suivante s'affiche : ValueError: zero-size array to reduction operation maximum which has no identity Que dois_je faire ? Merci d'avance pour votre aide

Excellent explanation..Can you please also tell what are the parameter values one should start with - like C, gamma ..Also, again a very basic question.. Can we say that lesser the % of support vectors (count of SVs/total records) better my model/richer my data is- assuming the datasize to be the same.. Waiting for more on parameter tuning..Really appreciate the knowledge shared..

Kirana 15 Jun, 2017

Hi could you please explain why SVM perform well on small dataset?

Chris 20 Jun, 2017

Another nice kernel for the problem stated in the article is the radial basis kernel.

实用指南-在python中使用Scikit-learn进行数据预处理 - 数据分析网 22 Jun, 2017

[…] 资源：阅读这篇文章来理解SVM support vector machines。 […]

chiru 23 Jun, 2017

wow excellent

Zhen Zhang 26 Jun, 2017

very appreciating for explaining

Andrey 27 Jun, 2017

Nice tutorial. The new feature to separate data would be something like z = y - x^2 as most dots following the parabola will have lower z than stars.

Very intuitive explanation. Thank you! Good to add SVM for Regression of Continuous variables.

neha 11 Jul, 2017

this is so simple method that anyone can get easily thnx for that but also explain the 4 senario of svm.

Nirav Pingle 20 Jul, 2017

Great article for understanding of SVM: But, When and Why do we use the SVM algorithm can anyone make that help me understand because until this thing is clear there may not be use of this article. Thanks in advance.

Mostafa 02 Aug, 2017

It is one of best explanation of machine learning technique that i have seen! and new variable: i think Z=|x| and new Axis are Z and Y

venkat 03 Aug, 2017

higher degree polynomial will separate the points in the problem,

Tirthankar 08 Aug, 2017

I guess the required feature is z = x^2 / y^2 For the red points, z will be close to 1 but for the blue points z values will be significantly more than 1

murtaza ali 09 Aug, 2017

amazing article no doubt! It makes me clear all the concept and deep points regarding SVM. many thanks.

katherine 19 Aug, 2017

The best explanation ever! Thank you!

Rahul 20 Aug, 2017

z = x^2+y^2

Applied text classification on Email Spam Filtering [part 1] – Sarah Mestiri 01 Sep, 2017

[…] [1] Naive Bayes and Text Classification. [2]Naive Bayes by Example. [3] Andrew Ng explanation of Naive Bayes video 1 and video 2 [4] Please explain SVM like I am 5 years old. [5] Understanding Support Vector Machines from examples. […]

roshan 07 Sep, 2017

new variable = ABS(Y)

Robert 13 Sep, 2017

Man, I was looking for definition of SVM for my diploma, but I got interested in explanation part of this article. Keep up good work!

Aman Goel 15 Sep, 2017

we can use 'poly' kernel with degree=2

Nethra Kulkarni 21 Sep, 2017

Hi.. Very well written, great article !:). Thanks so much share knowledge on SVM.

S Sen Sharma 23 Sep, 2017

z=y-x^2

Dalon 02 Oct, 2017

Wonderful, easy to understand explanation.

Eka A 11 Oct, 2017

Thanks a lot for your explanations, they were really helpful and easy to understand

Prachi 13 Oct, 2017

Really helpful article. Can you please elaborate on time complexity for implementation of SVMs? Are they polynomial time or non-polynomial time executing algorithms?

Harsha Vardhan 14 Oct, 2017

Good explanation,but it would be better if some one explains the equations and math formulation involved in SVM. y=w.x+b

Kevin Mekulu 19 Oct, 2017

It would be a parabola z = a*x^2 + b*y^2 + c*x + d*y + e

shefali 03 Nov, 2017

valuable explanation!!

Lena 06 Nov, 2017

Thanks for your great overview and explanation about SVM's!! Only one question: In the very beginning you define what a Support Vector is. Shouldn't you emphasize, that it is only a subset of the data points? Otherwise you could wonder if all data points are Support Vectors. Cheers, Lena

vami 15 Nov, 2017

Thirupathi Thangavel 28 Dec, 2017

ValueError: The gamma value of 0.0 is invalid. Use 'auto' to set gamma to a value of 1 / n_features.

Sakshi 29 Dec, 2017

Thanks for the great article. I have a question...for data like tweets of twitter....svm is helpful or it would be better to choose another option? which will be helpful?.

Fran 05 Jan, 2018

I would do |x| or 2 linear SVMs -ax+b, till 0 and then cx+d afterward. Concatenating linear SVM is way more powerful than expected. A very nice explanation, I totally love your work.

Fernando Ortega Gallego 12 Jan, 2018

I think you should change the title "How to implement SVM in Python and R?" because you don't actually implement SVM, you only use implementations of SVM through some libraries like scikit-learn for Python.

Shivam Misra 12 Jan, 2018

|X|

panimalar 18 Jan, 2018

thank u sir ,it is easy to understand

Pallavi 22 Jan, 2018

Hi Sunil, Great explanation of SVM in very simple and understandable format. Can you please the equivalent R code for generating the mesh plots? I think the hyper-plane looks like a parabola. So the new feature is z = y^2+y+x^2+x+c. Kindly let me know if this is correct. Thanks & Regards, Pallavi

Sundar Jubilant 24 Jan, 2018

Hi Sunil, Wonderful article. Can you also cover what are gamma and cost values and how to select them?

John 09 Feb, 2018

z = x^2 + y

Nide 13 Feb, 2018

Hi there. Great article. I have a question. You wrote: "if you have large number of features (>1000) because it is more likely that the data is linearly separable in high dimensional space." Could you explain why this is probably the case? Thanks

Edin 13 Feb, 2018

Hello, great article there. Just one quesiton. You wrote: "if you have large number of features (>1000) because it is more likely that the data is linearly separable in high dimensional space" Could you explain why this is probably the case? Thanks

Pavan Kumar 07 Mar, 2018

It may be z=x^2+y

Jose 10 Mar, 2018

y=x^2

anoop 21 Mar, 2018

z=ax^2 + by^2 + c

quandapro 28 Mar, 2018

Nice. new variable is z = abs(x). Then replace x coordinates with z coordinates

Shravani 29 Mar, 2018

Can you please tell me how an image can be represented as a single point in support vector space..... and how histograms used in support vector classification... I know it is not related to this article... but I am asking in general...

Athul 31 Mar, 2018

z = |x|

Boris Epshtein 28 Apr, 2018

My God, your English is so broken!

Deyire Yusuf Umar 02 May, 2018

PARABOLA

Jason 02 May, 2018

I think the boundaryf between two type of snapshot could be a curve (of a part of circle). So I prefer kernel Z=sqrt(X^2+(Y-c)^2)

ILA 08 May, 2018

Thanks a lot. I like how you define a problem and then solve it. It makes things clear.

Rizwan Ali 24 May, 2018

hi, can anyone help me for forecasting of stock exchange by SVM in rstudio, i hv a project for my mphil course work.

Prachi 26 May, 2018

z=x-y^2

Imran ullah 02 Jan, 2022

The content is good but you can not use the gamma parameter if you want to separate your data linearly.

K Mala 08 Jul, 2022

explanations are easy to undersdand

Rakesh Nandkumar Lipare 15 Oct, 2022

Polynomial kernel, Nice and simple explanation with required knowledge

Bijumon 24 Nov, 2022

Really good article.

Deputy 05 Jul, 2023

This is great article. And the explanation are clear

Praneeth 06 Jul, 2023

New Variable (z) = x**2 - y + c (some constant)

Praneeth 06 Jul, 2023

z = x**2 - y + c (some constant)