Skip to content

Latest commit

 

History

History
72 lines (63 loc) · 6.26 KB

File metadata and controls

72 lines (63 loc) · 6.26 KB

SVR - Support Vector Regression

Notes:
  • Please remember the below list of words in your mind before getting into the Session.
    1. hyper-plane.
    2. Margins.
    3. Support Vectors.
    4. Kernal.
    5. SVM = Support Vector Machine.
    6. SVR = Support Vector Regression

1. Introduction

  • Hi Welcome to 4'th part of Regression models - Support Vector Regression[SVR] which is sub class of a familiar ML-Model Support Vector Machine[SVM].
  • I recommend you to look in to SVM-Support Vector Machine intution to get more understanding on SVR.
  • The Support-Vector algorithm is a nonlinear generalization of the Generalized Portrait algorithm, it is firmly grounded in the framework of statistical learning theory, or VC theory.
  • In a nutshell, VC theory characterizes properties of learning machines which enable them to generalize well to unseen data.
  • Our goal is to find a function f(x) that has at most ε deviation from the actually obtained targets yi for all the training data, and at the same time is as flat as possible.
  • It is considered a non parametric technique because it relies on kernel functions.
  • In other words, we do not care about errors as long as they are less than ε, but will not accept any deviation larger than this
  • The key difference between SVM and SVR is that SVM works with non-linear(Discrete) data in the other hand SVR works with Linear(continuous) data.
  • When it comes to SVR the following terms used very frequently
    1. Kernal:
      1. The function which is used to map lower dimentional data into a higher dimentional data.
      2. Generally it could be linear, polynomial, sigmoid and rbf.
      3. This is known as Kernel-trick.
    2. Support Vectors:
      1. Actual data-points which are close to line of boundary.
      2. The distance between data-point and boundary line
    3. Hyper Plane:
      1. In SVM, Hyper Plane is the important factor which separates the data classes, on we can say it the line between data classes.
      2. In SVR, Hyper Plane will play the same role as SVM, but as we discussed before the key change is, here it is a line between linear(continuous) data.
      3. We can consider this as a best fit line predicted by our kernal.
    4. Boundary Plane or Decision Boundary(margin of tolerance):
      1. The name Boundary itself defines its usage, yes it is going to define the upper/lower boundaries from the Hyper-Plane.
      2. It act as a margin between Hyper-Plane and data points. One interesting factor is The support vectors can be on the Boundary line or outside boundary line.
      3. Here the distance between Boundary-Plane and Hyper-Plane are considered as epsilon(ε) distance. So the lines that we draw are at ‘’ and ‘ ’ distance from Hyper Plane.
      4. Below image shows the above points discussed.

Ref: http://www.saedsayad.com/support_vector_machine_reg.htm

  • Why SVR and how it differs from Linear Regression ?
    • In Normal Linear Regression model's like (SLR, MLR and Poly), we tried to mimize the error rate and we derived our best fit line.
    • In SVR, we tried to fit the error within the boundary threshold, so we can get more accuracy and better prediction in our model.

2. Support Vector Regression in Python:

3. Additional Use-cases:

  1. Create a Machine Learning Model to predict Car Price and compare your results with Multi-Linear and Polynomial results.[Car Price Prediction]
  2. Create a Machine Learning Model to predict Air Quality.[Air Quality]
  3. Create a Machine Learning model to predict ERP-estimated relative performance of a computer and compare it with Multi-Linear and Polynomial results.[ERP Prediction]
  4. Create a Machine Learning model to predict Wine Quality using SVR.[Wine Quality]
Reference URL:

Use Cases reference
SVM - Support Vector Machine
How Does SVR Works
how does support vector regression work intuitively

See Also