| ... | @@ -42,7 +42,22 @@ $`y_{i} = w . x_i + b >= 0`$ for positive points |
... | @@ -42,7 +42,22 @@ $`y_{i} = w . x_i + b >= 0`$ for positive points |
|
|
|
|
|
|
|
$`y_{i} = w . x_i + b <= 0`$ for negative points
|
|
$`y_{i} = w . x_i + b <= 0`$ for negative points
|
|
|
|
|
|
|
|
</div>...
|
|
</div>
|
|
|
|
|
|
|
|
Note that we do not know what w or b is.
|
|
|
|
|
|
|
|
At this stage, we will add a custom rule, stating that the samples should be further away from that zero-line solution, say 1-unit distance:
|
|
|
|
|
|
|
|
<div>
|
|
|
|
|
|
|
|
$`y_{i} = w . x_i + b >= 1`$
|
|
|
|
|
|
|
|
$`y_{i} = w . x_i + b <= -1`$
|
|
|
|
|
|
|
|
</div>
|
|
|
|
|
|
|
|
We are adding this rule to limit of degrees of freedom, as we can propose many solutions to the classification problem. This little touch is what we call ‘margin’. The margin is defined as the as the perpendicular distance between the decision boundary and the closest of the data points. When we set an objective such as “maximize the margin”, then we end up with a particular solution. Herein, the solution is found via a subset of points called support vectors.
|
|
|
|
|
|
|
|
|
|
|
|
## Additional Sources
|
|
## Additional Sources
|
|
|
|
|
|
| ... | |
... | |
| ... | | ... | |