Hinge Loss

From our SVM model, we know that hinge loss=max(0,1yf(x))loss = max(0, 1- y*f(x)).

Looking at the graph for SVM in Fig 4, we can see that for yf(x)1y*f(x) \geq 1, hinge loss is ‘0’. However, when yf(x)<1y*f(x) < 1, then hinge loss increases massively. As yf(x)y*f(x) increases with every misclassified point (very wrong points in Fig 5), the upper bound of hinge loss 1yf(x){1- y*f(x)} also increases exponentially.

Hence, the points that are farther away from the decision margins have a greater loss value, thus penalising those points.

Conclusion: This is just a basic understanding of what loss functions are and how hinge loss works. I will be posting other articles with greater understanding of ‘Hinge loss’ shortly

Last updated