What advantage does Guassian kernel have than any other kernels, such as linear kernel, polynomial kernel and so on?

Guassian kernel is so important in SVM as we know. The parameter gamma is designed for this kind of kernel. My question is what makes Guassian kernel so unique? What advantage does it have over other kernels?

Topic kernel svm machine-learning

Category Data Science


Mathematically speaking, one can show that an RBF is simply a low-band pass filter. To put it simply, RBF induces a smooth solution. Sometimes that is an important characteristic that matters. However, it isn't always important or useful. For example, it's somewhat well known that a smoothing prior does not outperform a linear/polynomial kernel for problems like Text Classification.

There's also other reasons though why people tend to stick with an RBF kernel instead of trying to design their own or use a polynomial. One main one is that an RBF is non-parametric -- meaning that it's complexity grows with data. While if one uses a polynomial kernel, then you have a parametric model which has a finite and fix size. There essentially comes a point when your polynomial kernel becomes saturated no matter how much data you throw at it. Hence, it's generally seen as a good benchmark to start with an RBF unless you have reason to believe otherwise.

There's other properties that make it mathematically useful, like it's invariant, stationary contain less parameters than polynomial kernels.

TL;DR generally works well.


The advantange is that it is related to euclidean distance, and if your problem benefits from this type of feature, e.g. you have a highly non-linear problem where values nearby some vector point tend to be of a different class, it will be able to find relationships that other kernels would not, and create highly complex decision boundaries.

About

Geeks Mental is a community that publishes articles and tutorials about Web, Android, Data Science, new techniques and Linux security.