SVMs can efficiently perform a non-linear classification using what is called the kernel trick, implicitly mapping their inputs into high-dimensional feature spaces.
The original maximum-margin hyperplane algorithm proposed by Vapnik in 1963 constructed a linear classifier. However, in 1992, Bernhard E. Boser, Isabelle M. Guyon and Vladimir N. Vapnik suggested a way to create nonlinear classifiers by applying the kernel trick .
The following links provides resources on applications of SVMs for two class and multi class cladsification
My professor always says SVM the best first choice for any classification task. The reason: SVM is one of the most robust and accurate algorithm among the other classification algorithms.
SVMs can efficiently perform a non-linear classification using what is called the kernel trick, implicitly mapping their inputs into high-dimensional feature spaces.
The original maximum-margin hyperplane algorithm proposed by Vapnik in 1963 constructed a linear classifier. However, in 1992, Bernhard E. Boser, Isabelle M. Guyon and Vladimir N. Vapnik suggested a way to create nonlinear classifiers by applying the kernel trick .
The following links provides resources on applications of SVMs for two class and multi class cladsification
Since SVM is one of the most used techniques, you should try it. The problem is to set parameters. You can try Optimum-Path Forest as well. It is parameterless.
thanks, all and thanks Behrouz for sharing the links. It will be the great help for me .
We can use SVM when a number of features are high compared to a number of data points in the dataset. By using the correct kernel and setting an optimum set of parameters.
SVM is one of the best classifier but not the best. In fact, no one could be the best. It depends upon the problem which classifier would be suitable.
As for as, SVM is concerned, it is a suitable classifier in following cases:
1) When number of features (variables) and number of training data is very large (say millions of features and millions of instances (data)).
2) When sparsity in the problem is very high, i.e., most of the features have zero value.
3) It is the best for document classification problems where sparsity is high and features/instances are also very high.
4) It also performs very well for problems like image classification, genes classsification, drug disambiguation etc. where number of features are high.
It is one of the best, because of following things:
1) It uses Kernel trick
2) It is Optimal margin based classification technique in Machine Learning.
3) Good number of algorithms are proposed which utilizes problem structures and other smaller-smaller things like problem shrinking during optimization etc.