This book focuses on Least Squares Support Vector Machines (LS-SVMs) which are reformulations to standard SVMs. LS-SVMs are closely related to regularization networks and Gaussian processes but additionally emphasize and exploit primal-dual interpretations from optimization theory. The authors explain the natural links between LS-SVM classifiers and kernel Fisher discriminant analysis. Bayesian inference of LS-SVM models is discussed, together with methods for imposing spareness and employing robust statistics. The framework is further extended towards unsupervised learning by considering PCA analysis and its kernel version as a one-class modelling problem. This leads to new primal-dual support vector machine formulations for kernel PCA and kernel CCA analysis. Furthermore, LS-SVM formulations are given for recurrent networks and control. In general, support vector machines may pose heavy computational challenges for large data sets. For this purpose, a method of fixed size LS-SVM is proposed where the estimation is done in the primal space in relation to a Nystrom sampling with active selection of support vectors. The methods are illustrated with several examples.... polynomial and RBF kernels for LS-SVM classifiers, standard SVM classifiers with linear and RBF kernel; the decision tree ... The Matlab SVM toolbox http:// theoval.sys.uea.ac.uk/~gcc/svm/tool- box with SMO solver was used to train andanbsp;...

Title | : | Least Squares Support Vector Machines |

Author | : | Johan A. K. Suykens, Tony Van Gestel, Jos De Brabanter |

Publisher | : | World Scientific - 2002-01-01 |

Continue