ICASSP 2006 - May 15-19, 2006 - Toulouse, France

Technical Program

Paper Detail

Paper:MLSP-L1.6
Session:Learning Theory I
Time:Wednesday, May 17, 11:40 - 12:00
Presentation: Lecture
Topic: Machine Learning for Signal Processing: Learning Theory and Modeling
Title: An Optimal Basis for Feature Extraction with Support Vector Machine Classification using the Radius-Margin Bound
Authors: Jeff Fortuna, David Capson, McMaster University, Canada
Abstract: A method is presented for deriving an optimal basis for features classified with a support vector machine. The method is based on minimizing the leave-one-out error which is approximated by the radius--margin bound. A gradient descent method provides a learning rule for the basis in an outer loop of an iteration. The inner loop performs support vector machine training and provides support vector coefficients on which the gradient descent depends. In this way, the derivation of a basis for feature extraction and the support vector machine are jointly optimized. The efficacy of the method is illustrated with examples from multi-dimensional synthetic data sets.



IEEESignal Processing Society

©2018 Conference Management Services, Inc. -||- email: webmaster@icassp2006.org -||- Last updated Friday, August 17, 2012