Variable Selection for Support Vector Machines in Moderately High Dimensions.

TitleVariable Selection for Support Vector Machines in Moderately High Dimensions.
Publication TypeJournal Article
Year of Publication2016
AuthorsZhang, Xiang, Yichao Wu, Lan Wang, and Runze Li
JournalJ R Stat Soc Series B Stat Methodol
Volume78
Issue1
Pagination53-76
Date Published2016 Jan
ISSN1369-7412
Abstract

The support vector machine (SVM) is a powerful binary classification tool with high accuracy and great flexibility. It has achieved great success, but its performance can be seriously impaired if many redundant covariates are included. Some efforts have been devoted to studying variable selection for SVMs, but asymptotic properties, such as variable selection consistency, are largely unknown when the number of predictors diverges to infinity. In this work, we establish a unified theory for a general class of nonconvex penalized SVMs. We first prove that in ultra-high dimensions, there exists one local minimizer to the objective function of nonconvex penalized SVMs possessing the desired oracle property. We further address the problem of nonunique local minimizers by showing that the local linear approximation algorithm is guaranteed to converge to the oracle estimator even in the ultra-high dimensional setting if an appropriate initial estimator is available. This condition on initial estimator is verified to be automatically valid as long as the dimensions are moderately high. Numerical examples provide supportive evidence.

DOI10.1111/rssb.12100
Alternate JournalJ R Stat Soc Series B Stat Methodol
Original PublicationVariable selection for support vector machines in moderately high dimensions.
PubMed ID26778916
PubMed Central IDPMC4709852
Grant ListP50 DA010075 / DA / NIDA NIH HHS / United States
P50 DA039838 / DA / NIDA NIH HHS / United States
R01 CA149569 / CA / NCI NIH HHS / United States
P01 CA142538 / CA / NCI NIH HHS / United States
P50 DA036107 / DA / NIDA NIH HHS / United States