Papers
Topics
Authors
Recent
Search
2000 character limit reached

Algorithms for Sparse Support Vector Machines

Published 14 Oct 2021 in stat.ME | (2110.07691v1)

Abstract: Many problems in classification involve huge numbers of irrelevant features. Model selection reveals the crucial features, reduces the dimensionality of feature space, and improves model interpretation. In the support vector machine literature, model selection is achieved by $\ell_1$ penalties. These convex relaxations seriously bias parameter estimates toward 0 and tend to admit too many irrelevant features. The current paper presents an alternative that replaces penalties by sparse-set constraints. Penalties still appear, but serve a different purpose. The proximal distance principle takes a loss function $L(\boldsymbol{\beta})$ and adds the penalty $\frac{\rho}{2}\mathrm{dist}(\boldsymbol{\beta}, S_k)2$ capturing the squared Euclidean distance of the parameter vector $\boldsymbol{\beta}$ to the sparsity set $S_k$ where at most $k$ components of $\boldsymbol{\beta}$ are nonzero. If $\boldsymbol{\beta}\rho$ represents the minimum of the objective $f\rho(\boldsymbol{\beta})=L(\boldsymbol{\beta})+\frac{\rho}{2}\mathrm{dist}(\boldsymbol{\beta}, S_k)2$, then $\boldsymbol{\beta}_\rho$ tends to the constrained minimum of $L(\boldsymbol{\beta})$ over $S_k$ as $\rho$ tends to $\infty$. We derive two closely related algorithms to carry out this strategy. Our simulated and real examples vividly demonstrate how the algorithms achieve much better sparsity without loss of classification power.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.