Feature elimination in kernel machines in moderately high dimensions

Sayan Dasgupta, Yair Goldberg, Michael R. Kosorok

Research output: Contribution to journalArticlepeer-review

Abstract

We develop an approach for feature elimination in statistical learning with kernel machines, based on recursive elimination of features. We present theoretical properties of this method and show that it is uniformly consistent in finding the correct feature space under certain generalized assumptions. We present a few case studies to show that the assumptions are met in most practical situations and present simulation results to demonstrate performance of the proposed approach.

Original languageEnglish
Pages (from-to)497-526
Number of pages30
JournalAnnals of Statistics
Volume47
Issue number1
DOIs
StatePublished - Feb 2019

Bibliographical note

Publisher Copyright:
© Institute of Mathematical Statistics, 2019

Keywords

  • Kernel machines
  • Recursive feature elimination
  • Support vector machines
  • Variable selection

ASJC Scopus subject areas

  • Statistics and Probability
  • Statistics, Probability and Uncertainty

Fingerprint

Dive into the research topics of 'Feature elimination in kernel machines in moderately high dimensions'. Together they form a unique fingerprint.

Cite this