Feature selection using stochastic approximation with Barzilai and Borwein non-monotone gains

Aksakalli, Vural and D. Yenice, Zeren and Malekipirbazari, Milad and Kargar, Kamyar (2021) Feature selection using stochastic approximation with Barzilai and Borwein non-monotone gains. Computers and Operations Research, 132: 105334. ISSN 0305-0548

Full text not available from this repository.

Abstract

With recent emergence of machine learning problems with massive number of features, feature selection (FS) has become an ever-increasingly important tool to mitigate the effects of the so-called curse of dimensionality. FS aims to eliminate redundant and irrelevant features for models that are faster to train, easier to understand, and less prone to overfitting. This study presents a wrapper FS method based on Simultaneous Perturbation Stochastic Approximation (SPSA) with Barzilai and Borwein (BB) non-monotone gains within a pseudo-gradient descent framework wherein performance is measured via cross-validation. We illustrate that SPSA with BB gains (SPSA-BB) provides dramatic improvements in terms of the number of iterations for convergence with minimal degradation in cross-validated error performance over the current state-of-the art approach with monotone gains (SPSA-MON). In addition, SPSA-BB requires only one internal parameter and therefore it eliminates the need for careful fine-tuning of numerous other internal parameters as in SPSA-MON or comparable meta-heuristic FS methods such as genetic algorithms (GA). Our particular implementation includes gradient averaging as well as gain smoothing for better convergence properties. We present computational experiments on various public datasets with Nearest Neighbors and Naive Bayes classifiers as wrappers. We present comparisons of SPSA-BB against full set of features, SPSA-MON, as well as seven popular meta-heuristics based FS algorithms including GA and particle swarm optimization. Our results indicate that SPSA-BB converges to a good feature set in about 50 iterations on the average regardless of the number of features (whether a dozen or more than 1000 features) and its performance is quite competitive. SPSA-BB can be considered extremely fast for a wrapper method and therefore it stands as a high-performing new feature selection method that is also computationally feasible in practice.

Item Type:
Journal Article
Journal or Publication Title:
Computers and Operations Research
Uncontrolled Keywords:
/dk/atira/pure/subjectarea/asjc/1700/1700
Subjects:
?? barzilai and borwein methodexplainable artificial intelligencefeature selectiongenetic algorithmgradient descentstochastic approximationgeneral computer sciencemodelling and simulationmanagement science and operations researchcomputer science(all) ??
ID Code:
219532
Deposited By:
Deposited On:
10 May 2024 08:25
Refereed?:
Yes
Published?:
Published
Last Modified:
24 Sep 2024 12:05