Sparse Penalized Forward Selection for Support Vector Classification. Issue 2 (2nd April 2016)
- Record Type:
- Journal Article
- Title:
- Sparse Penalized Forward Selection for Support Vector Classification. Issue 2 (2nd April 2016)
- Main Title:
- Sparse Penalized Forward Selection for Support Vector Classification
- Authors:
- Ghosal, Subhashis
Turnbull, Bradley
Zhang, Hao Helen
Hwang, Wook Yeon - Abstract:
- Abstract : We propose a new binary classification and variable selection technique especially designed for high-dimensional predictors. Among many predictors, typically, only a small fraction of them have significant impact on prediction. In such a situation, more interpretable models with better prediction accuracy can be obtained by variable selection along with classification. By adding an ℓ1 -type penalty to the loss function, common classification methods such as logistic regression or support vector machines (SVM) can perform variable selection. Existing penalized SVM methods all attempt to jointly solve all the parameters involved in the penalization problem altogether. When data dimension is very high, the joint optimization problem is very complex and involves a lot of memory allocation. In this article, we propose a new penalized forward search technique that can reduce high-dimensional optimization problems to one-dimensional optimization by iterating the selection steps. The new algorithm can be regarded as a forward selection version of the penalized SVM and its variants. The advantage of optimizing in one dimension is that the location of the optimum solution can be obtained with intelligent search by exploiting convexity and a piecewise linear or quadratic structure of the criterion function. In each step, the predictor that is most able to predict the outcome is chosen in the model. The search is then repeatedly used in an iterative fashion until convergenceAbstract : We propose a new binary classification and variable selection technique especially designed for high-dimensional predictors. Among many predictors, typically, only a small fraction of them have significant impact on prediction. In such a situation, more interpretable models with better prediction accuracy can be obtained by variable selection along with classification. By adding an ℓ1 -type penalty to the loss function, common classification methods such as logistic regression or support vector machines (SVM) can perform variable selection. Existing penalized SVM methods all attempt to jointly solve all the parameters involved in the penalization problem altogether. When data dimension is very high, the joint optimization problem is very complex and involves a lot of memory allocation. In this article, we propose a new penalized forward search technique that can reduce high-dimensional optimization problems to one-dimensional optimization by iterating the selection steps. The new algorithm can be regarded as a forward selection version of the penalized SVM and its variants. The advantage of optimizing in one dimension is that the location of the optimum solution can be obtained with intelligent search by exploiting convexity and a piecewise linear or quadratic structure of the criterion function. In each step, the predictor that is most able to predict the outcome is chosen in the model. The search is then repeatedly used in an iterative fashion until convergence occurs. Comparison of our new classification rule with ℓ1 -SVM and other common methods show very promising performance, in that the proposed method leads to much leaner models without compromising misclassification rates, particularly for high-dimensional predictors. … (more)
- Is Part Of:
- Journal of computational and graphical statistics. Volume 25:Issue 2(2016)
- Journal:
- Journal of computational and graphical statistics
- Issue:
- Volume 25:Issue 2(2016)
- Issue Display:
- Volume 25, Issue 2 (2016)
- Year:
- 2016
- Volume:
- 25
- Issue:
- 2
- Issue Sort Value:
- 2016-0025-0002-0000
- Page Start:
- 493
- Page End:
- 514
- Publication Date:
- 2016-04-02
- Subjects:
- High dimension -- Penalization -- Sparsity -- SVM -- Variable selection
Mathematical statistics -- Data processing -- Periodicals
Mathematical statistics -- Graphic methods -- Periodicals
519.50285 - Journal URLs:
- http://pubs.amstat.org/loi/jcgs ↗
http://www.catchword.com/titles/10857117.htm ↗
http://www.tandf.co.uk/journals/titles/10618600.asp ↗
http://www.tandfonline.com/ ↗ - DOI:
- 10.1080/10618600.2015.1023395 ↗
- Languages:
- English
- ISSNs:
- 1061-8600
- Deposit Type:
- Legaldeposit
- View Content:
- Available online (eLD content is only available in our Reading Rooms) ↗
- Physical Locations:
- British Library DSC - 4963.451000
British Library DSC - BLDSS-3PM
British Library HMNTS - ELD Digital store - Ingest File:
- 2114.xml