Abstract
The support vector machine (SVM) is a popular classification method which enjoys good performance in many real applications. The SVM can be viewed as a penalized minimization problem in which the objective function is the expectation of hinge loss function with respect to the standard non-smooth empirical measure corresponding to the true underlying measure. We further extend this viewpoint and propose a smoothed SVM by substituting a kernel density estimator for the measure in the expectation calculation. The resulting method is called density convoluted support vector machine (DCSVM). We argue that the DCSVM is particularly more interesting than the standard SVM in the context of high-dimensional classification. We systematically study the rate of convergence of the elastic-net penalized DCSVM under general random design setting. We further develop novel efficient algorithm for computing elastic-net penalized DCSVM. Simulation studies and ten benchmark datasets are used to demonstrate the superior classification performance of elastic-net DCSVM over other competitors, and it is demonstrated in these numerical studies that the computation of DCSVM can be more than 100 times faster than that of the SVM.
Original language | English (US) |
---|---|
Pages (from-to) | 2523-2536 |
Number of pages | 14 |
Journal | IEEE Transactions on Information Theory |
Volume | 69 |
Issue number | 4 |
DOIs | |
State | Published - Apr 1 2023 |
Bibliographical note
Funding Information:The work of Hui Zou was supported in part by NSF under Grant 1915842 and Grant 2015120.
Publisher Copyright:
IEEE
Keywords
- Classification
- DCSVM
- kernel density smoother
- support vector machines
- ultra-high dimension