Hong, X.
ORCID: https://orcid.org/0000-0002-6832-2298, Mitchell, R. and Di Fatta, G.
(2019)
Simplex basis function based sparse least squares support vector regression.
Neurocomputing, 330.
pp. 394-402.
ISSN 0925-2312
doi: 10.1016/j.neucom.2018.11.025
Abstract/Summary
In this paper, a novel sparse least squares support vector regression algorithm, referred to as LSSVR-SBF, is introduced which uses a new low rank kernel based on simplex basis function, which has a set of nonlinear parameters. It is shown that the proposed model can be represented as a sparse linear regression model based on simplex basis functions. We propose a fast algorithm for least squares support vector regression solution at the cost of O(N) by avoiding direct kernel matrix inversion. An iterative estimation algorithm has been proposed to optimize the nonlinear parameters associated with the simplex basis functions with the aim of minimizing model mean square errors using the gradient descent algorithm. The proposed fast least square solution and the gradient descent algorithm are alternatively applied. Finally it is shown that the model has a dual representation as a piecewise linear model with respect to the system input. Numerical experiments are carried out to demonstrate the effectiveness of the proposed approaches.
Altmetric Badge
| Item Type | Article |
| URI | https://reading-clone.eprints-hosting.org/id/eprint/80789 |
| Identification Number/DOI | 10.1016/j.neucom.2018.11.025 |
| Refereed | Yes |
| Divisions | Science > School of Mathematical, Physical and Computational Sciences > Department of Computer Science |
| Publisher | Elsevier |
| Download/View statistics | View download statistics for this item |
Downloads
Downloads per month over past year
University Staff: Request a correction | Centaur Editors: Update this record
Download
Download