非稀疏性是最小二乘支持向量机(Least squares support vector machine,LS-SVM)的主要不足,因此稀疏化是LS-SVM研究的重要内容.在目前LS-SVM稀疏化研究中,多数算法采用的是基于迭代选择的稀疏化策略,但是时间复杂度和稀疏化效果还不够理想.为了进一步改进LS-SVM稀疏化方法的性能,文中提出了一种基于全局代表点选择的快速LS-SVM稀疏化算法(Global-representation-based sparse least squares support vector machine,GRS-LSSVM).在综合考虑数据局部密度和全局离散度的基础上,给出了数据全局代表性指标来评估每个数据的全局代表性.利用该指标,在全部数据中,一次性地选择出其中最具有全局代表性的数据并构成稀疏化后的支持向量集,然后在此基础上求解决策超平面,是该算法的基本思路.该算法对LS-SVM的非迭代稀疏化研究进行了有益的探索.通过与传统的迭代稀疏化方法进行比较,实验表明GRS-LSSVM具有稀疏度高、稳定性好、计算复杂度低的优点.
For lack of sparseness on least squares support vector machine(LS-SVM), the study on sparsity of LS-SVM is an important topic. Currently, most of the sparse LS-SVM methods are based on the iteration selection strategy.Consequently, they do not perform well in computation complexity and sparsity. To improve the performance of sparse LS-SVM method, a fast method, global-representation-based sparse least squares support vector machine(GRS-LSSVM),is proposed based on the selection of global representative points in this paper. To evaluate datum s representation, an index is given based on local density and global discrete degree. In the algorithm, firstly, the top global representative data are selected from all data in one step using the index to construct the support vector set of sparse LS-SVM, and then the set is used to compute the decision hyperplane of sparse LS-SVM. This algorithm explores the non-iteration on sparse LS-SVM. Experimental results show that the proposed method has higher sparseness degree, more stability, and lower computational complexity than the traditional iteration algorithms.