TY - JOUR
T1 - Condensed vector machines
T2 - Learning fast machine for large data
AU - Nguyen, Dung Duc
AU - Matsumoto, Kazunori
AU - Takishima, Yasuhiro
AU - Hashimoto, Kazuo
N1 - Funding Information:
Manuscript received February 19, 2009; revised August 26, 2010; accepted August 27, 2010. Date of publication October 18, 2010; date of current version November 30, 2010. This work was supported in part by the Vietnam National Foundation for Science and Technology Development under NAFOSTED Grant 102.02.16.09.
PY - 2010/12
Y1 - 2010/12
N2 - Scalability is one of the main challenges for kernel-based methods and support vector machines (SVMs). The quadratic demand in memory for storing kernel matrices makes it impossible for training on million-size data. Sophisticated decomposition algorithms have been proposed to efficiently train SVMs using only important examples, which ideally are the final support vectors (SVs). However, the ability of the decomposition method is limited to large-scale applications where the number of SVs is still too large for a computer's capacity. From another perspective, the large number of SVs slows down SVMs in the testing phase, making it impractical for many applications. In this paper, we introduce the integration of a vector combination scheme to simplify the SVM solution into an incremental working set selection for SVM training. The main objective of the integration is to maintain a minimal number of final SVs, bringing a minimum resource demand and faster training time. Consequently, the learning machines are more compact and run faster thanks to the small number of vectors included in their solution. Experimental results on large benchmark datasets shows that the proposed condensed SVMs achieve both training and testing efficiency while maintaining a generalization ability equivalent to that of normal SVMs.
AB - Scalability is one of the main challenges for kernel-based methods and support vector machines (SVMs). The quadratic demand in memory for storing kernel matrices makes it impossible for training on million-size data. Sophisticated decomposition algorithms have been proposed to efficiently train SVMs using only important examples, which ideally are the final support vectors (SVs). However, the ability of the decomposition method is limited to large-scale applications where the number of SVs is still too large for a computer's capacity. From another perspective, the large number of SVs slows down SVMs in the testing phase, making it impractical for many applications. In this paper, we introduce the integration of a vector combination scheme to simplify the SVM solution into an incremental working set selection for SVM training. The main objective of the integration is to maintain a minimal number of final SVs, bringing a minimum resource demand and faster training time. Consequently, the learning machines are more compact and run faster thanks to the small number of vectors included in their solution. Experimental results on large benchmark datasets shows that the proposed condensed SVMs achieve both training and testing efficiency while maintaining a generalization ability equivalent to that of normal SVMs.
KW - Decomposition algorithm
KW - Kernel method
KW - optimization
KW - reduced set method
KW - support vector machine
UR - http://www.scopus.com/inward/record.url?scp=78650072699&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=78650072699&partnerID=8YFLogxK
U2 - 10.1109/TNN.2010.2079947
DO - 10.1109/TNN.2010.2079947
M3 - Article
C2 - 20959266
AN - SCOPUS:78650072699
SN - 1045-9227
VL - 21
SP - 1903
EP - 1914
JO - IEEE Transactions on Neural Networks
JF - IEEE Transactions on Neural Networks
IS - 12
M1 - 5605254
ER -