The Hybrid Dynamic Prototype Construction and Parameter Optimization with Genetic Algorithm for Support Vector Machine
Abstract
The optimized hybrid artificial intelligence model is a potential tool to deal with construction engineering and management problems. Support vector machine (SVM) has achieved excellent performance in a wide variety of applications. Nevertheless, how to effectively reduce the training complexity for SVM is still a serious challenge. In this paper, a novel order-independent approach for instance selection, called the dynamic condensed nearest neighbor (DCNN) rule, is proposed to adaptively construct prototypes in the training dataset and to reduce the redundant or noisy instances in a classification process for the SVM. Furthermore, a hybrid model based on the genetic algorithm (GA) is proposed to simultaneously optimize the prototype construction and the SVM kernel parameters setting to enhance the classification accuracy. Several UCI benchmark datasets are considered to compare the proposed hybrid GA-DCNN-SVM approach with the previously published GA-based method. The experimental results illustrate that the proposed hybrid model outperforms the existing method and effectively improves the classification performance for the SVM.
References
V. N. Vapnik, Statistical learning theory, 1st ed. New York: Wiley, 1998.
D. Anguita, A. Ghio, L. Oneto and S. Ridella, "In-Sample and out-of-sample model selection and error estimation for
support vector machines," IEEE Transactions on Neural Networks and Learning Systems, vol. 23, pp.1390-1406, 2012.
C. L. Huang and C. J. Wang, "A GA-based feature selection and parameters optimization for support vector machines,"
Expert systems with applications, vol. 31, pp. 231-240, 2006.
P. E. Hart, "The condensed nearest neighbor rule," IEEE Transactions on Information Theory, vol. 14, pp. 515-516,
F. Angiulli and A. Astorino, "Scaling up support vector machines using nearest neighbor condensation," IEEE
Transactions on Neural Networks, vol. 21, pp. 351-357, 2010.
F. Angiulli, "Fast nearest neighbor condensation for Large data sets classification," IEEE Trans. Knowledge and data
engineering, vol. 19, pp. 1450-1464, 2007.
A. Abroudi and F. Farokhi, "Prototype selection for training artificial neural networks based on Fast Condensed Nearest
Neighbor rule," IEEE Conference on Open Systems (ICOS), pp. 1-4, 2012.
X. Zhao, D. Li, Bo Yang, H. Chen, X. Yang, C. L. Yu and S. Y. Liu, "A two-stage feature selection method with its
application," Computers & Electrical Engineering, vol. 47, pp. 114-125, Oct. 2015.
W. Shu and H. Shen, "Incremental feature selection based on rough set in dynamic incomplete data," Pattern
Recognition, vol. 47, pp. 3890-3906, Dec. 2014.
J. R. Rico-Juan and J. M. Iñesta, "Adaptive training set reduction for nearest neighbor classification," Neurocomputing,
vol. 138, pp. 316-324, Aug. 2014.
J. L. Chen and C. S. Yang, "Optimizing the proportion of prototypes generation in nearest neighbor classification,"
International Conference on Machine Learning and Cybernetics (ICMLC), vol. 4, pp. 1695-1699, 2013.
Y. Miao, X. Tao, Y. Sun, Y. Li and J. Lu, "Risk-based adaptive metric learning for nearest neighbor classification,"
Neurocomputing, vol. 156, pp. 33-41, May 2015.
J. H. Holland, Adaption in natural and artificial systems, Cambridge, MIT Press Cambridge, MA, USA, 1992.
X. Sun, J. Wang, F.H. Mary and L. Kong, "Scale invariant texture classification via sparse representation,"
Neurocomputing, vol. 122, pp. 338-348, Dec. 2013.
S. Cateni, V. Colla and M. Vannucci, "A hybrid feature selection method for classification purposes," European
Modelling Symposium (EMS), pp. 39-44, 2014.
C. C. Chang and C. J. Lin, "Training nu-support vector regression: theory and algorithms," Neural Computation, vol. 14,
pp. 1959-1977, 2002.
Published
How to Cite
Issue
Section
License
Copyright Notice
Submission of a manuscript implies: that the work described has not been published before that it is not under consideration for publication elsewhere; that if and when the manuscript is accepted for publication. Authors can retain copyright in their articles with no restrictions. Also, author can post the final, peer-reviewed manuscript version (postprint) to any repository or website.
Since Jan. 01, 2019, IJETI will publish new articles with Creative Commons Attribution Non-Commercial License, under Creative Commons Attribution Non-Commercial 4.0 International (CC BY-NC 4.0) License.
The Creative Commons Attribution Non-Commercial (CC-BY-NC) License permits use, distribution and reproduction in any medium, provided the original work is properly cited and is not used for commercial purposes.