Optimal data division for empowering artificial neural network models employing a modified M-SPXY algorithm

Main Article Content

Wirote Apinantanakon
Khamron Sunat
Joel Alan Kinmond

Abstract

Data splitting is an important step in artificial neural network (ANN) models, which is found in the form of training and testing subsets. In general, a random data splitting method is favored to divide a pool of samples into subsets, without considering the quality of data for the training step of a neural network. The drawback of poor data splitting methods is that they poses ill effects to the performance of the neural network when the data involves complex matrices or multivariate modeling. In order to overcome this drawback, the current paper presents our proposed M-SPXY method. It is based on a modified version of Sample Set Partitioning, which relies on a joint X-y distances (SPXY) method. The proposed method has resulted in better performance, compared to the modified Kennard-Stone (KS) method, using Mahalanobis distances (MDKS). In our experiments, the proposed approach was employed to compare various data splitting methods using data sets from the repository of the University of California in Irvine (UCI), processed through an Extreme Learning Machine (ELM) neural network. Performance was measured in terms of classification accuracy. The results indicate that the classification accuracy of the proposed M-SPXY process is superior to that of the MDKS data splitting method.

Article Details

How to Cite
Apinantanakon, W., Sunat, K., & Kinmond, J. A. (2019). Optimal data division for empowering artificial neural network models employing a modified M-SPXY algorithm. Engineering and Applied Science Research, 46(4), 276–284. Retrieved from https://ph01.tci-thaijo.org/index.php/easr/article/view/184408
Section
ORIGINAL RESEARCH

References

Haykin S. Neural networks: a comprehensive foundation. USA: Prentice Hall PTR; 1998.

Maier HR, Jain A, Dandy GC, Sudheer KP. Methods used for the development of neural networks for the prediction of water resource variables in river systems: Current status and future directions. Environ Model Software. 2010;25(8):891-909.

Baskin I, Palyulin V, Zefirov N. Neural networks in building QSAR models. In: Livingstone D, editor. Artificial neural networks. USA: Humana Press; 2009. p. 133-54.

Stojanović MB, Božić MM, Stanković MM, Stajić ZP. A methodology for training set instance selection using mutual information in time series prediction. Neurocomputing. 2014;141:236-45.

Picard RR, Cook RD. Cross-validation of regression models. J Am Stat Assoc. 1984;79:575-83.

May RJ, Maier HR, Dandy GC. Data splitting for artificial neural networks using SOM-based stratified sampling. Neural Networks. 2010;23(2):283-94.

Wu W, Walczak B, Massart DL, Heuerding S, Erni F, Last IR, et al. Artificial neural networks in classification of NIR spectral data: Design of the training set. Chemometr Intell Lab Syst. 1996;33(1):35-46.

Refaeilzadeh P, Tang L, Liu H. Cross-Validation. In: Liu L, ÖZsu MT, editors. Encyclopedia of database systems. USA: Springer; 2009. p. 532-8.

Pahikkala T, Suominen H, Boberg J, Salakoski T. Efficient hold-out for subset of regressors, In: Kolehmainen M, Toivanen P, Beliczynski B, editors. Adaptive and natural computing algorithms. Berlin: Springer Berlin Heidelberg; 2009. p. 350-9.

Kennard RW, Stone LA. Computer Aided Design of Experiments. Technometrics. 1969;11(1):137-48.

Galvão RKH, Araujo MCU, José GM, Pontes MJC, Silva EC, Saldanha TCB. A method for calibration and validation subset partitioning. Talanta. 2005;67(4):736-40.

Saptoro A, Tadé MO, Vuthaluru H. A modified kennard-stone algorithm for optimal division of data for developing artificial neural network models. Chem Prod Process Model. 2012;7(1):1-14.

Maesschalck RD, Jouan-Rimbaud D, Massart DL. The Mahalanobis distance. Chemometr Intell Lab Syst. 2000;50(1):1-18.

Yang KC, Huang CH, Yang C, Chao PY, Shih PH. Using artificial neural back-propagation network model to detect the outliers in semiconductor manufacturing machines. In: Ali M, Pan JS, Chen SM, Horng MF, editors. Modern Advances in Applied Intelligence. Berlin: Springer International Publishing; 2014. p. 240-9.

Sun Z, Wang J, Nie L, Li L, Cao D, Fan J, Zang H. Calibration transfer of near infrared spectrometers for the assessment of plasma ethanol precipitation process. Chemometr Intell Lab Syst. 2018;181:64-71.

Tian H, Zhang L, Li M, Wang Y, Sheng D, Liu J, et al. Weighted SPXY method for calibration set selection for composition analysis based on near-infrared spectroscopy. Infrared Phys Tech. 2018;95:88-92.

Gueorguieva N, Valova I, Georgiev G. M&MFCM: fuzzy c-means clustering with mahalanobis and minkowski distance metrics. Procedia Comput Sci. 2017;114:224-33.

Wu W, Mallet Y, Walczak B, Penninckx W, Massart DL, Heurding S, et al. Comparison of regularised discriminant analysis, linear discriminant analysis and quadratic discriminant analysis, applied to NIR data. Anal Chim Acta. 1996;329(3):257-65.

Jouan-Rimbaud D, Massart DL, Saby CA, Puel C. Characterisation of the representativity of selected sets of samples in multivariate calibration and pattern recognition. Anal Chim Acta. 1997;350(1-2):149-61.

Jouan-Rimbaud D, Massart DL, Saby CA, Puel C. Determination of the representativity between two multidimensional data sets by a comparison of their structure. Chemometr Intell Lab. 1998;40(2):129-44.

Shenk JS, Westerhaus MO. Population definition, sample selection, and calibration procedures for near infrared reflectance spectroscopy. Crop Sci. 1991;31(2):469-74.

Dean T, Isaksson T. Standardisation: What is it and how is it done? Part 1. NIR news. 1993;4:8- 9.

Bouveresse E, Massart DL. Standardization of near-infrared spectrometric instruments: a review. Vibr Spectrosc. 1996;11:3-15.

Dean T, Kowalski BR. Multivariate instrument standardization: review of the state of the art. In: Brown SD, editor. Computer Assisted Analytical Spectroscopy. Chichester, UK: Wiley; 1996. p. 175-87.

Huang GB, Zhu QY, Siew CK. Extreme learning machine: Theory and applications. Neurocomputing. 2006;70(1-3):489-501.

Zhang Y, Zhang P. Optimization of nonlinear process based on sequential extreme learning machine. Chem Eng Sci. 2011;66(20):4702-10.

Sun Y, Yuan Y, Wang G. An OS-ELM based distributed ensemble classification framework in P2P networks. Neurocomputing. 2011;74(16):2438-43.

Zhao LJ, Chai TY, Diao XK, Yuan DC. Multi-class Classification with One-Against-One Using Probabilistic Extreme Learning Machine. In: Wang J, Yen G, Polycarpou M, editors. Advances in Neural Networks. Berlin: Springer Berlin Heidelberg; 2012. p. 10-19.

Qiu S, Gao L, Wang J. Classification and regression of ELM, LVQ and SVM for E-nose data of strawberry juice. J Food Eng. 2015;144:77-85.

Chacko B, Vimal Krishnan VR, Raju G, Babu Anto P. Handwritten character recognition using wavelet energy and extreme learning machine. Int J Mach Learn Cybern. 2012;3(2):149-61.

Li G, Liu M, Dong M. A new online learning algorithm for structure-adjustable extreme learning machine. Comput Math Appl. 2010;60(3):377-89.

Junhai Z, Jinggeng W, Xizhao W. Ensemble online sequential extreme learning machine for large data set classification. 2014 IEEE International Conference on Systems, Man, and Cybernetics (SMC); 2014 Oct 5-8; San Diego, USA. USA: IEEE; 2014. p. 2250-5.

Soliman OS, Aboelhamd E. Classification of breast cancer using differential evolution and least squares support vector machine. Int J Emerg Trends Technol Comput Sci. 2014;3:155-61.

Dutta RK, Karmakar NK, Si T. Artificial neural network training using fireworks algorithm in medical data mining. Int J Comput Appl. 2016;137:1-5.