Voir la notice de l'article provenant de la source Library of Science
@article{IJAMCS_2024_34_1_a10, author = {Czmil, Sylwester and Kluska, Jacek and Czmil, Anna}, title = {An empirical study of a simple incremental classifier based on vector quantization and adaptive resonance theory}, journal = {International Journal of Applied Mathematics and Computer Science}, pages = {149--165}, publisher = {mathdoc}, volume = {34}, number = {1}, year = {2024}, language = {en}, url = {http://geodesic.mathdoc.fr/item/IJAMCS_2024_34_1_a10/} }
TY - JOUR AU - Czmil, Sylwester AU - Kluska, Jacek AU - Czmil, Anna TI - An empirical study of a simple incremental classifier based on vector quantization and adaptive resonance theory JO - International Journal of Applied Mathematics and Computer Science PY - 2024 SP - 149 EP - 165 VL - 34 IS - 1 PB - mathdoc UR - http://geodesic.mathdoc.fr/item/IJAMCS_2024_34_1_a10/ LA - en ID - IJAMCS_2024_34_1_a10 ER -
%0 Journal Article %A Czmil, Sylwester %A Kluska, Jacek %A Czmil, Anna %T An empirical study of a simple incremental classifier based on vector quantization and adaptive resonance theory %J International Journal of Applied Mathematics and Computer Science %D 2024 %P 149-165 %V 34 %N 1 %I mathdoc %U http://geodesic.mathdoc.fr/item/IJAMCS_2024_34_1_a10/ %G en %F IJAMCS_2024_34_1_a10
Czmil, Sylwester; Kluska, Jacek; Czmil, Anna. An empirical study of a simple incremental classifier based on vector quantization and adaptive resonance theory. International Journal of Applied Mathematics and Computer Science, Tome 34 (2024) no. 1, pp. 149-165. http://geodesic.mathdoc.fr/item/IJAMCS_2024_34_1_a10/
[1] Alcalá-Fdez, J., Fernandez, A., Luengo, J. , Derrac, J. , García, S., Sanchez, L. and Herrera, F. (2011). KEEL data-mining software tool: Data set repository, integration of algorithms and experimental analysis framework, Journal of Multiple-Valued Logic and Soft Computing 17(2-3): 255-287.
[2] Altman, N.S. (1992). An introduction to kernel and nearest-neighbor nonparametric regression, The American Statistician 46(3): 175-185.
[3] Banerjee, S., Bhattacharjee, P. and Das, S. (2017). Performance of deep learning algorithms vs. shallow models, in extreme conditions - Some empirical studies, in B.U. Shankar et al. (Eds), Pattern Recognition and Machine Intelligence, Springer International Publishing, Cham, pp. 565-574.
[4] Bifet, A. and Gavaldà, R. (2009). Adaptive learning from evolving data streams, in N.M Adams et al. (Eds), Advances in Intelligent Data Analysis VIII, Springer, Berlin/Heidelberg, pp. 249-260.
[5] Breiman, L. (2001). Random forests, Machine Learning 45(1): 5-32.
[6] Breiman, L., Friedman, J.H., Olshen, R.A. and Stone, C.J. (1984). Classification and Regression Trees, Wadsworth International Group, Belmont.
[7] Carpenter, G.A. and Grossberg, S. (1987). A massively parallel architecture for a self-organizing neural pattern recognition machine, Computer Vision, Graphics, and Image Processing 37(1): 54-115.
[8] Carpenter, G.A., Grossberg, S. and Reynolds, J.H. (1991). ARTMAP: Supervised real-time learning and classification of nonstationary data by a self-organizing neural network, Neural Networks 4(5): 565-588.
[9] Carpenter, G., Grossberg, S., Markuzon, N. , Reynolds, J.H. and Rosen, D.B. (1992). Fuzzy ARTMAP: A neural network architecture for incremental supervised learning of analog multidimensional maps, IEEE Transactions on Neural Networks 3(5): 698-713.
[10] Chan, T.F., Golub, G.H. and LeVeque, R.J. (1979). Updating formulae and an pairwise algorithm for computing sample variances, Stanford Working Paper STAN-CS-79-773: 1-22, Stanford University, Stanford, http://i.stanford.edu/pub/cstr/reports/cs/tr/79/773/CS-TR-79-773.pdf.
[11] Chang, C.-C. and Lin, C.-J. (2011). LIBSVM: A library for support vector machine, ACM Transactions on Intelligent Systems and Technology 2(3): 1-27.
[12] Chen, T. and Guestrin, C. (2016). XGBoost: A scalable tree boosting system, Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, USA, pp. 785-794.
[13] Czmil, S. (2021). Python implementation of evolving vector quantization for classification of on-line data streams (Version 0.0.2), Computer software, https://github.com/sylwekczmil/evq.
[14] Czmil, S., Kluska, J. and Czmil, A. (2022). CACP: Classification algorithms comparison pipeline, SoftwareX 19: 101134.
[15] Duch, W., Adamczak, R. and Diercksen, G.H.F. (2000). Classification, association and pattern completion using neural similarity based methods, International Journal of Applied Mathematics and Computer Science 10(4): 747-766.
[16] Elsayad, A.M. (2009). Classification of ECG arrhythmia using learning vector quantization neural networks, 2009 International Conference on Computer Engineering & Systems, Cairo, Egypt, pp. 139-144.
[17] Fernández-Delgado, M., Cernadas, E., Barro, S. and Amorim, D. (2014). Do we need hundreds of classifiers to solve real world classification problems?, Journal of Machine Learning Research 15(90): 3133-3181.
[18] Friedman, J.H. (2001). Greedy function approximation: A gradient boosting machine, The Annals of Statistics 29(5): 1189-1232.
[19] Galbraith, B. (2017). Adaptive resonance theory models, Computer software, https://github.com/AIOpenLab/art.
[20] Gomes, H.M., Bifet, A., Read, J., Barddal, J.P., Enembreck, F., Pfharinger, B., Holmes, G. and Abdessalem, T. (2017). Adaptive random forests for evolving data stream classification, Machine Learning 106(9-10): 1469-1495.
[21] Hastie, T., Rosset, S., Zhu, J. and Zou, H. (2009). Multi-class AdaBoost, Statistics and Its Interface 2(3): 349-360.
[22] Hastie, T., Tibshirani, R. and Friedman, J. (2008). The Elements of Statistical Learning, Springer, New York.
[23] Holte, R.C. (1993). Very simple classification rules perform well on most commonly used data sets, Machine Learning 11(1): 63-90.
[24] Huang, J. and Ling, C.X. (2005). Using AUC and accuracy in evaluating learning algorithms, IEEE Transactions on Knowledge and Data Engineering 17(3): 299-310.
[25] Hulten, G., Spencer, L. and Domingos, P. (2001). Mining time-changing data streams, Proceedings of the 7th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD’01, San Francisco, USA, pp. 97-106.
[26] James, G., Witten, D., Hastie, T. and Tibshirani, R. (2013). An Introduction to Statistical Learning: With Applications in R, Springer, New York.
[27] Kasuba, T. (1993). Simplified fuzzy ARTMAP, AI Expert 8: 18-25.
[28] Kingma, D.P. and Ba, J. (2015). Adam: A method for stochastic optimization, Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, USA.
[29] Kluska, J. and Madera, M. (2021). Extremely simple classifier based on fuzzy logic and gene expression programming, Information Sciences 571: 560-579.
[30] Kohonen, T., Schroeder, M.R. and Huang, T.S. (2001). Self-Organizing Maps, 3rd Edn, Springer-Verlag, Berlin/Heidelberg.
[31] Kolter, J.Z. and Maloof, M.A. (2005). Using additive expert ensembles to cope with concept drift, Proceedings of the 22nd International Conference on Machine Learning (ICML-2005), Bonn, Germany, pp. 449-456.
[32] Kolter, J.Z. and Maloof, M.A. (2007). Dynamic weighted majority: An ensemble method for drifting concepts, Journal of Machine Learning Research 8(91): 2755-2790.
[33] Kulczycki, P. and Kowalski, P.A. (2015). Bayes classification for nonstationary patterns, International Journal of Computational Methods 12(02): 1550008.
[34] Kusy,M. and Zajdel, R. (2021). A weighted wrapper approach to feature selection, International Journal of Applied Mathematics and Computer Science 31(4): 685-696, DOI: 10.34768/amcs-2021-0047.
[35] Lang, K.J. and Witbrock, M.J. (1988). Learning to tell two spirals apart, The 1988 Connectionist Models Summer School, Pittsburgh, USA, pp. 52-59.
[36] Lee, S., Chang, K. and Baek, J.-G. (2021). Incremental learning using generative-rehearsal strategy for fault detection and classification, Expert Systems with Applications 184: 115477.
[37] Leo, J. and Kalita, J. (2022). Incremental deep neural network learning using classification confidence thresholding, IEEE Transactions on Neural Networks and Learning Systems 33(12): 7706-7716.
[38] Lughofer, E. (2008a). Evolving vector quantization for classification of on-line data streams, 2008 International Conference on Computational Intelligence for Modelling Control & Automation (CIMCA 2008), Vienna, Austria, pp. 779-784.
[39] Lughofer, E. (2008b). Extensions of vector quantization for incremental clustering, Pattern Recognition 41(3): 995-1011.
[40] Luo, Y., Yin, L., Bai, W. and Mao, K. (2020). An appraisal of incremental learning methods, Entropy 22(11): 1190.
[41] Manapragada, C., Webb, G.I. and Salehi, M. (2018). Extremely fast decision tree, Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, pp. 1953-1962.
[42] Montiel, J., Read, J., Bifet, A. and Abdessalem, T. (2018). Scikit-Multiflow: A multi-output streaming framework, Journal of Machine Learning Research 19(72): 1-5.
[43] Oza, N.C. and Russell, S.J. (2001). Online bagging and boosting, in T.S. Richardson and T.S. Jaakkola (Eds), Proceedings of the 8th International Workshop on Artificial Intelligence and Statistics, Key West, USA, pp. 229-236.
[44] Pedregosa, F., Varoquaux, G., Gramfort, A. Michel, V., Thirion, B., Grisel, O., Blondel, M., Müller, A., Nothman, J., Louppe, G., Prettenhofer, P., Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cournapeau, D., Brucher, M., Perrot, M. and Duchesnay, E. (2011). Scikit-learn: Machine learning in Python, Journal of Machine Learning Research 12: 2825-2830.
[45] Polikar, R., Upda, L., Upda, S. and Honavar, V. (2001). Learn++: An incremental learning algorithm for supervised neural networks, IEEE Transactions on Systems, Man, and Cybernetics C: Applications and Reviews 31(4): 497-508.
[46] Pratama, M., Pedrycz, W. and Lughofer, E. (2018). Evolving ensemble fuzzy classifier, IEEE Transactions on Fuzzy Systems 26(5): 2552-2567.
[47] Pratama, M., Pedrycz, W. and Webb, G.I. (2020). An incremental construction of deep neuro fuzzy system for continual learning of nonstationary data streams, IEEE Transactions on Fuzzy Systems 28(7): 1315-1328.
[48] Rutkowski, L. and Cierniak, R. (1996). Image compression by competitive learning neural networks and predictive vector quantization, Applied Mathematics and Computer Science 6(3): 431-445.
[49] Shevchuk, Y. (2015). NeuPy (Version 1.18.5), Computer software, http://neupy.com/.
[50] Shi, X., Wong, Y.D., Li, M.Z.-F., Palanisamy, C. and Chai, C. (2019). A feature learning approach based on XGBoost for driving assessment and risk prediction, Accident Analysis & Prevention 129: 170-179.
[51] Skubalska-Rafajlowicz, E. (2000). One-dimensional Kohonen LVQ nets for multidimensional pattern recognition, International Journal of Applied Mathematics and Computer Science 10(4): 767-778.
[52] Sokolova, M. and Lapalme, G. (2009). A systematic analysis of performance measures for classification tasks, Information Processing & Management 45(4): 427-437.
[53] Stapor, K. (2018). Evaluating and comparing classifiers: Review, some recommendations and limitations, in M. Kurzynski et al. (Eds), Proceedings of the 10th International Conference on Computer Recognition Systems, CORES 2017, Springer International Publishing, Cham, pp. 12-21.
[54] Tantithamthavorn, C., McIntosh, S., Hassan, A.E. and Matsumoto, K. (2019). The impact of automated parameter optimization on defect prediction models, IEEE Transactions on Software Engineering 45(7): 683-711.
[55] Tibshirani, R., Hastie, T., Narasimhan, B. and Chu, G. (2002). Diagnosis of multiple cancer types by shrunken centroids of gene expression, Proceedings of the National Academy of Sciences 99(10): 6567-6572.
[56] Trawiński, B., Smętek, M., Telec, Z. and Lasota, T. (2012). Nonparametric statistical analysis for multiple comparison of machine learning regression algorithms, International Journal of Applied Mathematics and Computer Science 22(4): 867-881, DOI: 10.2478/v10006-012-0064-z.
[57] Vakil-Baghmisheh, M. and Pavešić, N. (2003). A fast simplified fuzzy ARTMAP network, Neural Processing Letters 17(3): 273-316.
[58] Villuendas-Rey, Y., Rey-Benguría, C.F., Ángel Ferreira-Santiago, Camacho-Nieto, O. and Yáñez-Márquez, C. (2017). The naïve associative classifier (NAC): A novel, simple, transparent, and accurate classification model evaluated on financial data, Neurocomputing 265: 105-115.
[59] Wolpert, D. and Macready, W. (1997). No free lunch theorems for optimization, IEEE Transactions on Evolutionary Computation 1(1): 67-82.
[60] Żabiński, T., Maczka, T. and Kluska, J. (2017). Industrial platform for rapid prototyping of intelligent diagnostic systems, in W. Mitkowski et al. (Eds), Trends in Advanced Intelligent Control, Optimization and Automation, Springer International Publishing, Cham, pp. 712-721.
[61] Żabiński, T., Maczka, T., Kluska, J., Kusy, M., Hajduk, Z. and Prucnal, S. (2014). Failures prediction in the cold forging process using machine learning methods, in L. Rutkowski et al. (Eds), Artificial Intelligence and Soft Computing, Springer International Publishing, Cham, pp. 622-633.
[62] Škrjanc, I., Iglesias, J.A., Sanchis, A., Leite, D., Lughofer, E. and Gomide, F. (2019). Evolving fuzzy and neuro-fuzzy approaches in clustering, regression, identification, and classification: A survey, Information Sciences 490: 344-368.