AN EFFICIENT METHOD FOR MULTICLASS CLASSIFICATION IN TABULAR DATA USING MULTILAYER PERCEPTRON MODEL
Keywords:
Multilayer Perceptron (MLP), Neural Networks, Multiclass Classification, Tabular Data, Feature Scaling, Dropout Regularization, One-Hot Encoding, Categorical Cross-Entropy, Adam Optimizer, Early Stopping, Model Generalization, Structured Data, Machine Learning, Deep Learning, Hyperparameter TuningAbstract
In recent years, neural networks have shown remarkable promise in addressing various classification challenges, especially in fields like image and text processing. However, their effectiveness on structured tabular data remains relatively underexplored. This research examines the use of Multilayer Perceptron (MLP) neural networks for multiclass classification tasks on tabular datasets. Preprocessing steps, including feature scaling and one-hot encoding, were applied to enhance performance. Different MLP architectures were evaluated, with variations in the number of hidden layers and neurons, as well as the use of regularization methods such as dropout to mitigate overfitting. Early stopping was also utilized to improve generalization. The model was trained using categorical cross-entropy loss and the Adam optimizer, achieving a good balance between training accuracy and validation results. Findings suggest that MLPs can be effectively applied to multiclass classification on tabular data, though performance is highly dependent on hyperparameter selection and preprocessing techniques. This study outlines key approaches for optimizing MLP architectures for structured data and provides insights into best practices for neural network classification tasks in similar contexts.
References
LeCun, Y., Bengio, Y., & Hinton, G. (2015). "Deep Learning." Nature, 521(7553), 436-444. DOI: 10.1038/nature14539
Zhang, Z. (2017). "Improved Adam Optimizer for Deep Neural Networks." IEEE/ACM Transactions on Audio, Speech, and Language Processing, 26(1), 23-35.DOI: 10.1109/TASLP.2017.2771221
Choi, E., Bahadori, M. T., Sun, J., Kulas, J., Schuetz, A., & Stewart, W. F. (2016). "RETAIN: An Interpretable Predictive Model for Healthcare Using Reverse Time Attention Mechanism." Advances in Neural Information Processing Systems, 29, 3504-3512.DOI: 10.48550/arXiv.1608.05745
Shickel, B., Tighe, P. J., Bihorac, A., & Rashidi, P. (2018). "Deep EHR: A Survey of Recent Advances in Deep Learning Techniques for Electronic Health Record (EHR) Analysis." IEEE Journal of Biomedical and Health Informatics, 22(5), 1589-1604. DOI: 10.1109/JBHI.2017.2767063
Xu, Y., Mo, T., Feng, Q., Zhong, P., Lai, M., & Chang, S. (2016). "Deep Learning of Feature Representation with Multiple Instance Learning for Medical Image Analysis." IEEE Transactions on Neural Networks and Learning Systems, 27(11), 2214-2226. DOI: 10.1109/TNNLS.2015.2424995
Lipton, Z. C., Kale, D. C., Elkan, C., & Wetzel, R. (2016). "Learning to Diagnose with LSTM Recurrent Neural Networks." arXiv preprint arXiv:1511.03677. DOI: 10.48550/arXiv.1511.03677
Suresh, H., Hunt, N., Johnson, A., Celi, L. A., Szolovits, P., & Ghassemi, M. (2017). "Clinical Intervention Prediction and Understanding with Deep Neural Networks." In Proceedings of Machine Learning for Healthcare, 243-256. DOI: 10.48550/arXiv.1706.03456
Esteva, A., Kuprel, B., Novoa, R. A., Ko, J., Swetter, S. M., Blau, H. M., & Thrun, S. (2017). "Dermatologist-level Classification of Skin Cancer with Deep Neural Networks." Nature, 542(7639), 115-118. DOI: 10.1038/nature21056
Rajkomar, A., Oren, E., Chen, K., Dai, A. M., Hajaj, N., Liu, P. J., ... & Dean, J. (2018). "Scalable and Accurate Deep Learning with Electronic Health Records." npj Digital Medicine, 1(1), 1-10. DOI: 10.1038/s41746-018-0029-1
Miotto, R., Li, L., Kidd, B. A., & Dudley, J. T. (2016). "Deep Patient: An Unsupervised Representation to Predict the Future of Patients from the Electronic Health Records." Scientific Reports, 6(1), 1-10. DOI: 10.1038/srep26094
Published
How to Cite
Issue
Section
License
Copyright (c) 2024 Sharipov D.K., Saidov A.D.
This work is licensed under a Creative Commons Attribution 4.0 International License.