The innovation of Machine Learning (ML) techniques is evolving from basic techniques to optimized techniques, considerably improving the performance of prediction models. In the proposed work, the study primarily explores fundamental ML classification methods to classify banking customers based on their credit information. The classification of customers targets five categories: Outstanding, Excellent, Good, Satisfactory, and Bad. The aim is to assess the profitable customer categories and gain successful business by offering resources. The basic classification algorithms used in the proposed work are K-Nearest Neighbour (K-NN), Support Vector Machines (SVM), Decision Tree (DT), and Random Forest (RF) Classifiers. Using standard evaluation metrics, the performance of the classifiers are evaluated. Based on the metrics the comparative analysis is conducted, and comprehended the performance metrics need to be elevated. To manipulate this, Hyperparameter GridSearchCV (HGSCV) optimization is adopted, which is putative for its exhaustive search capabilities. However, the present accuracy scores of algorithms could be slightly improved while applying the HGSCV. Subsequently, the analysis moves on to an advanced optimization meta-heuristic optimized technique known as Particle Swarm Optimization (PSO). In this approach, the GlobalBestPSO method is implemented to tune the classifiers. The performance of the optimized classifiers such as GlobalBestPSO-SVM (gbestPSO-SVM), GlobalBestPSO-KNN (gbestPSO-KNN), GlobalBestPSO-DT (gbestPSO-DT), and GlobalBestPSO-RF (gbestPSO-RF) classifier are evaluated by analyzing the chosen set of parameters. The comparison of test results demonstrates the outstanding performance metrics in the optimized method, with accuracy outperformed with exceeding 0.95 score. The proposed hybrid model, integrating GlobalBestPSO with basic classifiers, superiors both traditional classifiers and tuned model HGSCV. The analysis is concluded to figure out the performance metrics of boosted classifiers, which optimized with the GlobalBestPSO, offers superior performance than others beyond all metrics.
J VEERALAGAN, PRIYA SM. Hyper Tuning Using Gridsearchcv on Machine Learning Models for Prognosticating Dementia. Research Square Platform LLC; 2022.
2.
Etaiwi W, Biltawi M, Naymat G. Evaluation of classification algorithms for banking customer’s behavior under Apache Spark Data Processing System. Procedia Computer Science. 2017;113:559–64.
3.
Sarker IH. Machine Learning: Algorithms, Real-World Applications and Research Directions. SN Computer Science. 2021;2(3).
4.
Smeureanu I, Ruxanda G, Badea LM. CUSTOMER SEGMENTATION IN PRIVATE BANKING SECTOR USING MACHINE LEARNING TECHNIQUES. Journal of Business Economics and Management. 2013;14(5):923–39.
5.
Zeinulla E, Bekbayeva K, Yazici A. Comparative study of the classification models for prediction of bank telemarketing. 2018 IEEE 12th International Conference on Application of Information and Communication Technologies (AICT). IEEE; 2018. p. 1–5.
6.
Dawood EAE, Elfakhrany E, Maghraby FA. Improve Profiling Bank Customer’s Behavior Using Machine Learning. IEEE Access. 2019;7:109320–7.
7.
Charbuty B, Abdulazeez A. Classification Based on Decision Tree Algorithm for Machine Learning. Journal of Applied Science and Technology Trends. 2021;2(01):20–8.
8.
Anuradha, Gupta G. A self explanatory review of decision tree classifiers. International Conference on Recent Advances and Innovations in Engineering (ICRAIE-2014). IEEE; 2014. p. 1–7.
9.
Patel H, Prajapati P. Study and analysis of decision tree-based classification algorithms. International Journal of Computer Sciences and Engineering. 2018;(10):74–8.
10.
Taha Chicho B, Mohsin Abdulazeez A, Qader Zeebaree D, Assad Zebari D. Machine Learning Classifiers Based Classification For IRIS Recognition. Qubahan Academic Journal. 2021;1(2):106–18.
11.
Khorshid S, Abdulazeez A. Breast cancer diagnosis based on k-nearest neighbors: a review. PalArch’s Journal of Archaeology of Egypt/Egyptology. 2021;(4):1927–51.
12.
Zebari DA, Zeebaree DQ, Abdulazeez AM, Haron H, Hamed HNA. Improved Threshold Based and Trainable Fully Automated Segmentation for Breast Cancer Boundary and Pectoral Muscle in Mammogram Images. IEEE Access. 2020;8:203097–116.
13.
Ahmed NS, Hikmat Sadiq M. Clarify of the Random Forest Algorithm in an Educational Field. 2018 International Conference on Advanced Science and Engineering (ICOASE). IEEE; 2018. p. 179–84.
14.
Priyanka NA, Kumar D. Decision tree classifier: a detailed survey. International Journal of Information and Decision Sciences. 2020;12(3):246.
15.
Pandey A, Singh P. A systematic survey of classification algorithms for cancer detection. Int J Data Informatics Intell Comput. 2022;(2):34–50.
16.
Wang S, Lu H, Khan A, Hajati F, Khushi M, Uddin S. A machine learning software tool for multiclass classification. Software Impacts. 2022;13:100383.
17.
Belete DM, Huchaiah MD. Grid search in hyperparameter optimization of machine learning models for prediction of HIV/AIDS test results. International Journal of Computers and Applications. 2021;44(9):875–86.
18.
Religia YR, Pranoto GT, Suwancita IM. Analysis of the Use of Particle Swarm Optimization on Naïve Bayes for Classification of Credit Bank Applications. JISA(Jurnal Informatika dan Sains). 2021;4(2):133–7.
19.
Chopard B, Tomassini M. Particle Swarm Optimization. Natural Computing Series. Springer International Publishing; 2018. p. 97–102.
20.
Kanaka Vardhini K, Sitamahalakshmi T. A Review on Nature-based Swarm Intelligence Optimization Techniques and its Current Research Directions. Indian Journal of Science and Technology. 2016;9(10).
21.
Li J, Ding L, Li B. A Novel Naive Bayes Classification Algorithm Based on Particle Swarm Optimization. The Open Automation and Control Systems Journal. 2014;6(1):747–53.
22.
Lamba A, Kumar D. Survey on KNN and its variants. Int J Adv Res Comput Commun Eng. 2016;(5):430–5.
23.
Ibrahim I, Abdulazeez A. The Role of Machine Learning Algorithms for Diagnosing Diseases. Journal of Applied Science and Technology Trends. 2021;2(01):10–9.
24.
Ray S. A Quick Review of Machine Learning Algorithms. 2019 International Conference on Machine Learning, Big Data, Cloud and Parallel Computing (COMITCon). IEEE; 2019. p. 35–9.
25.
ALMASI ON, ROUHANI M. Fast and de-noise support vector machine training method based on fuzzy clustering method for large real world datasets. TURKISH JOURNAL OF ELECTRICAL ENGINEERING & COMPUTER SCIENCES. 2016;24:219–33.
26.
Cheushev V, Simovici DA, Shmerko V, Yanushkevich S. Functional entropy and decision trees. Proceedings. 1998 28th IEEE International Symposium on Multiple- Valued Logic (Cat. No.98CB36138). IEEE Comput. Soc; p. 257–62.
27.
Molala R, Entropy. Information Gain, Gini Index-The Crux of a Decision Tree. Medium. 2020;
28.
Sarica A, Cerasa A, Quattrone A. Random Forest Algorithm for the Classification of Neuroimaging Data in Alzheimer’s Disease: A Systematic Review. Frontiers in Aging Neuroscience. 2017;9.
29.
DeCastro-García N, Muñoz Castañeda ÁL, Escudero García D, Carriegos MV. Effect of the Sampling of a Dataset in the Hyperparameter Optimization Phase over the Efficiency of a Machine Learning Algorithm. Complexity. 2019;2019(1).
30.
Eberhart R, Kennedy J. A new optimizer using particle swarm theory. MHS’95. Proceedings of the Sixth International Symposium on Micro Machine and Human Science. IEEE; p. 39–43.
31.
Kennedy J, Eberhart R. Particle swarm optimization. Vol. 4, Proceedings of ICNN’95 - International Conference on Neural Networks. IEEE; p. 1942–8.
The statements, opinions and data contained in the journal are solely those of the individual authors and contributors and not of the publisher and the editor(s). We stay neutral with regard to jurisdictional claims in published maps and institutional affiliations.