Prediction of Loan Eligibility Approval
Keywords:
Support Vector Machine (SVM), K-Nearest Neighbors (KNN), Synthetic Minority Oversampling Technique (SMOTE), Gaussian Naïve Bayes (GNB), Machine Learning (ML)Abstract
The data available through Kaggle includes loan-specific features comprising income level along with credit history as well as asset ownership information and the loan value and education level of the applicant. The main goal consists of creating an accurate forecasting model to categorize loan applications into eligible or ineligible decisions. Seven supervised learning approaches including Decision Tree and Random Forest and Logistic Regression together with SVM and KNN and GNB have been applied for evaluation. SMOTE was used to overcome the unequal distribution of loan status classification labels. The accuracy rate of 98.55% was achieved by Random Forest among the group of models and resulted in superior performance compared to other classifiers regarding precision and recall metrics. The integrated Django web application allows users to provide new data inputs which generate real-time eligibility predictions using the trained model. Such a system represents an efficient and expandable solution designed for institutions trying to automate their loan application review methods.
Downloads
References
T. Kaur and M. Goel, “Loan Approval Prediction using ML Techniques,” International Journal of Scientific Research in Computer Science, Engineering and Information Technology, vol. 5, no. 3, pp. 245–250, 2019.
S. B. Patil and A. V. Mudholkar, “Application of ML for Predicting Loan Approval,” International Journal of Computer Applications, vol. 179, no. 32, pp. 12–16, Apr. 2018.
P. Sharma and R. Gupta, “An Analysis of Loan Prediction Using Random Forest Algorithm,” International Journal of Advanced Research in Computer Science, vol. 9, no. 2, pp. 789–793, 2018.
N. Choudhury and S. Pal, “A Comparative Study of ML Algorithms for Loan Approval Prediction,” Procedia Computer Science, vol. 167, pp. 2594–2601, 2020.
L. Breiman, “Random Forests,” ML, vol. 45, no. 1, pp. 5–32, 2001.
A. Fernandez et al., “SMOTE for Learning from Imbalanced Data: Progress and Challenges, Marking the 15-Year Anniversary,” Journal of Artificial Intelligence Research, vol. 61, pp. 863–905, 2018.
M. Galar, A. Fernández, E. Barrenechea, H. Bustince, and F. Herrera, “A Review on Ensembles for the Class Imbalance Problem: Bagging-, Boosting-, and Hybrid-Based Approaches,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 42, no. 4, pp. 463–484, 2012.
D. Dua and C. Graff, “UCI ML Repository,” University of California, Irvine, School of Information and Computer Sciences, 2017. [Online]. Available: https://archive.ics.uci.edu/ml
J. Han, M. Kamber, and J. Pei, Data Mining: Concepts and Techniques, 3rd ed., Morgan Kaufmann Publishers, 2011.
B. Lantz, ML with Python: Essential Techniques for Predictive Analysis, 2nd ed., Packt Publishing, 2019.
P. Domingos, “A Few Useful Things to Know About ML,” Communications of the ACM, vol. 55, no. 10, pp. 78–87, 2012.
S. Raschka and V. Mirjalili, Python ML: ML and Deep Learning with Python, scikit-learn, and TensorFlow 2, 3rd ed., Packt Publishing, 2019.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 International Journal of Scientific Research in Computer Science, Engineering and Information Technology

This work is licensed under a Creative Commons Attribution 4.0 International License.