Abstract
This paper compares model development strategies based on different performance metrics. The study was conducted in the area of credit risk modeling with the usage of diverse metrics, including general-purpose Area Under the ROC curve (AUC), problem-dedicated Expected Maximum Profit (EMP) and the novel case-tailored Calculated Profit (CP). The metrics were used to optimize competitive credit risk scoring models based on two predictive algorithms that are widely used in the financial industry: Logistic Regression and extreme gradient boosting machine (XGBoost). A dataset provided by the American Fannie Mae agency was utilized to conduct the study. In addition to the baseline study, the paper also includes a stability analysis. In each case examined the proposed CP metric that allowed us to achieve the most profitable loan portfolio.
Funder
Ministry of Education, Poland
Subject
General Physics and Astronomy
Reference45 articles.
1. Weerts, H.J.P., Mueller, A.C., and Vanschoren, J. (2020). Importance of Tuning Hyperparameters of Machine Learning Algorithms. arXiv.
2. A comparative study on the effect of feature selection on classification accuracy;Karabulut;Procedia Technol.,2012
3. Algorithm selection via meta-learning and sample-based active testing;Abdulrahman;CEUR-WS,2015
4. Performance analysis of cost-sensitive learning methods with application to imbalanced medical data;Mienye;Inform. Med. Unlocked,2021
5. Li Yiheng, C.W. (2020). A Comparative Performance Assessment of Ensemble Learning for Credit Scoring. Mathematics, 8.
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献