Abstract
Abstract
One of the most common metrics to evaluate neural network classifiers is the area under the receiver operating characteristic curve (AUC). However, optimisation of the AUC as the loss function during network training is not a standard procedure. Here we compare minimising the cross-entropy (CE) loss and optimising the AUC directly. In particular, we analyse the loss function landscape (LFL) of approximate AUC (appAUC) loss functions to discover the organisation of this solution space. We discuss various surrogates for AUC approximation and show their differences. We find that the characteristics of the appAUC landscape are significantly different from the CE landscape. The approximate AUC loss function improves testing AUC, and the appAUC landscape has substantially more minima, but these minima are less robust, with larger average Hessian eigenvalues. We provide a theoretical foundation to explain these results. To generalise our results, we lastly provide an overview of how the LFL can help to guide loss function analysis and selection.
Funder
Agence Nationale de la Recherche
Subject
Artificial Intelligence,Human-Computer Interaction,Software
Reference44 articles.
1. Surrogate regret bounds for bipartite ranking via strongly proper losses;Agarwal;J. Mach. Learn. Res.,2014
2. Shaping the learning landscape in neural networks around wide flat minima;Baldassi;Proc. Natl Acad. Sci.,2020
3. Energy landscapes for machine learning;Ballard;Phys. Chem. Chem. Phys.,2017
4. The topology of multidimensional potential energy surfaces: theory and application to peptide structure and kinetics;Becker;J. Chem. Phys.,1997
5. Loss functions for binary class probability estimation and classification: structure and applications;Buja;Work. Draft,2005
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献