Author:
Huang Hanwen,Yang Qinglong
Abstract
Abstract
Large-margin classifiers are popular methods for classification. We derive the asymptotic expression for the generalization error of a family of large-margin classifiers in the limit of both sample size n and dimension p going to ∞ with fixed ratio α = n/p. This family covers a broad range of commonly used classifiers including support vector machine, distance weighted discrimination, and penalized logistic regression. Our result can be used to establish the phase transition boundary for the separability of two classes. We assume that the data are generated from a single multivariate Gaussian distribution with arbitrary covariance structure. We explore two special choices for the covariance matrix: spiked population model and two layer neural networks with random first layer weights. The method we used for deriving the closed-form expression is from statistical physics known as the replica method. Our asymptotic results match simulations already when n, p are of the order of a few hundreds. For two layer neural networks, we reproduce the recently observed ‘double descent’ phenomenology for several classification models. We also discuss some statistical insights that can be drawn from these analysis.
Subject
Statistics, Probability and Uncertainty,Statistics and Probability,Statistical and Nonlinear Physics
Reference46 articles.
1. The committee machine: computational to statistical gaps in learning a two-layers neural network;Aubin;J. Stat. Mech.,2019
2. Kernels as features: on kernels, margins, and low-dimensional mappings;Balcan;Mach. Learn.,2006
3. The adaptive interpolation method: a simple scheme to prove replica formulas in bayesian inference;Barbier,2017
4. The LASSO risk for Gaussian matrices;Bayati;IEEE Trans. Inf. Theory,2012
5. Overfitting or perfect fitting? Risk bounds for classification and regression rules that interpolate;Belkin,2018
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献