Author:
Pandit Parthe,Sahraee-Ardakan Mojtaba,Rangan Sundeep,Schniter Philip,Fletcher Alyson K
Abstract
Abstract
We consider the problem of estimating the input and hidden variables of a stochastic multi-layer neural network (NN) from an observation of the output. The hidden variables in each layer are represented as matrices with statistical interactions along both rows as well as columns. This problem applies to matrix imputation, signal recovery via deep generative prior models, multi-task and mixed regression, and learning certain classes of two-layer NNs. We extend a recently-developed algorithm—multi-layer vector approximate message passing, for this matrix-valued inference problem. It is shown that the performance of the proposed multi-layer matrix vector approximate message passing algorithm can be exactly predicted in a certain random large-system limit, where the dimensions N × d of the unknown quantities grow as N → ∞ with d fixed. In the two-layer neural-network learning problem, this scaling corresponds to the case where the number of input features as well as training samples grow to infinity but the number of hidden nodes stays fixed. The analysis enables a precise prediction of the parameter and test error of the learning.
Subject
Statistics, Probability and Uncertainty,Statistics and Probability,Statistical and Nonlinear Physics
Reference47 articles.
1. The committee machine: computational to statistical gaps in learning a two-layers neural network;Aubin,2018
2. Optimal errors and phase transitions in high-dimensional generalized linear models;Barbier;Proc. Natl Acad. Sci. USA,2019
3. The dynamics of message passing on dense graphs, with applications to compressed sensing;Bayati;IEEE Trans. Inf. Theory,2011
4. Compressed sensing using generative models;Bora,2017
5. Sketched clustering via hybrid approximate message passing;Byrne;IEEE Trans. Signal Process.,2019
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献