Author:
Marzen Sarah E.,Riechers Paul M.,Crutchfield James P.
Abstract
AbstractRecurrent neural networks are used to forecast time series in finance, climate, language, and from many other domains. Reservoir computers are a particularly easily trainable form of recurrent neural network. Recently, a “next-generation” reservoir computer was introduced in which the memory trace involves only a finite number of previous symbols. We explore the inherent limitations of finite-past memory traces in this intriguing proposal. A lower bound from Fano’s inequality shows that, on highly non-Markovian processes generated by large probabilistic state machines, next-generation reservoir computers with reasonably long memory traces have an error probability that is at least $$\sim 60\%$$
∼
60
%
higher than the minimal attainable error probability in predicting the next observation. More generally, it appears that popular recurrent neural networks fall far short of optimally predicting such complex processes. These results highlight the need for a new generation of optimized recurrent neural network architectures. Alongside this finding, we present concentration-of-measure results for randomly-generated but complex processes. One conclusion is that large probabilistic state machines—specifically, large $$\epsilon$$
ϵ
-machines—are key to generating challenging and structurally-unbiased stimuli for ground-truthing recurrent neural network architectures.
Funder
Air Force Office of Scientific Research
Templeton World Charity Foundation
Foundational Questions Institute
U.S. Army Research Office
U.S. Department of Energy
Publisher
Springer Science and Business Media LLC
Reference40 articles.
1. Lipton, Z.C., Berkowitz, J. & Elkan, C. A critical review of recurrent neural networks for sequence learning. arXiv preprint arXiv:1506.00019, (2015).
2. Schrauwen, B., Verstraeten, D. & Van Campenhout, J. An overview of reservoir computing: theory, applications and implementations. In Proceedings of the 15th European Symposium on Artificial Neural Networks (p. 471-482 2007).
3. Hsu, A. & Marzen, S. E. Strange properties of linear reservoirs in the infinitely large limit for prediction of continuous-time signals. J. Stat. Phys. 190(2), 1–16 (2023).
4. Krishnamurthy, K., Can, T. & Schwab, D. J. Theory of gating in recurrent neural networks. Phys. Rev. X 12(1), 011011 (2022).
5. Gauthier, D. J., Bollt, E., Griffith, A. & Barbosa, W. A. S. Next generation reservoir computing. Nat. Commun. 12(1), 1–8 (2021).