Affiliation:
1. Donders Institute for Brain, Cognition, and Behaviour, Radboud University, 6525 XZ Nijmegen, The Netherlands
Abstract
Uncovering the mechanisms behind long-term memory is one of the most fascinating open problems in neuroscience and artificial intelligence. Artificial associative memory networks have been used to formalize important aspects of biological memory. Generative diffusion models are a type of generative machine learning techniques that have shown great performance in many tasks. Similar to associative memory systems, these networks define a dynamical system that converges to a set of target states. In this work, we show that generative diffusion models can be interpreted as energy-based models and that, when trained on discrete patterns, their energy function is (asymptotically) identical to that of modern Hopfield networks. This equivalence allows us to interpret the supervised training of diffusion models as a synaptic learning process that encodes the associative dynamics of a modern Hopfield network in the weight structure of a deep neural network. Leveraging this connection, we formulate a generalized framework for understanding the formation of long-term memory, where creative generation and memory recall can be seen as parts of a unified continuum.
Reference53 articles.
1. Neural networks and physical systems with emergent collective computational abilities;Hopfield;Proc. Natl. Acad. Sci. USA,1982
2. Associative memories via artificial neural networks;Michel;IEEE Control Syst. Mag.,1990
3. Hybrid computing using a neural network with dynamic external memory;Graves;Nature,2016
4. Lopez-Paz, D., and Ranzato, M. (2017). Gradient episodic memory for continual learning. Adv. Neural Inf. Process. Syst., 30.
5. Improved schemes for episodic memory-based lifelong learning;Guo;Adv. Neural Inf. Process. Syst.,2020
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献