Author:
Ivanov Dmitry A.,Larionov Denis A.,Kiselev Mikhail V.,Dylov Dmitry V.
Abstract
AbstractWe propose a sparse computation method for optimizing the inference of neural networks in reinforcement learning (RL) tasks. Motivated by the processing abilities of the brain, this method combines simple neural network pruning with a delta-network algorithm to account for the input data correlations. The former mimics neuroplasticity by eliminating inefficient connections; the latter makes it possible to update neuron states only when their changes exceed a certain threshold. This combination significantly reduces the number of multiplications during the neural network inference for fast neuromorphic computing. We tested the approach in popular deep RL tasks, yielding up to a 100-fold reduction in the number of required multiplications without substantial performance loss (sometimes, the performance even improved).
Funder
Russian Foundation for Basic Research
Publisher
Springer Science and Business Media LLC
Reference33 articles.
1. Ivanov, D., Chezhegov, A., Kiselev, M., Grunin, A. & Larionov, D. Neuromorphic artificial intelligence systems. Frontiers in Neuroscience 16 (2022).
2. Attwell, D. & Laughlin, S. B. An energy budget for signaling in the grey matter of the brain. Journal of Cerebral Blood Flow & Metabolism 21, 1133–1145 (2001).
3. Bellemare, M., Veness, J. & Bowling, M. Investigating contingency awareness using atari 2600 games. In Proceedings of the AAAI Conference on Artificial Intelligence 26, 864–871 (2012).
4. Hudspeth, A. J., Jessell, T. M., Kandel, E. R., Schwartz, J. H. & Siegelbaum, S. A. Principles of neural science (McGraw-Hill, Health Professions Division, 2013).
5. Blalock, D., Ortiz, J. J. G., Frankle, J. & Guttag, J. What is the state of neural network pruning? arXiv preprint arXiv:2003.03033 (2020).