Abstract
AbstractBits manipulation in traditional memory writing is commonly done through quasi-static operations. While simple to model, this method is known to reduce memory capacity. We demonstrate how a reinforcement learning agent can exploit the dynamical response of a simple multi-bit mechanical system to restore its memory. To do so, we introduce a model framework consisting of a chain of bi-stable springs manipulated on one end by the external action of the agent. We show that the agent learns how to reach all available states for three springs, even though some states are not reachable through adiabatic manipulation, and that training is significantly improved using transfer learning techniques. Interestingly, the agent also points to an optimal system design by taking advantage of the underlying physics. Indeed, the control time exhibits a non-monotonic dependence on the internal dissipation, reaching a minimum at a cross-over shown to verify a mechanically motivated scaling relation.
Funder
The Raymond and Beverly Sackler Post-Doctoral Scholarship
Publisher
Springer Science and Business Media LLC
Subject
General Physics and Astronomy
Reference44 articles.
1. Keim, N. C., Paulsen, J. D., Zeravcic, Z., Sastry, S. & Nagel, S. R. Memory formation in matter. Rev. Mod. Phys. 91, 035002 (2019).
2. Kovacs, A. Glass transition in amorphous polymers: a phenomenological study. Adv. Polym. Sci. 3, 394–507 (1963).
3. Strukov, D. B., Snider, G. S., Stewart, D. R. & Williams, R. S. The missing memristor found. Nature 453, 80–83 (2018).
4. Prados, A. & Trizac, E. Kovacs-like memory effect in driven granular gases. Phys. Rev. Lett. 112, 198001 (2014).
5. Jules, T., Lechenault, F. & Adda-Bedia, M. Plasticity and aging of folded elastic sheets. Phys. Rev. E 102, 033005 (2020).