1. R. Adamczak. “A note on the Hanson-Wright inequality for random vectors with dependencies”. Electron. Commun. Probab. 20 (2015), no. 72, 13. DOI: 10.1214/ECP.v20-3829.
2. K. Ahn, C. Yun, and S. Sra. “SGD with shuffling: optimal rates without component convexity and large epoch requirements”. In: Advances in Neural Information Processing Systems. Ed. by H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, and H. Lin. Vol. 33. Curran Associates, Inc., 2020, pp. 17526–17535.
3. G. B. Arous, R. Gheissari, and A. Jagannath. “Online stochastic gradient descent on non-convex losses from high-dimensional inference”. The Journal of Machine Learning Research 22.1 (2021), pp. 4788–4838.
4. G. B. Arous, R. Gheissari, and A. Jagannath. “High-dimensional limit theorems for SGD: Effective dynamics and critical scaling”. In: Advances in Neural Information Processing Systems. Ed. by A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho. 2022.
5. S. Asmussen. Applied probability and queues. Second. Vol. 51. Applications of Mathematics (New York). Stochastic Modelling and Applied Probability. Springer-Verlag, New York, 2003, pp. xii+438.