1. Bach, F. and Levy, K.Y. (2019). A universal algorithm for variational inequalities adaptive to smoothness and noise. In A. Beygelzimer and D. Hsu (eds.), Proceedings of the Thirty-Second Conference on Learning Theory, volume 99 of Proceedings of Machine Learning Research, 164-194. PMLR, Phoenix, USA. URL http://proceedings.mlr.press/v99/bach19a.html. ArXiv:1902.01637.
2. Sample size selection in optimization methods for machine learning;Byrd;Mathematical Programming,2012
3. Deng, Q., Cheng, Y., and Lan, G. (2018). Optimal adaptive and accelerated stochastic gradient descent. arXiv:1810.00553.
4. First-order methods of smooth convex optimization with inexact oracle;Devolder;Mathematical Programming,2014
5. Adaptive subgradient methods for online learning and stochastic optimization;Duchi;Journal of Machine Learning Research,2011