Author:
Claeys Emmanuelle,Maumy-Bertrand Myriam,Gançarski Pierre
Publisher
Springer Nature Switzerland
Reference24 articles.
1. Agarwal, A., Dudík, M., Kale, S., Langford, J., Schapire, R.E.: Contextual bandit learning with predictable rewards. ArXiv e-prints, February 2012
2. Auer, P., Cesa-Bianchi, N., Fischer, P.: Finite-time analysis of the multiarmed bandit problem. Mach. Learn. 47(2), 235–256 (2002)
3. Avadhanula, V., Colini-Baldeschi, R., Leonardi, S., Sankararaman, K.A., Schrijvers, O.: Stochastic bandits for multi-platform budget optimization in online advertising. CoRR (2021)
4. Bastani, H., Bayati, M.: Online decision-making with high-dimensional covariates. SSRN Electron. J. (2015)
5. Bietti, A., Agarwal, A., Langford, J.: A contextual bandit bake-off. JLMR (2021)