Abstract
AbstractWe consider mean-field control problems in discrete time with discounted reward, infinite time horizon and compact state and action space. The existence of optimal policies is shown and the limiting mean-field problem is derived when the number of individuals tends to infinity. Moreover, we consider the average reward problem and show that the optimal policy in this mean-field limit is $$\varepsilon $$
ε
-optimal for the discounted problem if the number of individuals is large and the discount factor close to one. This result is very helpful, because it turns out that in the special case when the reward does only depend on the distribution of the individuals, we obtain a very interesting subclass of problems where an average reward optimal policy can be obtained by first computing an optimal measure from a static optimization problem and then achieving it with Markov Chain Monte Carlo methods. We give two applications: Avoiding congestion an a graph and optimal positioning on a market place which we solve explicitly.
Funder
Karlsruher Institut für Technologie (KIT)
Publisher
Springer Science and Business Media LLC
Subject
Applied Mathematics,Control and Optimization
Reference36 articles.
1. Bäuerle, N., Lange, D.: Optimal control of partially observable piecewise deterministic Markov processes. SIAM J. Control Optim. 56(2), 1441–1462 (2018)
2. Bäuerle, N., Rieder, U.: Markov Decision Processes with Applications to Finance. Springer-Verlag, Berlin Heidelberg (2011)
3. Bäuerle, N.: Convex stochastic fluid programs with average cost. J. Math. Anal. Appl. 259(1), 137–156 (2001)
4. Bertsekas, D.P., Tsitsiklis, J.N.: Neuro-dynamic Programming. Athena Scientific, Belmont, Mass (1996)
5. Biswas, A.: Mean field games with ergodic cost for discrete time Markov processes. arXiv preprint arXiv:1510.08968 (2015)
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献