Author:
Shah Kunal,Laxkar Pradeep,Chakrabarti Prasun
Reference6 articles.
1. Bostrom, N.: Superintelligence: Paths, Dangers, Strategies. Oxford University Press, Sept 3rd 2014. Accessed 9 Dec 2017
2. Anderson, S.L.: Asimov’s “three laws of robotics” and machine metaethics. AI & Soc. 22(4), 477–493 (2008)
3. Gouberman A., Siegle M.: Markov reward models and markov decision processes in discrete and continuous time: performance evaluation and optimization. In: Remke, A., Stoelinga, M. (eds.) Stochastic Model Checking. Rigorous Dependability Analysis Using Model Checking Techniques for Stochastic Systems. ROCKS 2012. Lecture Notes in Computer Science, vol. 8453. Springer, Berlin, Heidelberg (2014)
4. Bradtke, S.J., Duff, M.O.: Reinforcement learning methods for continuous-time Markov decision problems. In Advances in Neural Information Processing Systems, pp. 393–400 (1995)
5. Baird, L.: Residual algorithms: Reinforcement learning with function approximation. In: Machine Learning Proceedings, pp. 30–37 (1995)
Cited by
11 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献