Multi-Agent Imitation Learning for Driving Simulation

作者: Raunak P. Bhattacharyya , Derek J. Phillips , Blake Wulfe , Jeremy Morton , Alex Kuefler

DOI: 10.1109/IROS.2018.8593758

关键词:

摘要: Simulation is an appealing option for validating the safety of autonomous vehicles. Generative Adversarial Imitation Learning (GAIL) has recently been shown to learn representative human driver models. These models were learned through training in single-agent environments, but they have difficulty generalizing multi-agent driving scenarios. We argue these difficulties arise because observations at and test time are sampled from different distributions. This difference makes such unsuitable simulation scenes, where multiple agents must interact realistically over long horizons. extend GAIL address shortcomings a parameter-sharing approach grounded curriculum learning. Compared with policies, policies generated by our PS-GAIL method prove superior interacting stably setting capturing emergent behavior drivers.

参考文章(25)
Diederik P. Kingma, Jimmy Ba, Adam: A Method for Stochastic Optimization arXiv: Learning. ,(2014)
Michael L. Littman, Markov games as a framework for multi-agent reinforcement learning Machine Learning Proceedings 1994. pp. 157- 163 ,(1994) , 10.1016/B978-1-55860-335-6.50027-1
John Schulman, None, Trust Region Policy Optimization international conference on machine learning. pp. 1889- 1897 ,(2015)
Lucian Busoniu, Robert Babuska, Bart De Schutter, Multi-Agent Reinforcement Learning: A Survey international conference on control, automation, robotics and vision. pp. 1- 6 ,(2006) , 10.1109/ICARCV.2006.345353
Pieter Abbeel, Andrew Y. Ng, Apprenticeship learning via inverse reinforcement learning Twenty-first international conference on Machine learning - ICML '04. pp. 1- 8 ,(2004) , 10.1145/1015330.1015430
Hidetoshi Shimodaira, Improving predictive inference under covariate shift by weighting the log-likelihood function Journal of Statistical Planning and Inference. ,vol. 90, pp. 227- 244 ,(2000) , 10.1016/S0378-3758(00)00115-4
Liviu Panait, Sean Luke, Cooperative Multi-Agent Learning: The State of the Art Autonomous Agents and Multi-Agent Systems. ,vol. 11, pp. 387- 434 ,(2005) , 10.1007/S10458-005-2631-2
Reza Olfati-Saber, J. Alex Fax, Richard M. Murray, Consensus and Cooperation in Networked Multi-Agent Systems Proceedings of the IEEE. ,vol. 95, pp. 215- 233 ,(2007) , 10.1109/JPROC.2006.887293
A.K. Das, R. Fierro, V. Kumar, J.P. Ostrowski, J. Spletzer, C.J. Taylor, A vision-based formation control framework international conference on robotics and automation. ,vol. 18, pp. 813- 825 ,(2002) , 10.1109/TRA.2002.803463
Stefan Schaal, Is imitation learning the route to humanoid robots Trends in Cognitive Sciences. ,vol. 3, pp. 233- 242 ,(1999) , 10.1016/S1364-6613(99)01327-3