Predictive Inverse Optimal Control for Linear-Quadratic-Gaussian Systems

作者: Brian D. Ziebart , Xiangli Chen

DOI:

关键词:

摘要: Predictive inverse optimal control is a powerful approach for estimating the policy of an agent from observed demonstrations. Its usefulness has been established in number large-scale sequential decision settings characterized by complete state observability. However, many real decisions are made situations where not fully known to making decisions. Though extensions predictive partially observable Markov processes have developed, their applicability limited complexities inference those representations. In this work, we extend linearquadratic-Gaussian setting. We establish close connections between laws setting and probabilistic predictions under our approach. demonstrate effectiveness benefit policies that influenced partial observability on both synthetic datasets.

参考文章(5)
Eyal Amir, Deepak Ramachandran, Bayesian inverse reinforcement learning international joint conference on artificial intelligence. ,vol. 51, pp. 2586- 2591 ,(2007)
Pieter Abbeel, Andrew Y. Ng, Apprenticeship learning via inverse reinforcement learning Twenty-first international conference on Machine learning - ICML '04. pp. 1- 8 ,(2004) , 10.1145/1015330.1015430
Andrew Y Ng, Stuart Russell, None, Algorithms for Inverse Reinforcement Learning international conference on machine learning. ,vol. 67, pp. 663- 670 ,(2000) , 10.2460/AJVR.67.2.323
Thomas M. Cover, Joy A. Thomas, Elements of information theory ,(1991)
Tamer Basar, A New Approach to Linear Filtering and Prediction Problems Journal of Basic Engineering. ,vol. 82, pp. 35- 45 ,(1960) , 10.1115/1.3662552