Deep Spatial Autoencoders for Visuomotor Learning

作者: Pieter Abbeel , Chelsea Finn , Yan Duan , Sergey Levine , Trevor Darrell

DOI:

关键词: Control theoryRepresentation (mathematics)RobotReinforcement learningComputer visionTask (project management)AutoencoderArtificial intelligenceFeature (computer vision)Robot learningComputer science

摘要: Reinforcement learning provides a powerful and flexible framework for automated acquisition of robotic motion skills. However, applying reinforcement requires sufficiently detailed representation the state, including configuration task-relevant objects. We present an approach that automates state-space construction by state directly from camera images. Our method uses deep spatial autoencoder to acquire set feature points describe environment current task, such as positions objects, then learns skill with these using efficient based on local linear models. The resulting controller reacts continuously learned points, allowing robot dynamically manipulate objects in world closed-loop control. demonstrate our PR2 tasks include pushing free-standing toy block, picking up bag rice spatula, hanging loop rope hook at various positions. In each automatically track their robot's arm.

参考文章(30)
Emanuel Todorov, Weiwei Li, Iterative Linear Quadratic Regulator Design for Nonlinear Biological Movement Systems Iterative Linear Quadratic Regulator Design for Nonlinear Biological Movement Systems. pp. 222- 229 ,(2004)
Katharina Mülling, Jan Peters, Yasemin Altün, Relative entropy policy search national conference on artificial intelligence. pp. 1607- 1612 ,(2010)
Marc Deisenroth, Carl Rasmussen, Dieter Fox, Learning to Control a Low-Cost Manipulator using Data-Efficient Reinforcement Learning robotics science and systems. ,vol. 07, ,(2011) , 10.15607/RSS.2011.VII.008
J. Andrew Bagnell, Jeff Schneider, Covariant policy search international joint conference on artificial intelligence. pp. 1019- 1024 ,(2003) , 10.1184/R1/6552458.V1
Leemon Baird, Residual Algorithms: Reinforcement Learning with Function Approximation Machine Learning Proceedings 1995. pp. 30- 37 ,(1995) , 10.1016/B978-1-55860-377-6.50013-X
Volodymyr Mnih, Ioannis Antonoglou, Koray Kavukcuoglu, Daan Wierstra, Martin A. Riedmiller, Alex Graves, David Silver, Playing Atari with Deep Reinforcement Learning arXiv: Learning. ,(2013)
Sascha Lange, Martin Riedmiller, Arne Voigtlander, Autonomous reinforcement learning on raw visual input data in a real world application international joint conference on neural network. pp. 1- 8 ,(2012) , 10.1109/IJCNN.2012.6252823
Thomas Lampe, Martin Riedmiller, Acquiring visual servoing reaching and grasping skills using neural reinforcement learning international joint conference on neural network. pp. 1- 8 ,(2013) , 10.1109/IJCNN.2013.6707053
Gerhard Neumann, Marc Peter Deisenroth, Jan Peters, A Survey on Policy Search for Robotics ,(2013)
Gen Endo, Jun Morimoto, Takamitsu Matsubara, Jun Nakanishi, Gordon Cheng, Learning CPG-based Biped Locomotion with a Policy Gradient Method: Application to a Humanoid Robot The International Journal of Robotics Research. ,vol. 27, pp. 213- 228 ,(2008) , 10.1177/0278364907084980