Intrinsically Motivated Hierarchical Skill Learning in Structured Environments

作者: Christopher M Vigorito , Andrew G Barto

DOI: 10.1109/TAMD.2010.2050205

关键词:

摘要: We present a framework for intrinsically motivated developmental learning of abstract skill hierarchies by reinforcement agents in structured environments. Long-term can drastically improve an agent's efficiency solving ensembles related tasks complex domain. In domains composed many features, understanding the causal relationships between actions and their effects on different features environment greatly facilitate learning. Using Bayesian network structure (learning techniques dynamic programming algorithms), we show that learn incrementally autonomously both hierarchy skills exploit this structure. Furthermore, novel active scheme employs intrinsic motivation to maximize with which is learned. As new acquired using current set skills, more are learned, turn allow agent discover structure, so on. This bootstrapping property makes our approach process results steadily increasing domain knowledge behavioral complexity as continues explore its environment.

参考文章(3)
L. G. Valiant, A theory of the learnable symposium on the theory of computing. ,vol. 27, pp. 1134- 1142 ,(1984) , 10.1145/800057.808710
Richard S. Sutton, Doina Precup, Satinder Singh, Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning Artificial Intelligence. ,vol. 112, pp. 181- 211 ,(1999) , 10.1016/S0004-3702(99)00052-1
A.G. Barto, R.S. Sutton, Reinforcement Learning: An Introduction ,(1988)