作者: Anil Anthony Bharath , Murray Shanahan , Kai Arulkumaran , Nat Dilokthanakul
DOI:
关键词:
摘要: In this paper we combine one method for hierarchical reinforcement learning - the options framework with deep Q-networks (DQNs) through use of different "option heads" on policy network, and a supervisory network choosing between options. We utilise our setup to investigate effects architectural constraints in subtasks positive negative transfer, across range capacities. empirically show that augmented DQN has lower sample complexity when simultaneously without degrading performance transfer.