We introduce FeUdal Networks (FuNs): a novel architecture for hierarchical
reinforcement learning. Our approach is inspired by the feudal reinforcement
learning proposal of Dayan and Hinton, and gains power and efficacy by
decoupling end-to-end learning across multiple levels -- allowing it to utilise
different resolutions of time. Our framework employs a Manager module and a
Worker module. The Manager operates at a lower temporal resolution and sets
abstract goals which are conveyed to and enacted by the Worker. The Worker
generates primitive actions at every tick of the environment. The decoupled
structure of FuN conveys several benefits -- in addition to facilitating very
long timescale credit assignment it also encourages the emergence of
sub-policies associated with different goals set by the Manager. These
properties allow FuN to dramatically outperform a strong baseline agent on
tasks that involve long-term credit assignment or memorisation. We demonstrate
the performance of our proposed system on a range of tasks from the ATARI suite
and also from a 3D DeepMind Lab environment.
Description
[1703.01161] FeUdal Networks for Hierarchical Reinforcement Learning
%0 Journal Article
%1 vezhnevets2017feudal
%A Vezhnevets, Alexander Sasha
%A Osindero, Simon
%A Schaul, Tom
%A Heess, Nicolas
%A Jaderberg, Max
%A Silver, David
%A Kavukcuoglu, Koray
%D 2017
%K hierarchical reinforcement-learning
%T FeUdal Networks for Hierarchical Reinforcement Learning
%U http://arxiv.org/abs/1703.01161
%X We introduce FeUdal Networks (FuNs): a novel architecture for hierarchical
reinforcement learning. Our approach is inspired by the feudal reinforcement
learning proposal of Dayan and Hinton, and gains power and efficacy by
decoupling end-to-end learning across multiple levels -- allowing it to utilise
different resolutions of time. Our framework employs a Manager module and a
Worker module. The Manager operates at a lower temporal resolution and sets
abstract goals which are conveyed to and enacted by the Worker. The Worker
generates primitive actions at every tick of the environment. The decoupled
structure of FuN conveys several benefits -- in addition to facilitating very
long timescale credit assignment it also encourages the emergence of
sub-policies associated with different goals set by the Manager. These
properties allow FuN to dramatically outperform a strong baseline agent on
tasks that involve long-term credit assignment or memorisation. We demonstrate
the performance of our proposed system on a range of tasks from the ATARI suite
and also from a 3D DeepMind Lab environment.
@article{vezhnevets2017feudal,
abstract = {We introduce FeUdal Networks (FuNs): a novel architecture for hierarchical
reinforcement learning. Our approach is inspired by the feudal reinforcement
learning proposal of Dayan and Hinton, and gains power and efficacy by
decoupling end-to-end learning across multiple levels -- allowing it to utilise
different resolutions of time. Our framework employs a Manager module and a
Worker module. The Manager operates at a lower temporal resolution and sets
abstract goals which are conveyed to and enacted by the Worker. The Worker
generates primitive actions at every tick of the environment. The decoupled
structure of FuN conveys several benefits -- in addition to facilitating very
long timescale credit assignment it also encourages the emergence of
sub-policies associated with different goals set by the Manager. These
properties allow FuN to dramatically outperform a strong baseline agent on
tasks that involve long-term credit assignment or memorisation. We demonstrate
the performance of our proposed system on a range of tasks from the ATARI suite
and also from a 3D DeepMind Lab environment.},
added-at = {2019-04-23T14:18:23.000+0200},
author = {Vezhnevets, Alexander Sasha and Osindero, Simon and Schaul, Tom and Heess, Nicolas and Jaderberg, Max and Silver, David and Kavukcuoglu, Koray},
biburl = {https://www.bibsonomy.org/bibtex/25488b15003c09188e885d8022cd67811/kirk86},
description = {[1703.01161] FeUdal Networks for Hierarchical Reinforcement Learning},
interhash = {78949828baa884e70ab49a6689b70fae},
intrahash = {5488b15003c09188e885d8022cd67811},
keywords = {hierarchical reinforcement-learning},
note = {cite arxiv:1703.01161},
timestamp = {2019-04-23T14:18:23.000+0200},
title = {FeUdal Networks for Hierarchical Reinforcement Learning},
url = {http://arxiv.org/abs/1703.01161},
year = 2017
}