Through multi-agent competition, the simple objective of hide-and-seek, and
standard reinforcement learning algorithms at scale, we find that agents create
a self-supervised autocurriculum inducing multiple distinct rounds of emergent
strategy, many of which require sophisticated tool use and coordination. We
find clear evidence of six emergent phases in agent strategy in our
environment, each of which creates a new pressure for the opposing team to
adapt; for instance, agents learn to build multi-object shelters using moveable
boxes which in turn leads to agents discovering that they can overcome
obstacles using ramps. We further provide evidence that multi-agent competition
may scale better with increasing environment complexity and leads to behavior
that centers around far more human-relevant skills than other self-supervised
reinforcement learning methods such as intrinsic motivation. Finally, we
propose transfer and fine-tuning as a way to quantitatively evaluate targeted
capabilities, and we compare hide-and-seek agents to both intrinsic motivation
and random initialization baselines in a suite of domain-specific intelligence
tests.
Description
[1909.07528] Emergent Tool Use From Multi-Agent Autocurricula
%0 Generic
%1 baker2019HideAndSeek
%A Baker, Bowen
%A Kanitscheider, Ingmar
%A Markov, Todor
%A Wu, Yi
%A Powell, Glenn
%A McGrew, Bob
%A Mordatch, Igor
%D 2019
%K DRLAlgoComparison PPO reinforcement_learning
%T Emergent Tool Use From Multi-Agent Autocurricula
%U http://arxiv.org/abs/1909.07528
%X Through multi-agent competition, the simple objective of hide-and-seek, and
standard reinforcement learning algorithms at scale, we find that agents create
a self-supervised autocurriculum inducing multiple distinct rounds of emergent
strategy, many of which require sophisticated tool use and coordination. We
find clear evidence of six emergent phases in agent strategy in our
environment, each of which creates a new pressure for the opposing team to
adapt; for instance, agents learn to build multi-object shelters using moveable
boxes which in turn leads to agents discovering that they can overcome
obstacles using ramps. We further provide evidence that multi-agent competition
may scale better with increasing environment complexity and leads to behavior
that centers around far more human-relevant skills than other self-supervised
reinforcement learning methods such as intrinsic motivation. Finally, we
propose transfer and fine-tuning as a way to quantitatively evaluate targeted
capabilities, and we compare hide-and-seek agents to both intrinsic motivation
and random initialization baselines in a suite of domain-specific intelligence
tests.
@misc{baker2019HideAndSeek,
abstract = {Through multi-agent competition, the simple objective of hide-and-seek, and
standard reinforcement learning algorithms at scale, we find that agents create
a self-supervised autocurriculum inducing multiple distinct rounds of emergent
strategy, many of which require sophisticated tool use and coordination. We
find clear evidence of six emergent phases in agent strategy in our
environment, each of which creates a new pressure for the opposing team to
adapt; for instance, agents learn to build multi-object shelters using moveable
boxes which in turn leads to agents discovering that they can overcome
obstacles using ramps. We further provide evidence that multi-agent competition
may scale better with increasing environment complexity and leads to behavior
that centers around far more human-relevant skills than other self-supervised
reinforcement learning methods such as intrinsic motivation. Finally, we
propose transfer and fine-tuning as a way to quantitatively evaluate targeted
capabilities, and we compare hide-and-seek agents to both intrinsic motivation
and random initialization baselines in a suite of domain-specific intelligence
tests.},
added-at = {2020-01-24T08:37:51.000+0100},
author = {Baker, Bowen and Kanitscheider, Ingmar and Markov, Todor and Wu, Yi and Powell, Glenn and McGrew, Bob and Mordatch, Igor},
biburl = {https://www.bibsonomy.org/bibtex/2bb3e531318e3b5cc98b53fb3c81152db/lanteunis},
description = {[1909.07528] Emergent Tool Use From Multi-Agent Autocurricula},
interhash = {58c402ce569035c1505d70cd284d6c35},
intrahash = {bb3e531318e3b5cc98b53fb3c81152db},
keywords = {DRLAlgoComparison PPO reinforcement_learning},
note = {cite arxiv:1909.07528},
timestamp = {2020-01-24T08:37:51.000+0100},
title = {Emergent Tool Use From Multi-Agent Autocurricula},
url = {http://arxiv.org/abs/1909.07528},
year = 2019
}