日本語
 
Help Privacy Policy ポリシー/免責事項
  詳細検索ブラウズ

アイテム詳細


公開

会議論文

Feudal Multi-Agent Hierarchies for Cooperative Reinforcement Learning

MPS-Authors
/persons/resource/persons217460

Dayan,  P
Department of Computational Neuroscience, Max Planck Institute for Biological Cybernetics, Max Planck Society;
Max Planck Institute for Biological Cybernetics, Max Planck Society;

External Resource
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
フルテキスト (公開)
公開されているフルテキストはありません
付随資料 (公開)
There is no public supplementary material available
引用

Ahilan, S., & Dayan, P. (2019). Feudal Multi-Agent Hierarchies for Cooperative Reinforcement Learning. In Annual Conference of the American Library Association (ALA 2019) (pp. 1-5).


引用: https://hdl.handle.net/21.11116/0000-0005-4B86-0
要旨
We investigate how reinforcement learning agents can learn tocooperate. Drawing inspiration from human societies, in whichsuccessful coordination of many individuals is often facilitated byhierarchical organisation, we introduce Feudal Multi-agent Hierar-chies (FMH). In this framework, a ‘manager’ agent, which is taskedwith maximising the environmentally-determined reward func-tion, learns to communicate subgoals to multiple, simultaneously-operating, ‘worker’ agents. Workers, which are rewarded for achiev-ing managerial subgoals, take concurrent actions in the world. Weoutline the structure of FMH and demonstrate its potential for de-centralised learning and control. We find that, given an adequate setof subgoals from which to choose, FMH performs, and particularlyscales, substantially better than cooperative approaches that use ashared reward function.