English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
 
 
DownloadE-Mail
  Probing Compositional Inference in Natural and Artificial Agents

Jagadish, A., Saanum, T., Wang, J., Binz, M., & Schulz, E. (2022). Probing Compositional Inference in Natural and Artificial Agents. In 5th Multidisciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2022) (pp. 275-279).

Item is

Basic

show hide
Genre: Conference Paper

Files

show Files

Locators

show
hide
Description:
-
OA-Status:
Description:
-
OA-Status:

Creators

show
hide
 Creators:
Jagadish, AK1, Author           
Saanum, T1, Author           
Wang, JX, Author
Binz, M1, Author           
Schulz, E1, Author           
Affiliations:
1Research Group Computational Principles of Intelligence, Max Planck Institute for Biological Cybernetics, Max Planck Society, ou_3189356              

Content

show
hide
Free keywords: -
 Abstract: People can easily evoke previously encountered concepts, compose them, and apply the result to novel contexts in
a zero-shot manner. What computational mechanisms underpin this ability? To study this question, we propose an
extension to the structured multi-armed bandit paradigm, which has been used to probe human function learning in
previous works. This new paradigm involves a learning curriculum where agents first perform two sub-tasks in which
rewards were sampled from differently structured reward functions, followed by a third sub-task in which rewards
were set to a composition of the previously encountered reward functions. This setup allows us to investigate how
people reason compositionally over learned functions, while still being simple enough to be tractable. Human behavior
in such tasks has been predominantly modeled by computational models with hard-coded structures such as Bayesian
grammars. We indeed find that such a model performs well on our task. However, they do not explain how people learn
to compose reward functions via trial and error but have, instead, been hand-designed to generalize compositionally
by expert researchers. How could the ability to compose ever emerge through trial and error? We propose a model
based on the principle of meta-learning to tackle this challenge and find that – upon training on the previously described
curriculum – meta-learned agents exhibit characteristics comparable to those of a Bayesian agent with compositional
priors. Model simulations suggest that both models can compose earlier learned functions to generalize in a zero-shot
manner. We complemented these model simulations results with a behavioral study, in which we investigated how
human participants approach our task. We find that they are indeed able to perform zero-shot compositional reasoning
as predicted by our models. Taken together, our study paves a way for studying compositional reinforcement learning
in humans, symbolic, and sub-symbolic agents.

Details

show
hide
Language(s):
 Dates: 2022-032022-06
 Publication Status: Published online
 Pages: -
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: -
 Degree: -

Event

show
hide
Title: 5th Multidisciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2022)
Place of Event: Providence, RI, USA
Start-/End Date: 2022-06-08 - 2022-06-11

Legal Case

show

Project information

show

Source 1

show
hide
Title: 5th Multidisciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2022)
Source Genre: Proceedings
 Creator(s):
Affiliations:
Publ. Info: -
Pages: - Volume / Issue: - Sequence Number: 1.67 Start / End Page: 275 - 279 Identifier: -