Autonomous Learning Conference Paper 2024

SENSEI: Semantic Exploration Guided by Foundation Models to Learn Versatile World Models

Exploring useful behavior is a keystone of reinforcement learning (RL). Existing approaches to intrinsic motivation, following general principles such as information gain, mostly uncover low-level interactions. In contrast, children’s play suggests that they engage in semantically meaningful high-level behavior by imitating or interacting with their caregivers. Recent work has focused on using foundation models to inject these semantic biases into exploration. However, these methods often rely on unrealistic assumptions, such as environments already embedded in language or access to high-level actions. To bridge this gap, we propose SEmaNtically Sensible ExploratIon (Sensei), a framework to equip model-based RL agents with intrinsic motivation for semantically meaningful behavior. To do so, we distill an intrinsic reward signal of interestingness from Vision Language Model (VLM) annotations. The agent learns to predict and maximize these intrinsic rewards using a world model learned directly from intrinsic rewards, image observations, and low-level actions. We show that in both robotic and video game-like simulations Sensei manages to discover a variety of meaningful behaviors. We believe Sensei provides a general tool for integrating feedback from foundation models into autonomous agents, a crucial research direction as openly available VLMs become more powerful.

Author(s): Sancaktar, Cansu and Gumbsch, Christian and Zadaianchuk, Andrii and Kolev, Pavel and Martius, Georg
Book Title: The Training Agents with Foundation Models Workshop at RLC
Year: 2024
Project(s):
Bibtex Type: Conference Paper (inproceedings)
State: Published
URL: https://openreview.net/forum?id=dHNVY5qMiP
Note: indicates equal contribution

BibTex

@inproceedings{sancaktargumbsch2024:sensei,
  title = {SENSEI: Semantic Exploration Guided by Foundation Models to Learn Versatile World Models},
  booktitle = {The Training Agents with Foundation Models Workshop at RLC},
  abstract = {Exploring useful behavior is a keystone of reinforcement learning (RL). Existing approaches to intrinsic motivation, following general principles such as information gain, mostly uncover low-level interactions. In contrast, children’s play suggests that they engage in semantically meaningful high-level behavior by imitating or interacting with their caregivers. Recent work has focused on using foundation models to inject these semantic biases into exploration. However, these methods often rely on unrealistic assumptions, such as environments already embedded in language or access to high-level actions. To bridge this gap, we propose SEmaNtically Sensible ExploratIon (Sensei), a framework to equip model-based RL agents with intrinsic motivation for semantically meaningful behavior. To do so, we distill an intrinsic reward signal of interestingness from Vision Language Model (VLM) annotations. The agent learns to predict and maximize these intrinsic rewards using a world model learned directly from intrinsic rewards, image observations, and low-level actions. We show that in both robotic and video game-like simulations Sensei manages to discover a variety of meaningful behaviors. We believe Sensei provides a general tool for integrating feedback from foundation models into autonomous agents, a crucial research direction as openly available VLMs become more powerful.},
  year = {2024},
  note = {indicates equal contribution},
  slug = {sancaktargumbsch2024-sensei},
  author = {Sancaktar, Cansu and Gumbsch, Christian and Zadaianchuk, Andrii and Kolev, Pavel and Martius, Georg},
  url = {https://openreview.net/forum?id=dHNVY5qMiP}
}