Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
π Abstract
Abstract: Humans leverage rich internal models of the world to reason about the future,
imagine counterfactuals, and adapt flexibly to new situations. In Reinforcement
Learning (RL), world models aim to capture how the environment evolves in
response to the agent's actions, facilitating planning and generalization.
However, typical world models directly operate on the environment variables
(e.g. pixels, physical attributes), which can make their training slow and
cumbersome; instead, it may be advantageous to rely on high-level latent
dimensions that capture relevant multimodal variables. Global Workspace (GW)
Theory offers a cognitive framework for multimodal integration and information
broadcasting in the brain, and recent studies have begun to introduce efficient
deep learning implementations of GW. Here, we evaluate the capabilities of an
RL system combining GW with a world model. We compare our GW-Dreamer with
various versions of the standard PPO and the original Dreamer algorithms. We
show that performing the dreaming process (i.e., mental simulation) inside the
GW latent space allows for training with fewer environment steps. As an
additional emergent property, the resulting model (but not its comparison
baselines) displays strong robustness to the absence of one of its observation
modalities (images or simulation attributes). We conclude that the combination
of GW with World Models holds great potential for improving decision-making in
RL agents.
Authors (3)
LΓ©opold MaytiΓ©
Roland Bertin Johannet
Rufin VanRullen
Submitted
February 28, 2025
Key Contributions
Introduces GW-Dreamer, an RL system combining Global Workspace (GW) Theory with a world model. This approach leverages high-level latent dimensions for multimodal integration and information broadcasting, aiming to improve planning, generalization, and the ability to imagine counterfactuals, outperforming standard PPO and the original Dreamer algorithm.
Business Value
Enables the development of more sophisticated RL agents capable of complex planning and adaptation, crucial for advanced robotics, autonomous vehicles, and intelligent decision-making systems.