Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Credit assignmen, disentangling each agent's contribution to a shared reward,
is a critical challenge in cooperative multi-agent reinforcement learning
(MARL). To be effective, credit assignment methods must preserve the
environment's optimal policy. Some recent approaches attempt this by enforcing
return equivalence, where the sum of distributed rewards must equal the team
reward. However, their guarantees are conditional on a learned model's
regression accuracy, making them unreliable in practice. We introduce
Temporal-Agent Reward Redistribution (TAR$^2$), an approach that decouples
credit modeling from this constraint. A neural network learns unnormalized
contribution scores, while a separate, deterministic normalization step
enforces return equivalence by construction. We demonstrate that this method is
equivalent to a valid Potential-Based Reward Shaping (PBRS), which guarantees
the optimal policy is preserved regardless of model accuracy. Empirically, on
challenging SMACLite and Google Research Football (GRF) benchmarks, TAR$^2$
accelerates learning and achieves higher final performance than strong
baselines. These results establish our method as an effective solution for the
agent-temporal credit assignment problem.
Authors (7)
Aditya Kapoor
Kale-ab Tessera
Mayank Baranwal
Harshad Khadilkar
Jan Peters
Stefano Albrecht
+1 more
Submitted
February 7, 2025
Key Contributions
Introduces Temporal-Agent Reward Redistribution (TAR^2), a novel MARL credit assignment method that decouples credit modeling from return equivalence constraints by using a deterministic normalization step, guaranteeing optimal policy preservation regardless of model accuracy.
Business Value
Enables more effective coordination and learning in multi-agent systems, crucial for applications like autonomous vehicle fleets, drone swarms, and complex robotic task execution.