Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 98% Match Research Paper RL Researchers,AI Engineers,Robotics Engineers,Game Developers 1 week ago

Redistributing Rewards Across Time and Agents for Multi-Agent Reinforcement Learning

reinforcement-learning › multi-agent
📄 Abstract

Abstract: Credit assignmen, disentangling each agent's contribution to a shared reward, is a critical challenge in cooperative multi-agent reinforcement learning (MARL). To be effective, credit assignment methods must preserve the environment's optimal policy. Some recent approaches attempt this by enforcing return equivalence, where the sum of distributed rewards must equal the team reward. However, their guarantees are conditional on a learned model's regression accuracy, making them unreliable in practice. We introduce Temporal-Agent Reward Redistribution (TAR$^2$), an approach that decouples credit modeling from this constraint. A neural network learns unnormalized contribution scores, while a separate, deterministic normalization step enforces return equivalence by construction. We demonstrate that this method is equivalent to a valid Potential-Based Reward Shaping (PBRS), which guarantees the optimal policy is preserved regardless of model accuracy. Empirically, on challenging SMACLite and Google Research Football (GRF) benchmarks, TAR$^2$ accelerates learning and achieves higher final performance than strong baselines. These results establish our method as an effective solution for the agent-temporal credit assignment problem.
Authors (7)
Aditya Kapoor
Kale-ab Tessera
Mayank Baranwal
Harshad Khadilkar
Jan Peters
Stefano Albrecht
+1 more
Submitted
February 7, 2025
arXiv Category
cs.MA
arXiv PDF

Key Contributions

Introduces Temporal-Agent Reward Redistribution (TAR^2), a novel MARL credit assignment method that decouples credit modeling from return equivalence constraints by using a deterministic normalization step, guaranteeing optimal policy preservation regardless of model accuracy.

Business Value

Enables more effective coordination and learning in multi-agent systems, crucial for applications like autonomous vehicle fleets, drone swarms, and complex robotic task execution.