Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: The ability to model interactions among agents is crucial for effective
coordination and understanding their cooperation mechanisms in multi-agent
reinforcement learning (MARL). However, previous efforts to model high-order
interactions have been primarily hindered by the combinatorial explosion or the
opaque nature of their black-box network structures. In this paper, we propose
a novel value decomposition framework, called Continued Fraction Q-Learning
(QCoFr), which can flexibly capture arbitrary-order agent interactions with
only linear complexity $\mathcal{O}\left({n}\right)$ in the number of agents,
thus avoiding the combinatorial explosion when modeling rich cooperation.
Furthermore, we introduce the variational information bottleneck to extract
latent information for estimating credits. This latent information helps agents
filter out noisy interactions, thereby significantly enhancing both cooperation
and interpretability. Extensive experiments demonstrate that QCoFr not only
consistently achieves better performance but also provides interpretability
that aligns with our theoretical analysis.
Authors (4)
Qinyu Xu
Yuanyang Zhu
Xuefei Wu
Chunlin Chen
Submitted
October 23, 2025
Key Contributions
Proposes QCoFr, a novel value decomposition framework for MARL that models arbitrary-order agent interactions with linear complexity, avoiding combinatorial explosion. It uses a variational information bottleneck for interpretable credit assignment, enhancing cooperation.
Business Value
Enables the development of more coordinated and understandable multi-agent systems, crucial for applications like autonomous vehicle fleets, robotic teams, and complex simulation environments. Improved interpretability aids debugging and trust.