Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 95% Match Research Paper AI Researchers,Machine Learning Engineers,Graph Data Scientists,Network Analysts 2 weeks ago

Unifying and Enhancing Graph Transformers via a Hierarchical Mask Framework

graph-neural-networks › graph-learning
📄 Abstract

Abstract: Graph Transformers (GTs) have emerged as a powerful paradigm for graph representation learning due to their ability to model diverse node interactions. However, existing GTs often rely on intricate architectural designs tailored to specific interactions, limiting their flexibility. To address this, we propose a unified hierarchical mask framework that reveals an underlying equivalence between model architecture and attention mask construction. This framework enables a consistent modeling paradigm by capturing diverse interactions through carefully designed attention masks. Theoretical analysis under this framework demonstrates that the probability of correct classification positively correlates with the receptive field size and label consistency, leading to a fundamental design principle: an effective attention mask should ensure both a sufficiently large receptive field and a high level of label consistency. While no single existing mask satisfies this principle across all scenarios, our analysis reveals that hierarchical masks offer complementary strengths, motivating their effective integration. Then, we introduce M3Dphormer, a Mixture-of-Experts-based Graph Transformer with Multi-Level Masking and Dual Attention Computation. M3Dphormer incorporates three theoretically grounded hierarchical masks and employs a bi-level expert routing mechanism to adaptively integrate multi-level interaction information. To ensure scalability, we further introduce a dual attention computation scheme that dynamically switches between dense and sparse modes based on local mask sparsity. Extensive experiments across multiple benchmarks demonstrate that M3Dphormer achieves state-of-the-art performance, validating the effectiveness of our unified framework and model design.
Authors (5)
Yujie Xing
Xiao Wang
Bin Wu
Hai Huang
Chuan Shi
Submitted
October 21, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

This paper introduces a unified hierarchical mask framework for Graph Transformers (GTs) that reveals an equivalence between architecture and attention mask construction. This framework enables consistent modeling of diverse interactions and establishes a design principle: effective attention masks require large receptive fields and high label consistency, leading to enhanced flexibility and performance.

Business Value

Enables more flexible and powerful graph-based AI models, leading to improved performance in areas like social network analysis, drug discovery, and recommendation systems.