Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 96% Match Research Paper GNN Researchers,ML Researchers,AI Safety Researchers,Data Scientists 2 weeks ago

Quantifying Distributional Invariance in Causal Subgraph for IRM-Free Graph Generalization

graph-neural-networks › graph-learning
📄 Abstract

Abstract: Out-of-distribution generalization under distributional shifts remains a critical challenge for graph neural networks. Existing methods generally adopt the Invariant Risk Minimization (IRM) framework, requiring costly environment annotations or heuristically generated synthetic splits. To circumvent these limitations, in this work, we aim to develop an IRM-free method for capturing causal subgraphs. We first identify that causal subgraphs exhibit substantially smaller distributional variations than non-causal components across diverse environments, which we formalize as the Invariant Distribution Criterion and theoretically prove in this paper. Building on this criterion, we systematically uncover the quantitative relationship between distributional shift and representation norm for identifying the causal subgraph, and investigate its underlying mechanisms in depth. Finally, we propose an IRM-free method by introducing a norm-guided invariant distribution objective for causal subgraph discovery and prediction. Extensive experiments on two widely used benchmarks demonstrate that our method consistently outperforms state-of-the-art methods in graph generalization.
Authors (6)
Yang Qiu
Yixiong Zou
Jun Wang
Wei Liu
Xiangyu Fu
Ruixuan Li
Submitted
October 23, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper proposes an IRM-free method for achieving out-of-distribution generalization in Graph Neural Networks by identifying causal subgraphs. It introduces the 'Invariant Distribution Criterion', showing that causal subgraphs exhibit smaller distributional variations, and leverages this by proposing a norm-guided invariant distribution objective to capture these causal structures without the need for costly environment annotations.

Business Value

Improves the reliability of GNNs in real-world scenarios where data distributions can shift, leading to more trustworthy AI systems in areas like drug discovery or social network analysis.