Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Graph Neural Networks (GNNs) have become a pivotal framework for modeling
graph-structured data, enabling a wide range of applications from social
network analysis to molecular chemistry. By integrating large language models
(LLMs), text-attributed graphs (TAGs) enhance node representations with rich
textual semantics, significantly boosting the expressive power of graph-based
learning. However, this sophisticated synergy introduces critical
vulnerabilities, as Graph-LLMs are susceptible to adversarial attacks on both
their structural topology and textual attributes. Although specialized attack
methods have been designed for each of these aspects, no work has yet unified
them into a comprehensive approach. In this work, we propose the Interpretable
Multi-Dimensional Graph Attack (IMDGA), a novel human-centric adversarial
attack framework designed to orchestrate multi-level perturbations across both
graph structure and textual features. IMDGA utilizes three tightly integrated
modules to craft attacks that balance interpretability and impact, enabling a
deeper understanding of Graph-LLM vulnerabilities. Through rigorous theoretical
analysis and comprehensive empirical evaluations on diverse datasets and
architectures, IMDGA demonstrates superior interpretability, attack
effectiveness, stealthiness, and robustness compared to existing methods. By
exposing critical weaknesses in TAG representation learning, this work uncovers
a previously underexplored semantic dimension of vulnerability in Graph-LLMs,
offering valuable insights for improving their resilience. Our code and
resources are publicly available at
https://anonymous.4open.science/r/IMDGA-7289.
Authors (8)
Bowen Fan
Zhilin Guo
Xunkai Li
Yihan Zhou
Bing Zhou
Zhenjun Li
+2 more
Submitted
October 14, 2025
Key Contributions
Proposes the Interpretable Multi-Dimensional Graph Attack (IMDGA), a novel human-centric adversarial attack framework that unifies and orchestrates multi-level perturbations across both graph structure and textual features in Graph-LLMs. This addresses the lack of comprehensive attack methods that consider both aspects simultaneously.
Business Value
Understanding and mitigating vulnerabilities in Graph-LLMs is crucial for deploying these powerful models safely in sensitive applications like fraud detection or cybersecurity.