Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper Researchers in GNNs and HDC,ML engineers focused on efficiency,Hardware designers for AI accelerators 1 week ago

HyperGraphX: Graph Transductive Learning with Hyperdimensional Computing and Message Passing

graph-neural-networks › graph-learning
📄 Abstract

Abstract: We present a novel algorithm, \hdgc, that marries graph convolution with binding and bundling operations in hyperdimensional computing for transductive graph learning. For prediction accuracy \hdgc outperforms major and popular graph neural network implementations as well as state-of-the-art hyperdimensional computing implementations for a collection of homophilic graphs and heterophilic graphs. Compared with the most accurate learning methodologies we have tested, on the same target GPU platform, \hdgc is on average 9561.0 and 144.5 times faster than \gcnii, a graph neural network implementation and HDGL, a hyperdimensional computing implementation, respectively. As the majority of the learning operates on binary vectors, we expect outstanding energy performance of \hdgc on neuromorphic and emerging process-in-memory devices.
Authors (4)
Guojing Cong
Tom Potok
Hamed Poursiami
Maryam Parsa
Submitted
October 28, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

Introduces \hdgc, a novel algorithm combining graph convolution with hyperdimensional computing (HDC) operations (binding and bundling) for transductive graph learning. This approach significantly outperforms GNNs and HDC methods in prediction accuracy and achieves dramatic speedups (9561x and 144.5x faster than GCNII and HDGL respectively), with expected outstanding energy performance on neuromorphic hardware due to binary vector operations.

Business Value

Enables faster and more energy-efficient analysis of large-scale graph data, unlocking new possibilities for real-time graph-based AI applications and reducing the cost of AI computation.