Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: This paper introduces a novel approach that integrates graph theory into
self-supervised representation learning. Traditional methods focus on
intra-instance variations generated by applying augmentations. However, they
often overlook important inter-instance relationships. While our method retains
the intra-instance property, it further captures inter-instance relationships
by constructing k-nearest neighbor (KNN) graphs for both teacher and student
streams during pretraining. In these graphs, nodes represent samples along with
their latent representations. Edges encode the similarity between instances.
Following pretraining, a representation refinement phase is performed. In this
phase, Graph Neural Networks (GNNs) propagate messages not only among immediate
neighbors but also across multiple hops, thereby enabling broader contextual
integration. Experimental results on CIFAR-10, ImageNet-100, and ImageNet-1K
demonstrate accuracy improvements of 7.3%, 3.2%, and 1.0%, respectively, over
state-of-the-art methods. These results highlight the effectiveness of the
proposed graph based mechanism. The code is publicly available at
https://github.com/alijavidani/SSL-GraphNNCLR.
Authors (3)
Ali Javidani
Babak Nadjar Araabi
Mohammad Amin Sadeghi
Submitted
October 25, 2025
IEEE Signal Processing Letters, vol. 32, pp. 3730-3734, 2025
Key Contributions
This paper introduces a novel self-supervised representation learning approach that leverages inter-instance relationships by constructing k-nearest neighbor (KNN) graphs. By using Graph Neural Networks (GNNs) for message passing across these graphs, it enables richer contextual integration beyond traditional intra-instance augmentations, leading to improved representation quality and downstream task accuracy.
Business Value
Enhances the performance of machine learning models by learning richer, more context-aware representations from unlabeled data. This can lead to more accurate AI systems in various applications without the need for extensive labeled datasets.