Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 85% Match Research Paper Machine Learning Researchers,AI Safety Researchers,Neuroscience Researchers 20 hours ago

Contrastive Consolidation of Top-Down Modulations Achieves Sparsely Supervised Continual Learning

ai-safety › alignment
📄 Abstract

Abstract: Biological brains learn continually from a stream of unlabeled data, while integrating specialized information from sparsely labeled examples without compromising their ability to generalize. Meanwhile, machine learning methods are susceptible to catastrophic forgetting in this natural learning setting, as supervised specialist fine-tuning degrades performance on the original task. We introduce task-modulated contrastive learning (TMCL), which takes inspiration from the biophysical machinery in the neocortex, using predictive coding principles to integrate top-down information continually and without supervision. We follow the idea that these principles build a view-invariant representation space, and that this can be implemented using a contrastive loss. Then, whenever labeled samples of a new class occur, new affine modulations are learned that improve separation of the new class from all others, without affecting feedforward weights. By co-opting the view-invariance learning mechanism, we then train feedforward weights to match the unmodulated representation of a data sample to its modulated counterparts. This introduces modulation invariance into the representation space, and, by also using past modulations, stabilizes it. Our experiments show improvements in both class-incremental and transfer learning over state-of-the-art unsupervised approaches, as well as over comparable supervised approaches, using as few as 1% of available labels. Taken together, our work suggests that top-down modulations play a crucial role in balancing stability and plasticity.

Key Contributions

This paper introduces Task-Modulated Contrastive Learning (TMCL), a novel approach inspired by biological brains to achieve sparsely supervised continual learning. TMCL leverages predictive coding and contrastive loss to build view-invariant representations and integrates new classes with top-down modulations without compromising generalization, addressing catastrophic forgetting in machine learning.

Business Value

Enables AI systems to learn continuously and adapt to new information with limited data, crucial for applications requiring long-term operation and evolving environments.