Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
π Abstract
Abstract: Memory Mosaics [Zhang et al., 2025], networks of associative memories, have
demonstrated appealing compositional and in-context learning capabilities on
medium-scale networks (GPT-2 scale) and synthetic small datasets. This work
shows that these favorable properties remain when we scale memory mosaics to
large language model sizes (llama-8B scale) and real-world datasets.
To this end, we scale memory mosaics to 10B size, we train them on one
trillion tokens, we introduce a couple architectural modifications ("Memory
Mosaics v2"), we assess their capabilities across three evaluation dimensions:
training-knowledge storage, new-knowledge storage, and in-context learning.
Throughout the evaluation, memory mosaics v2 match transformers on the
learning of training knowledge (first dimension) and significantly outperforms
transformers on carrying out new tasks at inference time (second and third
dimensions). These improvements cannot be easily replicated by simply
increasing the training data for transformers. A memory mosaics v2 trained on
one trillion tokens still perform better on these tasks than a transformer
trained on eight trillion tokens.
Authors (2)
Jianyu Zhang
LΓ©on Bottou
Key Contributions
This work demonstrates that Memory Mosaics, a network of associative memories, retain their appealing compositional and in-context learning capabilities when scaled to large language model sizes (LLaMA-8B) and real-world datasets. Memory Mosaics v2 significantly outperform transformers in new knowledge storage and in-context learning.
Business Value
Paves the way for more capable and efficient large language models that can better learn and adapt to new information, potentially leading to more dynamic and personalized AI applications.