Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 88% Match Research Paper Computer vision researchers,AI researchers,Robotics engineers,ML engineers 1 week ago

Rethinking Visual Intelligence: Insights from Video Pretraining

computer-vision › diffusion-models
📄 Abstract

Abstract: Large language models (LLMs) have demonstrated that large-scale pretraining enables systems to adapt rapidly to new problems with little supervision in the language domain. This success, however, has not translated as effectively to the visual domain, where models, including LLMs, continue to struggle with compositional understanding, sample efficiency, and general-purpose problem-solving. We investigate Video Diffusion Models (VDMs) as a promising direction for bridging this gap. Pretraining on spatiotemporal data endows these models with strong inductive biases for structure and dynamics, which we hypothesize can support broad task adaptability. To test this, we design a controlled evaluation in which both a pretrained LLM and a pretrained VDM are equipped with lightweight adapters and presented with tasks in their natural modalities. Across benchmarks including ARC-AGI, ConceptARC, visual games, route planning, and cellular automata, VDMs demonstrate higher data efficiency than their language counterparts. Taken together, our results indicate that video pretraining offers inductive biases that support progress toward visual foundation models.
Authors (7)
Pablo Acuaviva
Aram Davtyan
Mariam Hassan
Sebastian Stapf
Ahmad Rahimi
Alexandre Alahi
+1 more
Submitted
October 28, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

This paper investigates Video Diffusion Models (VDMs) as a promising direction for improving visual intelligence, hypothesizing that pretraining on spatiotemporal data endows models with strong inductive biases for structure and dynamics. Through controlled evaluations across various benchmarks, VDMs demonstrate higher data efficiency compared to pretrained LLMs equipped with similar adapters.

Business Value

Could lead to more capable AI systems in vision-heavy applications like robotics, autonomous driving, and video analysis, requiring less data for training.