Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 90% Match Research Paper VLM researchers,Robotics engineers,AI researchers in physical reasoning,Computer vision scientists 1 week ago

Physics Context Builders: A Modular Framework for Physical Reasoning in Vision-Language Models

computer-vision › scene-understanding
📄 Abstract

Abstract: Physical reasoning remains a significant challenge for Vision-Language Models (VLMs). This limitation arises from an inability to translate learned knowledge into predictions about physical behavior. Although continual fine-tuning can mitigate this issue, it is expensive for large models and impractical to perform repeatedly for every task. This necessitates the creation of modular and scalable ways to teach VLMs about physical reasoning. To that end, we introduce Physics Context Builders (PCBs), a modular framework where specialized smaller VLMs are fine-tuned to generate detailed physical scene descriptions. These can be used as physical contexts to enhance the reasoning capabilities of larger VLMs. PCBs enable the separation of visual perception from reasoning, allowing us to analyze their relative contributions to physical understanding. We perform experiments on CLEVRER and on Falling Tower, a stability detection dataset with both simulated and real-world scenes, to demonstrate that PCBs provide substantial performance improvements, increasing average accuracy by up to 13.8% on complex physical reasoning tasks. Notably, PCBs also show strong Sim2Real transfer, successfully generalizing from simulated training data to real-world scenes.
Authors (5)
Vahid Balazadeh
Mohammadmehdi Ataei
Hyunmin Cheong
Amir Hosein Khasahmadi
Rahul G. Krishnan
Submitted
December 11, 2024
arXiv Category
cs.CV
arXiv PDF

Key Contributions

This paper introduces Physics Context Builders (PCBs), a modular framework that enhances Vision-Language Models (VLMs) with physical reasoning capabilities. Specialized smaller VLMs are fine-tuned to generate detailed physical scene descriptions, which are then used as context to improve the reasoning of larger VLMs, allowing for separation of visual perception and reasoning.

Business Value

Enables AI systems to better understand and predict physical interactions in the real world, crucial for developing safer autonomous systems (e.g., self-driving cars, robots) and more intuitive human-AI interaction.