Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 93% Match Research Paper AI researchers,VLM developers,multimodal AI engineers,computer vision specialists 1 week ago

SteerVLM: Robust Model Control through Lightweight Activation Steering for Vision Language Models

large-language-models › model-architecture
📄 Abstract

Abstract: This work introduces SteerVLM, a lightweight steering module designed to guide Vision-Language Models (VLMs) towards outputs that better adhere to desired instructions. Our approach learns from the latent embeddings of paired prompts encoding target and converse behaviors to dynamically adjust activations connecting the language modality with image context. This allows for fine-grained, inference-time control over complex output semantics without modifying model weights while preserving performance on off-target tasks. Our steering module requires learning parameters equal to 0.14% of the original VLM's size. Our steering module gains model control through dimension-wise activation modulation and adaptive steering across layers without requiring pre-extracted static vectors or manual tuning of intervention points. Furthermore, we introduce VNIA (Visual Narrative Intent Alignment), a multimodal dataset specifically created to facilitate the development and evaluation of VLM steering techniques. Our method outperforms existing intervention techniques on steering and hallucination mitigation benchmarks for VLMs and proposes a robust solution for multimodal model control through activation engineering.
Authors (4)
Anushka Sivakumar
Andrew Zhang
Zaber Hakim
Chris Thomas
Submitted
October 30, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

SteerVLM introduces a lightweight activation steering module for Vision-Language Models (VLMs) that enables fine-grained, inference-time control over outputs without modifying model weights. It learns from latent embeddings to dynamically adjust activations, preserving performance on off-target tasks and requiring only 0.14% of the original VLM's parameters. This offers a robust and efficient way to steer VLM behavior.

Business Value

Enables the creation of more controllable and reliable multimodal AI applications, such as image generation tools that precisely follow user prompts or visual assistants that adapt their responses based on nuanced instructions.