Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Recent advancements in language models (LMs) have marked a shift toward the
growing importance of post-training. Yet, post-training approaches such as
supervised fine-tuning (SFT) do not guarantee the effective use of knowledge
acquired during pretraining. We therefore introduce InfoSteer, a lightweight
method that encourages parametric information utilization in LMs during
post-training. Specifically, InfoSteer treats the feed-forward network (FFN)
layer as associate key-value memory and promotes the use of stored memory
vectors via forward-pass interventions or regularization during
backpropagation. This simple guidance during post-training phase yields
consistent performance improvements across diverse model families -- including
Qwen, Gemma and Llama -- spanning 15 downstream tasks in both in-distribution
(ID) and out-of-distribution (OOD) evaluations. Beyond performance gains, we
also find that steered LMs can adaptively allocate information by placing more
emphasis on generating semantically meaningful tokens, while using fewer
resources on simple transition ones (e.g., `\texttt{,}' or `\texttt{and}'). Our
work underscores that vanilla post-training does not fully exploit the
potential gained during pre-training, and that steering LMs in latent
representation space offers a promising approach to enhance both performance
and interpretability. The code is available at:
https://github.com/chili-lab/InfoSteer.
Authors (3)
Chunyuan Deng
Ruidi Chang
Hanjie Chen
Key Contributions
Introduces InfoSteer, a lightweight post-training method that enhances parametric information utilization in LMs by treating FFN layers as key-value memory. Through forward-pass interventions or regularization, it guides LMs to better leverage pre-trained knowledge, leading to consistent performance improvements across diverse models and tasks, including OOD evaluations.
Business Value
Improves the efficiency and effectiveness of post-training LLMs, leading to better performance on downstream tasks and reducing the need for extensive retraining.