Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 95% Match Research Paper ML Researchers,NLP Engineers,LLM Developers 6 days ago

Steering Information Utility in Key-Value Memory for Language Model Post-Training

large-language-models › training-methods
📄 Abstract

Abstract: Recent advancements in language models (LMs) have marked a shift toward the growing importance of post-training. Yet, post-training approaches such as supervised fine-tuning (SFT) do not guarantee the effective use of knowledge acquired during pretraining. We therefore introduce InfoSteer, a lightweight method that encourages parametric information utilization in LMs during post-training. Specifically, InfoSteer treats the feed-forward network (FFN) layer as associate key-value memory and promotes the use of stored memory vectors via forward-pass interventions or regularization during backpropagation. This simple guidance during post-training phase yields consistent performance improvements across diverse model families -- including Qwen, Gemma and Llama -- spanning 15 downstream tasks in both in-distribution (ID) and out-of-distribution (OOD) evaluations. Beyond performance gains, we also find that steered LMs can adaptively allocate information by placing more emphasis on generating semantically meaningful tokens, while using fewer resources on simple transition ones (e.g., `\texttt{,}' or `\texttt{and}'). Our work underscores that vanilla post-training does not fully exploit the potential gained during pre-training, and that steering LMs in latent representation space offers a promising approach to enhance both performance and interpretability. The code is available at: https://github.com/chili-lab/InfoSteer.
Authors (3)
Chunyuan Deng
Ruidi Chang
Hanjie Chen
Submitted
July 7, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

Introduces InfoSteer, a lightweight post-training method that enhances parametric information utilization in LMs by treating FFN layers as key-value memory. Through forward-pass interventions or regularization, it guides LMs to better leverage pre-trained knowledge, leading to consistent performance improvements across diverse models and tasks, including OOD evaluations.

Business Value

Improves the efficiency and effectiveness of post-training LLMs, leading to better performance on downstream tasks and reducing the need for extensive retraining.