Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Text summarization is a crucial task that requires the simultaneous
optimization of multiple objectives, including consistency, coherence,
relevance, and fluency, which presents considerable challenges. Although large
language models (LLMs) have demonstrated remarkable performance, enhanced by
reinforcement learning (RL), few studies have focused on optimizing the
multi-objective problem of summarization through RL based on LLMs. In this
paper, we introduce hypervolume optimization (HVO), a novel optimization
strategy that dynamically adjusts the scores between groups during the reward
process in RL by using the hypervolume method. This method guides the model's
optimization to progressively approximate the pareto front, thereby generating
balanced summaries across multiple objectives. Experimental results on several
representative summarization datasets demonstrate that our method outperforms
group relative policy optimization (GRPO) in overall scores and shows more
balanced performance across different dimensions. Moreover, a 7B foundation
model enhanced by HVO performs comparably to GPT-4 in the summarization task,
while maintaining a shorter generation length. Our code is publicly available
at https://github.com/ai4business-LiAuto/HVO.git
Authors (7)
Junjie Song
Yiwen Liu
Dapeng Li
Yin Sun
Shukun Fu
Siqi Chen
+1 more
Submitted
October 22, 2025
Key Contributions
Introduces Hypervolume Optimization (HVO) for multi-objective reinforcement learning in text summarization. HVO dynamically adjusts reward scores to guide LLMs towards the Pareto front, enabling the generation of summaries that are balanced across multiple objectives like consistency, coherence, relevance, and fluency.
Business Value
Improves the quality and utility of automated text summarization tools, making them more reliable for applications like news aggregation, document analysis, and content creation.