Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: As future superhuman models become increasingly complex, accurately
supervising their behavior may exceed human capabilities. Recent works have
demonstrated that in such scenarios, weak models can effectively supervise
strong models, a phenomenon known as weak-to-strong generalization. However, we
find that naive weak-to-strong generalization fails under distribution shifts,
often leading to worse performance of the strong model than its weak
supervisors. To address this, we propose RAVEN, a robust weak-to-strong
generalization framework that dynamically learns the optimal combinations of
weak models in addition to parameters of the strong model. We demonstrate the
effectiveness of RAVEN on image classification, text classification, and
preference alignment tasks. RAVEN outperforms alternative baselines by over 30%
on out-of-distribution tasks while matching or surpassing existing methods on
in-distribution tasks. Moreover, our results show that RAVEN assigns higher
weights to more accurate weak models, demonstrating its ability to
automatically identify trustworthy supervision.
Authors (4)
Myeongho Jeon
Jan Sobotka
Suhwan Choi
Maria Brbić
Submitted
October 24, 2025
Key Contributions
This paper introduces RAVEN, a robust weak-to-strong generalization framework that addresses the failure of naive methods under distribution shifts. RAVEN dynamically learns optimal combinations of weak models to supervise strong models, significantly outperforming baselines on OOD tasks across various domains like image classification, text classification, and preference alignment.
Business Value
Enables the training of more reliable and robust AI systems that can generalize better to unseen data, crucial for safety-critical applications and reducing costly failures in production.