View a PDF of the paper titled Weak-to-Strong Generalization under Distribution Shifts, by Myeongho Jeon and 3 other authors
View PDF
Abstract:As future superhuman models become increasingly complex, accurately supervising their behavior may exceed human capabilities. Recent works have demonstrated that in such scenarios, weak models can effectively supervise strong models, a phenomenon known as weak-to-strong generalization. However, we find that naive weak-to-strong generalization fails under distribution shifts, often leading to worse performance of the strong model than its weak supervisors. To address this, we propose RAVEN, a robust weak-to-strong generalization framework that dynamically learns the optimal combinations of weak models in addition to parameters of the strong model. We demonstrate the effectiveness of RAVEN on image classification, text classification, and preference alignment tasks. RAVEN outperforms alternative baselines by over 30% on out-of-distribution tasks while matching or surpassing existing methods on in-distribution tasks. Moreover, our results show that RAVEN assigns higher weights to more accurate weak models, demonstrating its ability to automatically identify trustworthy supervision.
Submission history
From: Jan Sobotka [view email]
[v1]
Fri, 24 Oct 2025 10:46:50 UTC (2,041 KB)
[v2]
Tue, 25 Nov 2025 21:37:10 UTC (2,033 KB)


![[2510.21332] Weak-to-Strong Generalization under Distribution Shifts Measuring Intelligence Efficiency of Local AI](https://skytik.cc/wp-content/uploads/2025/11/Measuring-Intelligence-Efficiency-of-Local-AI-768x448.png)