ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2509.24005
120
0

Does Weak-to-strong Generalization Happen under Spurious Correlations?

28 September 2025
Chenruo Liu
Yijun Dong
Qi Lei
ArXiv (abs)PDFHTML
Main:12 Pages
7 Figures
Bibliography:6 Pages
9 Tables
Appendix:16 Pages
Abstract

We initiate a unified theoretical and algorithmic study of a key problem in weak-to-strong (W2S) generalization: when fine-tuning a strong pre-trained student with pseudolabels from a weaker teacher on a downstream task with spurious correlations, does W2S happen, and how to improve it upon failures? We consider two sources of spurious correlations caused by group imbalance: (i) a weak teacher fine-tuned on group-imbalanced labeled data with a minority group of fraction ηℓ\eta_\ellηℓ​, and (ii) a group-imbalanced unlabeled set pseudolabeled by the teacher with a minority group of fraction ηu\eta_uηu​. Theoretically, a precise characterization of W2S gain at the proportional asymptotic limit shows that W2S always happens with sufficient pseudolabels when ηu=ηℓ\eta_u = \eta_\ellηu​=ηℓ​ but may fail when ηu≠ηℓ\eta_u \ne \eta_\ellηu​=ηℓ​, where W2S gain diminishes as (ηu−ηℓ)2(\eta_u - \eta_\ell)^2(ηu​−ηℓ​)2 increases. Our theory is corroborated by extensive experiments on various spurious correlation benchmarks and teacher-student pairs. To boost W2S performance upon failures, we further propose a simple, effective algorithmic remedy that retrains the strong student on its high-confidence data subset after W2S fine-tuning. Our algorithm is group-label-free and achieves consistent, substantial improvements over vanilla W2S fine-tuning.

View on arXiv
Comments on this paper