Source-Optimal Training is Transfer-Suboptimal

We prove that training a source model optimally for its own task is generically suboptimal when the objective is downstream transfer. We study the source-side optimization problem in L2-SP ridge regression and show a fundamental mismatch between the source-optimal and transfer-optimal source regularization: outside of a measure-zero set, . We characterize the transfer-optimal source penalty as a function of task alignment and identify an alignment-dependent reversal: with imperfect alignment (), transfer benefits from stronger source regularization, while in super-aligned regimes (), transfer benefits from weaker regularization. In isotropic settings, the decision of whether transfer helps is independent of the target sample size and noise, depending only on task alignment and source characteristics. We verify the linear predictions in a synthetic ridge regression experiment, and we present CIFAR-10 experiments as evidence that the source-optimal versus transfer-optimal mismatch can persist in nonlinear networks.
View on arXiv