ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2510.20616
132
0

On Optimal Hyperparameters for Differentially Private Deep Transfer Learning

23 October 2025
Aki Rehn
Linzh Zhao
Mikko Heikkilä
Antti Honkela
ArXiv (abs)PDFHTML
Main:10 Pages
21 Figures
Bibliography:4 Pages
3 Tables
Appendix:11 Pages
Abstract

Differentially private (DP) transfer learning, i.e., fine-tuning a pretrained model on private data, is the current state-of-the-art approach for training large models under privacy constraints. We focus on two key hyperparameters in this setting: the clipping bound CCC and batch size BBB. We show a clear mismatch between the current theoretical understanding of how to choose an optimal CCC (stronger privacy requires smaller CCC) and empirical outcomes (larger CCC performs better under strong privacy), caused by changes in the gradient distributions. Assuming a limited compute budget (fixed epochs), we demonstrate that the existing heuristics for tuning BBB do not work, while cumulative DP noise better explains whether smaller or larger batches perform better. We also highlight how the common practice of using a single (C,B)(C,B)(C,B) setting across tasks can lead to suboptimal performance. We find that performance drops especially when moving between loose and tight privacy and between plentiful and limited compute, which we explain by analyzing clipping as a form of gradient re-weighting and examining cumulative DP noise.

View on arXiv
Comments on this paper