525
v1v2v3v4v5 (latest)

Antidistillation Sampling

Main:10 Pages
11 Figures
Bibliography:4 Pages
Appendix:8 Pages
Abstract

Frontier models that generate extended reasoning traces inadvertently produce rich token sequences that can facilitate model distillation. Recognizing this vulnerability, model owners may seek sampling strategies that limit the effectiveness of distillation without compromising model performance. Antidistillation sampling provides exactly this capability. By strategically modifying a model's next-token probability distribution, antidistillation sampling poisons reasoning traces, rendering them significantly less effective for distillation while preserving the model's practical utility. For further details, seethis https URL.

View on arXiv
Comments on this paper