139

Stress-Aware Learning under KL Drift via Trust-Decayed Mirror Descent

Main:7 Pages
Bibliography:2 Pages
Appendix:6 Pages
Abstract

We study sequential decision-making under distribution drift. We propose entropy-regularized trust-decay, which injects stress-aware exponential tilting into both belief updates and mirror-descent decisions. On the simplex, a Fenchel-dual equivalence shows that belief tilt and decision tilt coincide. We formalize robustness via fragility (worst-case excess risk in a KL ball), belief bandwidth (radius sustaining a target excess), and a decision-space Fragility Index (drift tolerated at O(T)O(\sqrt{T}) regret). We prove high-probability sensitivity bounds and establish dynamic-regret guarantees of O~(T)\tilde{O}(\sqrt{T}) under KL-drift path length ST=t2KL(DtDt1)/2S_T = \sum_{t\ge2}\sqrt{{\rm KL}(D_t|D_{t-1})/2}. In particular, trust-decay achieves O(1)O(1) per-switch regret, while stress-free updates incur Ω(1)\Omega(1) tails. A parameter-free hedge adapts the tilt to unknown drift, whereas persistent over-tilting yields an Ω(λ2T)\Omega(\lambda^2 T) stationary penalty. We further obtain calibrated-stress bounds and extensions to second-order updates, bandit feedback, outliers, stress variation, distributed optimization, and plug-in KL-drift estimation. The framework unifies dynamic-regret analysis, distributionally robust objectives, and KL-regularized control within a single stress-adaptive update.

View on arXiv
Comments on this paper