Stress-Aware Learning under KL Drift via Trust-Decayed Mirror Descent
We study sequential decision-making under distribution drift. We propose entropy-regularized trust-decay, which injects stress-aware exponential tilting into both belief updates and mirror-descent decisions. On the simplex, a Fenchel-dual equivalence shows that belief tilt and decision tilt coincide. We formalize robustness via fragility (worst-case excess risk in a KL ball), belief bandwidth (radius sustaining a target excess), and a decision-space Fragility Index (drift tolerated at regret). We prove high-probability sensitivity bounds and establish dynamic-regret guarantees of under KL-drift path length . In particular, trust-decay achieves per-switch regret, while stress-free updates incur tails. A parameter-free hedge adapts the tilt to unknown drift, whereas persistent over-tilting yields an stationary penalty. We further obtain calibrated-stress bounds and extensions to second-order updates, bandit feedback, outliers, stress variation, distributed optimization, and plug-in KL-drift estimation. The framework unifies dynamic-regret analysis, distributionally robust objectives, and KL-regularized control within a single stress-adaptive update.
View on arXiv