Online Self-Concordant and Relatively Smooth Minimization, With Applications to Online Portfolio Selection and Learning Quantum States

Consider an online convex optimization problem where the loss functions are self-concordant barriers, smooth relative to a convex function , and possibly non-Lipschitz. We analyze the regret of online mirror descent with . Then, based on the result, we prove the following in a unified manner. Denote by the time horizon and the parameter dimension. 1. For online portfolio selection, the regret of , a variant of exponentiated gradient due to Helmbold et al., is when . This improves on the original regret bound for . 2. For online portfolio selection, the regret of online mirror descent with the logarithmic barrier is . The regret bound is the same as that of Soft-Bayes due to Orseau et al. up to logarithmic terms. 3. For online learning quantum states with the logarithmic loss, the regret of online mirror descent with the log-determinant function is also . Its per-iteration time is shorter than all existing algorithms we know.
View on arXiv