Online Self-Concordant and Relatively Smooth Minimization, With Applications to Online Portfolio Selection and Learning Quantum States (2210.00997v3)
Abstract: Consider an online convex optimization problem where the loss functions are self-concordant barriers, smooth relative to a convex function $h$, and possibly non-Lipschitz. We analyze the regret of online mirror descent with $h$. Then, based on the result, we prove the following in a unified manner. Denote by $T$ the time horizon and $d$ the parameter dimension. 1. For online portfolio selection, the regret of $\widetilde{\text{EG}}$, a variant of exponentiated gradient due to Helmbold et al., is $\tilde{O} ( T{2/3} d{1/3} )$ when $T > 4 d / \log d$. This improves on the original $\tilde{O} ( T{3/4} d{1/2} )$ regret bound for $\widetilde{\text{EG}}$. 2. For online portfolio selection, the regret of online mirror descent with the logarithmic barrier is $\tilde{O}(\sqrt{T d})$. The regret bound is the same as that of Soft-Bayes due to Orseau et al. up to logarithmic terms. 3. For online learning quantum states with the logarithmic loss, the regret of online mirror descent with the log-determinant function is also $\tilde{O} ( \sqrt{T d} )$. Its per-iteration time is shorter than all existing algorithms we know.
- On-line portfolio selection using multiplicative updates. Math. Financ., 8(4):325–347, 1998.
- Soft-Bayes: Prod for mixtures of experts with log-loss. In Proc. 28th Int. Conf. Algorithmic Learning Theory, pages 372–399, 2017.
- Thomas M. Cover. Universal portfolios. Math. Financ., 1(1):1–29, 1991.
- Universal portfolios with side information. IEEE Trans. Inf. Theory, 42(2):348–363, 1996.
- Efficient algorithms for universal portfolios. J. Mach. Learn. Res., 3:423–440, 2002.
- Yurii Nesterov. Barrier subgradient method. Math. Program., Ser. B, 127:31–56, 2011.
- Efficient online portfolio with logarithmic regret. In Adv. Neural Information Processing Systems 31, 2018.
- Damped Online Newton Step for Portfolio Selection. In Proc. 35th Annu. Conf. Learning Theory, pages 5561–5595, 2022.
- Pushing the efficiency-regret Pareto frontier for online learning of portfolios and quantum states. In Proc. 35th Annu. Conf. Learning Theory, pages 182–226, 2022.
- An Online Algorithm for Maximum-Likelihood Quantum State Tomography. In 24th Annu. Conf. Quantum Information Processing, 2021. arXiv:2012.15498.
- Interior-Point Polynomial Algorithms in Convex Programming. SIAM, Philadelphia, PA, 1994.
- A descent lemma beyond Lipschitz gradient continuity: first-order methods revisited and applications. Math. Oper. Res., 42(2):330–348, 2017.
- Relatively smooth convex optimization by first-order methods, and applications. SIAM J. Optim., 28(1):333–354, 2018.
- Dynamic regret of online mirror descent for relatively smooth convex cost functions. IEEE Control Syst. Lett., 6:2395–2400, 2022.
- Online variance minimization. In 19th Annu. Conf. Learning Theory, COLT 2006, pages 514–528, 2006.
- A combinatorial, primal-dual approach to semidefinite programs. In STOC ’07: Proc. 39th Annu. ACM Symp. Theory of computing, pages 227–236, 2007.
- Learning eigenvectors for free. In Adv. Neural Information Processing Systems 24, 2011.
- Online learning of quantum states. In Adv. Neural Information Processing Systems 31, 2018.
- Adaptive online learning of quantum states. 2022. arXiv:2206.00220 [cs.LG].
- Revisiting online quantum state learning. In Proc. AAAI Conf. Artificial Intelligence, 2020.
- Bandit principal component analysis. In Proc. 32nd Conf. Learning Theory, pages 1994–2024, 2019.
- Competing in the dark: An efficient algorithm for bandit linear optimization. In Proc. 21st Annu. Conf. Learning Theory, pages 263–273, 2008.
- Online learning with predictable sequences. In Proc. 26th Annu. Conf. Learning Theory, pages 993–1019, 2013.
- Elad Hazan. Introduction to online convex optimization. Found. Trends Opt., 2(3–4):157–325, 2016.
- Prediction, Learning, and Games. Cambridge Univ. Press, Cambridge, UK, 2006.
- Open Problem: Fast and Optimal Online Portfolio Selection. In Proc. 33rd Annu. Conf. Learning Theory, pages 3864–3869, 2020.
- Improved dynamic regret for non-degenerate functions. In Adv. in Neural Information Processing Systems 30, 2017.
- Smoothness, low-noise and fast rates. In Adv. Neural Information Processing Systems 23, 2010.
- Francesco Orabona. A modern introduction to online learning. 2022. arXiv:1912.13213v5 [cs.LG].
- Online optimization: Competing with dynamic comparators. In Proc. 18th Int. Conf. Artificial Intelligence and Statistics, pages 398–406, 2015.
- Online optimization in dynamic environments: Improved regret rates for strongly convex problems. In IEEE 55th Conf. Decision and Control (CDC), pages 7195–7201, 2016.
- Interior-Point Methods for Full-Information and Bandit Online Learning. IEEE Trans. Inf. Theory, 58(7):4164–4175, 2012.
- Yurii Nesterov. Lectures on Convex Optimization. Springer, Cham, CH, second edition, 2018.
- Marc Teboulle. A simplified view of first order methods for optimization. Math. Program., Ser. B, 170:67–96, 2018.
- Logarithmic regret algorithms for online convex optimization. Mach. Learn., 69(2):169–192, 2007.
- Rajendra Bhatia. Matrix Analysis. Springer, New York, NY, 1997.
- Introduction to Matrix Analysis and Applications. Springer, Cham, 2014.