Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Online Self-Concordant and Relatively Smooth Minimization, With Applications to Online Portfolio Selection and Learning Quantum States (2210.00997v3)

Published 3 Oct 2022 in stat.ML, cs.LG, math.OC, q-fin.PM, and quant-ph

Abstract: Consider an online convex optimization problem where the loss functions are self-concordant barriers, smooth relative to a convex function $h$, and possibly non-Lipschitz. We analyze the regret of online mirror descent with $h$. Then, based on the result, we prove the following in a unified manner. Denote by $T$ the time horizon and $d$ the parameter dimension. 1. For online portfolio selection, the regret of $\widetilde{\text{EG}}$, a variant of exponentiated gradient due to Helmbold et al., is $\tilde{O} ( T{2/3} d{1/3} )$ when $T > 4 d / \log d$. This improves on the original $\tilde{O} ( T{3/4} d{1/2} )$ regret bound for $\widetilde{\text{EG}}$. 2. For online portfolio selection, the regret of online mirror descent with the logarithmic barrier is $\tilde{O}(\sqrt{T d})$. The regret bound is the same as that of Soft-Bayes due to Orseau et al. up to logarithmic terms. 3. For online learning quantum states with the logarithmic loss, the regret of online mirror descent with the log-determinant function is also $\tilde{O} ( \sqrt{T d} )$. Its per-iteration time is shorter than all existing algorithms we know.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. On-line portfolio selection using multiplicative updates. Math. Financ., 8(4):325–347, 1998.
  2. Soft-Bayes: Prod for mixtures of experts with log-loss. In Proc. 28th Int. Conf. Algorithmic Learning Theory, pages 372–399, 2017.
  3. Thomas M. Cover. Universal portfolios. Math. Financ., 1(1):1–29, 1991.
  4. Universal portfolios with side information. IEEE Trans. Inf. Theory, 42(2):348–363, 1996.
  5. Efficient algorithms for universal portfolios. J. Mach. Learn. Res., 3:423–440, 2002.
  6. Yurii Nesterov. Barrier subgradient method. Math. Program., Ser. B, 127:31–56, 2011.
  7. Efficient online portfolio with logarithmic regret. In Adv. Neural Information Processing Systems 31, 2018.
  8. Damped Online Newton Step for Portfolio Selection. In Proc. 35th Annu. Conf. Learning Theory, pages 5561–5595, 2022.
  9. Pushing the efficiency-regret Pareto frontier for online learning of portfolios and quantum states. In Proc. 35th Annu. Conf. Learning Theory, pages 182–226, 2022.
  10. An Online Algorithm for Maximum-Likelihood Quantum State Tomography. In 24th Annu. Conf. Quantum Information Processing, 2021. arXiv:2012.15498.
  11. Interior-Point Polynomial Algorithms in Convex Programming. SIAM, Philadelphia, PA, 1994.
  12. A descent lemma beyond Lipschitz gradient continuity: first-order methods revisited and applications. Math. Oper. Res., 42(2):330–348, 2017.
  13. Relatively smooth convex optimization by first-order methods, and applications. SIAM J. Optim., 28(1):333–354, 2018.
  14. Dynamic regret of online mirror descent for relatively smooth convex cost functions. IEEE Control Syst. Lett., 6:2395–2400, 2022.
  15. Online variance minimization. In 19th Annu. Conf. Learning Theory, COLT 2006, pages 514–528, 2006.
  16. A combinatorial, primal-dual approach to semidefinite programs. In STOC ’07: Proc. 39th Annu. ACM Symp. Theory of computing, pages 227–236, 2007.
  17. Learning eigenvectors for free. In Adv. Neural Information Processing Systems 24, 2011.
  18. Online learning of quantum states. In Adv. Neural Information Processing Systems 31, 2018.
  19. Adaptive online learning of quantum states. 2022. arXiv:2206.00220 [cs.LG].
  20. Revisiting online quantum state learning. In Proc. AAAI Conf. Artificial Intelligence, 2020.
  21. Bandit principal component analysis. In Proc. 32nd Conf. Learning Theory, pages 1994–2024, 2019.
  22. Competing in the dark: An efficient algorithm for bandit linear optimization. In Proc. 21st Annu. Conf. Learning Theory, pages 263–273, 2008.
  23. Online learning with predictable sequences. In Proc. 26th Annu. Conf. Learning Theory, pages 993–1019, 2013.
  24. Elad Hazan. Introduction to online convex optimization. Found. Trends Opt., 2(3–4):157–325, 2016.
  25. Prediction, Learning, and Games. Cambridge Univ. Press, Cambridge, UK, 2006.
  26. Open Problem: Fast and Optimal Online Portfolio Selection. In Proc. 33rd Annu. Conf. Learning Theory, pages 3864–3869, 2020.
  27. Improved dynamic regret for non-degenerate functions. In Adv. in Neural Information Processing Systems 30, 2017.
  28. Smoothness, low-noise and fast rates. In Adv. Neural Information Processing Systems 23, 2010.
  29. Francesco Orabona. A modern introduction to online learning. 2022. arXiv:1912.13213v5 [cs.LG].
  30. Online optimization: Competing with dynamic comparators. In Proc. 18th Int. Conf. Artificial Intelligence and Statistics, pages 398–406, 2015.
  31. Online optimization in dynamic environments: Improved regret rates for strongly convex problems. In IEEE 55th Conf. Decision and Control (CDC), pages 7195–7201, 2016.
  32. Interior-Point Methods for Full-Information and Bandit Online Learning. IEEE Trans. Inf. Theory, 58(7):4164–4175, 2012.
  33. Yurii Nesterov. Lectures on Convex Optimization. Springer, Cham, CH, second edition, 2018.
  34. Marc Teboulle. A simplified view of first order methods for optimization. Math. Program., Ser. B, 170:67–96, 2018.
  35. Logarithmic regret algorithms for online convex optimization. Mach. Learn., 69(2):169–192, 2007.
  36. Rajendra Bhatia. Matrix Analysis. Springer, New York, NY, 1997.
  37. Introduction to Matrix Analysis and Applications. Springer, Cham, 2014.
Citations (7)

Summary

We haven't generated a summary for this paper yet.