Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 64 tok/s
Gemini 2.5 Pro 50 tok/s Pro
GPT-5 Medium 30 tok/s Pro
GPT-5 High 35 tok/s Pro
GPT-4o 77 tok/s Pro
Kimi K2 174 tok/s Pro
GPT OSS 120B 457 tok/s Pro
Claude Sonnet 4 37 tok/s Pro
2000 character limit reached

Tight Bounds for Online Convex Optimization with Adversarial Constraints (2405.09296v1)

Published 15 May 2024 in cs.LG and math.OC

Abstract: A well-studied generalization of the standard online convex optimization (OCO) is constrained online convex optimization (COCO). In COCO, on every round, a convex cost function and a convex constraint function are revealed to the learner after the action for that round is chosen. The objective is to design an online policy that simultaneously achieves a small regret while ensuring small cumulative constraint violation (CCV) against an adaptive adversary. A long-standing open question in COCO is whether an online policy can simultaneously achieve $O(\sqrt{T})$ regret and $O(\sqrt{T})$ CCV without any restrictive assumptions. For the first time, we answer this in the affirmative and show that an online policy can simultaneously achieve $O(\sqrt{T})$ regret and $\tilde{O}(\sqrt{T})$ CCV. We establish this result by effectively combining the adaptive regret bound of the AdaGrad algorithm with Lyapunov optimization - a classic tool from control theory. Surprisingly, the analysis is short and elegant.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (14)
  1. Adaptive subgradient methods for online learning and stochastic optimization. Journal of machine learning research, 12(7), 2011.
  2. Online convex optimization with hard constraints: Towards the best of two worlds and beyond. Advances in Neural Information Processing Systems, 35:36426–36439, 2022.
  3. Hazan, E. Introduction to online convex optimization. MIT Press, 2022.
  4. Adaptive algorithms for online convex optimization with long-term constraints. In International Conference on Machine Learning, pp. 402–411. PMLR, 2016.
  5. Trading regret for efficiency: online convex optimization with long term constraints. The Journal of Machine Learning Research, 13(1):2503–2528, 2012.
  6. Neely, M. J. Stochastic network optimization with application to communication and queueing systems. Synthesis Lectures on Communication Networks, 3(1):1–211, 2010.
  7. Online convex optimization with time-varying constraints. arXiv preprint arXiv:1702.04783, 2017.
  8. Orabona, F. A modern introduction to online learning. arXiv preprint arXiv:1912.13213, 2019.
  9. Playing in the dark: No-regret learning with adversarial constraints, 2023.
  10. Regret and cumulative constraint violation analysis for online convex optimization with long term constraints. In International Conference on Machine Learning, pp. 11998–12008. PMLR, 2021.
  11. Distributed online convex optimization with adversarial constraints: Reduced cumulative constraint violation bounds under slater’s condition. arXiv preprint arXiv:2306.00149, 2023.
  12. A low complexity algorithm with 𝒪⁢(T)𝒪𝑇\mathcal{O}(\sqrt{T})caligraphic_O ( square-root start_ARG italic_T end_ARG ) regret and 𝒪⁢(1)𝒪1\mathcal{O}(1)caligraphic_O ( 1 ) constraint violations for online convex optimization with long term constraints. Journal of Machine Learning Research, 21(1):1–24, 2020.
  13. Online convex optimization with stochastic constraints. Advances in Neural Information Processing Systems, 30, 2017.
  14. Online convex optimization for cumulative constraints. Advances in Neural Information Processing Systems, 31, 2018.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com