2000 character limit reached
Backward error analysis and the qualitative behaviour of stochastic optimization algorithms: Application to stochastic coordinate descent (2309.02082v1)
Published 5 Sep 2023 in math.OC, cs.NA, math.NA, and stat.ML
Abstract: Stochastic optimization methods have been hugely successful in making large-scale optimization problems feasible when computing the full gradient is computationally prohibitive. Using the theory of modified equations for numerical integrators, we propose a class of stochastic differential equations that approximate the dynamics of general stochastic optimization methods more closely than the original gradient flow. Analyzing a modified stochastic differential equation can reveal qualitative insights about the associated optimization method. Here, we study mean-square stability of the modified equation in the case of stochastic coordinate descent.