Emergent Mind

Faster Gradient-Free Algorithms for Nonsmooth Nonconvex Stochastic Optimization

(2301.06428)
Published Jan 16, 2023 in math.OC and cs.LG

Abstract

We consider the optimization problem of the form $\min{x \in \mathbb{R}d} f(x) \triangleq \mathbb{E}{\xi} [F(x; \xi)]$, where the component $F(x;\xi)$ is $L$-mean-squared Lipschitz but possibly nonconvex and nonsmooth. The recently proposed gradient-free method requires at most $\mathcal{O}( L4 d{3/2} \epsilon{-4} + \Delta L3 d{3/2} \delta{-1} \epsilon{-4})$ stochastic zeroth-order oracle complexity to find a $(\delta,\epsilon)$-Goldstein stationary point of objective function, where $\Delta = f(x0) - \inf{x \in \mathbb{R}d} f(x)$ and $x_0$ is the initial point of the algorithm. This paper proposes a more efficient algorithm using stochastic recursive gradient estimators, which improves the complexity to $\mathcal{O}(L3 d{3/2} \epsilon{-3}+ \Delta L2 d{3/2} \delta{-1} \epsilon{-3})$.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.