Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Stochastic modified equations for the asynchronous stochastic gradient descent (1805.08244v3)

Published 21 May 2018 in stat.ML and cs.LG

Abstract: We propose a stochastic modified equations (SME) for modeling the asynchronous stochastic gradient descent (ASGD) algorithms. The resulting SME of Langevin type extracts more information about the ASGD dynamics and elucidates the relationship between different types of stochastic gradient algorithms. We show the convergence of ASGD to the SME in the continuous time limit, as well as the SME's precise prediction to the trajectories of ASGD with various forcing terms. As an application of the SME, we propose an optimal mini-batching strategy for ASGD via solving the optimal control problem of the associated SME.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Jing An (18 papers)
  2. Jianfeng Lu (273 papers)
  3. Lexing Ying (159 papers)

Summary

We haven't generated a summary for this paper yet.