2000 character limit reached
On stochastic gradient Langevin dynamics with dependent data streams: the fully non-convex case (1905.13142v4)
Published 30 May 2019 in math.ST, math.PR, stat.ML, and stat.TH
Abstract: We consider the problem of sampling from a target distribution, which is \emph {not necessarily logconcave}, in the context of empirical risk minimization and stochastic optimization as presented in Raginsky et al. (2017). Non-asymptotic analysis results are established in the $L1$-Wasserstein distance for the behaviour of Stochastic Gradient Langevin Dynamics (SGLD) algorithms. We allow the estimation of gradients to be performed even in the presence of \emph{dependent} data streams. Our convergence estimates are sharper and \emph{uniform} in the number of iterations, in contrast to those in previous studies.
- Ngoc Huy Chau (2 papers)
- Sotirios Sabanis (37 papers)
- Ying Zhang (389 papers)
- Éric Moulines (21 papers)
- Miklos Rásonyi (1 paper)