Fast Stochastic Bregman Gradient Methods: Sharp Analysis and Variance Reduction - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Proceedings of the 38 th International Conference on Machine Learning Année : 2021

Fast Stochastic Bregman Gradient Methods: Sharp Analysis and Variance Reduction

Résumé

We study the problem of minimizing a relatively-smooth convex function using stochastic Bregman gradient methods. We first prove the convergence of Bregman Stochastic Gradient Descent (BSGD) to a region that depends on the noise (magnitude of the gradients) at the optimum. In particular, BSGD with a constant step-size converges to the exact minimizer when this noise is zero (\emph{interpolation} setting, in which the data is fit perfectly). Otherwise, when the objective has a finite sum structure, we show that variance reduction can be used to counter the effect of noise. In particular, fast convergence to the exact minimizer can be obtained under additional regularity assumptions on the Bregman reference function. We illustrate the effectiveness of our approach on two key applications of relative smoothness: tomographic reconstruction with Poisson noise and statistical preconditioning for distributed optimization.

Dates et versions

hal-03383164 , version 1 (18-10-2021)

Identifiants

Citer

Radu-Alexandru Dragomir, Hadrien Hendrikx, Mathieu Even. Fast Stochastic Bregman Gradient Methods: Sharp Analysis and Variance Reduction. ICML 2021- 38th International Conference on Machine Learning, Jul 2021, virtual, United States. pp.2815-2825. ⟨hal-03383164⟩
55 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More