RUS  ENG
Full version
JOURNALS // Doklady Rossijskoj Akademii Nauk. Mathematika, Informatika, Processy Upravlenia // Archive

Dokl. RAN. Math. Inf. Proc. Upr., 2023 Volume 514, Number 2, Pages 99–108 (Mi danma455)

SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES

Algorithms with gradient clipping for stochastic optimization with heavy-tailed noise

M. Yu. Danilova

Moscow Institute of Physics and Technology, Moscow, Russia

Abstract: This article provides a review of the results of several research studies, in which open questions related to the high-probability convergence analysis of stochastic first-order optimization methods under mild assumptions on the noise were gradually addressed. In the beginning, we introduce the concept of gradient clipping, which plays a pivotal role in the development of stochastic methods for successful operation in the case of heavy-tailed distributions. Next, we examine the importance of obtaining the highprobability convergence guarantees and their connection with in-expectation convergence guarantees. The concluding sections of the article are dedicated to presenting the primary findings related to minimization problems and the results of numerical experiments.

Keywords: convex optimization, stochastic optimization, first-order methods.

UDC: 004.8

Presented: A. A. Shananin
Received: 02.09.2023
Revised: 08.10.2023
Accepted: 15.10.2023

DOI: 10.31857/S2686954323601768


 English version:
Doklady Mathematics, 2023, 108:suppl. 2, S248–S256

Bibliographic databases:


© Steklov Math. Inst. of RAS, 2026