RUS  ENG
Full version
JOURNALS // Computer Research and Modeling // Archive

Computer Research and Modeling, 2024 Volume 16, Issue 7, Pages 1727–1746 (Mi crm1245)

SPECIAL ISSUE

On some mirror descent methods for strongly convex programming problems with Lipschitz functional constraints

O. S. Savchukab, M. S. Alkousaca, F. S. Stonyakinab

a Moscow Institute of Physics and Technology, 9 Institutskiy per., Dolgoprudny, Moscow region, 141701, Russia
b V. I. Vernadsky Crimean Federal University, 4 Academician Vernadsky ave., Simferopol, Republic of Crimea, 295007, Russia
c Innopolis University, 1 Universitetskaya st., Innopolis, 420500, Russia

Abstract: The paper is devoted to one approach to constructing subgradient methods for strongly convex programming problems with several functional constraints. More precisely, the strongly convex minimization problem with several strongly convex (inequality-type) constraints is considered, and first-order optimization methods for this class of problems are proposed. The special feature of the proposed methods is the possibility of using the strong convexity parameters of the violated functional constraints at nonproductive iterations, in theoretical estimates of the quality of the produced solution by the methods. The main task, to solve the considered problem, is to propose a subgradient method with adaptive rules for selecting steps and stopping rule of the method. The key idea of the proposed methods in this paper is to combine two approaches: a scheme with switching on productive and nonproductive steps and recently proposed modifications of mirror descent for convex programming problems, allowing to ignore some of the functional constraints on nonproductive steps of the algorithms. In the paper, it was described a subgradient method with switching by productive and nonproductive steps for strongly convex programming problems in the case where the objective function and functional constraints satisfy the Lipschitz condition. An analog of the proposed subgradient method, a mirror descent scheme for problems with relatively Lipschitz and relatively strongly convex objective functions and constraints is also considered. For the proposed methods, it obtained theoretical estimates of the quality of the solution, they indicate the optimality of these methods from the point of view of lower oracle estimates. In addition, since in many problems, the operation of finding the exact subgradient vector is quite expensive, then for the class of problems under consideration, analogs of the mentioned above methods with the replacement of the usual subgradient of the objective function or functional constraints by the $\delta$-subgradient were investigated. The noted approach can save computational costs of the method by refusing to require the availability of the exact value of the subgradient at the current point. It is shown that the quality estimates of the solution change by $O(\delta)$. The results of numerical experiments illustrating the advantages of the proposed methods in comparison with some previously known ones are also presented.

Keywords: subgradient method, mirror descent, strongly convex function, Lipschitz-continuous function, $\delta$-subgradient, productive step, nonproductive step

UDC: 519.8

Received: 29.10.2024
Revised: 13.11.2024
Accepted: 25.11.2024

DOI: 10.20537/2076-7633-2024-16-7-1727-1746



© Steklov Math. Inst. of RAS, 2026