
Backtracking linesearch is the de facto approach for minimizing continuously differentiable functions with locally Lipschitz gradient. In recent years, it has been shown that in the convex setting it is possible to avoid linesearch altogether, and to allow the stepsize to adapt based on a local smoothness estimate without any backtracks or evaluations of the function value. In this work we propose an adaptive proximal gradient method, adaPG, that uses novel estimates of the local smoothness modulus which leads to less conservative stepsize updates and that can additionally cope with nonsmooth terms. This idea is extended to the primal-dual setting where an adaptive three-term primal-dual algorithm, adaPD, is proposed which can be viewed as an extension of the PDHG method. Moreover, in this setting the "essentially" fully adaptive variant adaPD$^+$ is proposed that avoids evaluating the linear operator norm by invoking a backtracking procedure, that, remarkably, does not require extra gradient evaluations. Numerical simulations demonstrate the effectiveness of the proposed algorithms compared to the state of the art.
FOS: Computer and information sciences, 65K05, 90C06, 90C25, 90C30, 90C47, Technology, STEP-SIZE, Operations Research, Computer Science - Machine Learning, Mathematics, Applied, Convex minimization, MONOTONE INCLUSIONS, Locally Lipschitz gradient, SPLITTING ALGORITHM, Machine Learning (cs.LG), Primal-dual algorithms, 0102 Applied Mathematics, FOS: Mathematics, 4901 Applied mathematics, PRIMAL-DUAL ALGORITHM, Mathematics - Optimization and Control, 0802 Computation Theory and Mathematics, 4613 Theory of computation, Science & Technology, COMPOSITE, Operations Research & Management Science, 0103 Numerical and Computational Mathematics, SUM, Linesearch-free adaptive stepsizes, Computer Science, Software Engineering, Proximal gradient method, BARZILAI, Optimization and Control (math.OC), Physical Sciences, Computer Science, 4903 Numerical and computational mathematics, Mathematics, STADIUS-23-76
FOS: Computer and information sciences, 65K05, 90C06, 90C25, 90C30, 90C47, Technology, STEP-SIZE, Operations Research, Computer Science - Machine Learning, Mathematics, Applied, Convex minimization, MONOTONE INCLUSIONS, Locally Lipschitz gradient, SPLITTING ALGORITHM, Machine Learning (cs.LG), Primal-dual algorithms, 0102 Applied Mathematics, FOS: Mathematics, 4901 Applied mathematics, PRIMAL-DUAL ALGORITHM, Mathematics - Optimization and Control, 0802 Computation Theory and Mathematics, 4613 Theory of computation, Science & Technology, COMPOSITE, Operations Research & Management Science, 0103 Numerical and Computational Mathematics, SUM, Linesearch-free adaptive stepsizes, Computer Science, Software Engineering, Proximal gradient method, BARZILAI, Optimization and Control (math.OC), Physical Sciences, Computer Science, 4903 Numerical and computational mathematics, Mathematics, STADIUS-23-76
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 6 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
