
arXiv: 2109.02093
This paper proposes and justifies two globally convergent Newton-type methods to solve unconstrained and constrained problems of nonsmooth optimization by using tools of variational analysis and generalized differentiation. Both methods are coderivative-based and employ generalized Hessians (coderivatives of subgradient mappings) associated with objective functions, which are either of class $\mathcal{C}^{1,1}$, or are represented in the form of convex composite optimization, where one of the terms may be extended-real-valued. The proposed globally convergent algorithms are of two types. The first one extends the damped Newton method and requires positive-definiteness of the generalized Hessians for its well-posedness and efficient performance, while the other algorithm is of {the regularized Newton type} being well-defined when the generalized Hessians are merely positive-semidefinite. The obtained convergence rates for both methods are at least linear, but become superlinear under the semismooth$^*$ property of subgradient mappings. Problems of convex composite optimization are investigated with and without the strong convexity assumption {on smooth parts} of objective functions by implementing the machinery of forward-backward envelopes. Numerical experiments are conducted for Lasso problems and for box constrained quadratic programs with providing performance comparisons of the new algorithms and some other first-order and second-order methods that are highly recognized in nonsmooth optimization.
arXiv admin note: text overlap with arXiv:2101.10555
convex composite optimization, Nonsmooth analysis, Lasso problems, nonsmooth optimization, linear and superlinear convergence rates, global convergence, variational analysis, Optimization and Control (math.OC), Sensitivity, stability, parametric optimization, 90C31, 49J52, 49J53, FOS: Mathematics, generalized Newton methods, Mathematics - Optimization and Control, Set-valued and variational analysis
convex composite optimization, Nonsmooth analysis, Lasso problems, nonsmooth optimization, linear and superlinear convergence rates, global convergence, variational analysis, Optimization and Control (math.OC), Sensitivity, stability, parametric optimization, 90C31, 49J52, 49J53, FOS: Mathematics, generalized Newton methods, Mathematics - Optimization and Control, Set-valued and variational analysis
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 17 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Top 10% | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Top 10% | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Top 10% |
