Linear convergence of inexact descent method and inexact proximal gradient algorithms for lower-order regularization problems
نویسندگان
چکیده
منابع مشابه
Convergence Rates of Inexact Proximal-Gradient Methods for Convex Optimization
We consider the problem of optimizing the sum of a smooth convex function and a non-smooth convex function using proximal-gradient methods, where an error is present in the calculation of the gradient of the smooth term or in the proximity operator with respect to the non-smooth term. We show that both the basic proximal-gradient method and the accelerated proximal-gradient method achieve the s...
متن کاملEfficient Inexact Proximal Gradient Algorithm for Nonconvex Problems
The proximal gradient algorithm has been popularly used for convex optimization. Recently, it has also been extended for nonconvex problems, and the current state-of-the-art is the nonmonotone accelerated proximal gradient algorithm. However, it typically requires two exact proximal steps in each iteration, and can be inefficient when the proximal step is expensive. In this paper, we propose an...
متن کاملInexact proximal stochastic gradient method for convex composite optimization
We study an inexact proximal stochastic gradient (IPSG) method for convex composite optimization, whose objective function is a summation of an average of a large number of smooth convex functions and a convex, but possibly nonsmooth, function. Variance reduction techniques are incorporated in the method to reduce the stochastic gradient variance. The main feature of this IPSG algorithm is to a...
متن کاملRelatively Inexact Proximal Point Algorithm and Linear Convergence Analysis
Based on a notion of relatively maximal m -relaxed monotonicity, the approximation solvability of a general class of inclusion problems is discussed, while generalizing Rockafellar’s theorem 1976 on linear convergence using the proximal point algorithm in a real Hilbert space setting. Convergence analysis, based on this newmodel, is simpler and compact than that of the celebrated technique of R...
متن کاملDescentwise inexact proximal algorithms for smooth optimization
The proximal method is a standard regularization approach in optimization. Practical implementations of this algorithm require (i) an algorithm to compute the proximal point, (ii) a rule to stop this algorithm, (iii) an update formula for the proximal parameter. In this work we focus on (ii), when smoothness is present – so that Newton-like methods can be used for (i): we aim at giving adequate...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of Global Optimization
سال: 2020
ISSN: 0925-5001,1573-2916
DOI: 10.1007/s10898-020-00955-3