A Differential Equation for Modeling Nesterov's Accelerated Gradient Method: Theory and Insights

نویسندگان

  • Weijie Su
  • Stephen P. Boyd
  • Emmanuel J. Candès
چکیده

We derive a second-order ordinary differential equation (ODE), which is the limitof Nesterov’s accelerated gradient method. This ODE exhibits approximate equiv-alence to Nesterov’s scheme and thus can serve as a tool for analysis. We show thatthe continuous time ODE allows for a better understanding of Nesterov’s scheme.As a byproduct, we obtain a family of schemes with similar convergence rates.The ODE interpretation also suggests restarting Nesterov’s scheme leading to analgorithm, which can be rigorously proven to converge at a linear rate wheneverthe objective is strongly convex.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Acceleration and Averaging in Stochastic Descent Dynamics

[1] Nemirovski and Yudin. Problems Complexity and Method Efficiency in Optimization. Wiley-Interscience series in discrete mathematics. Wiley, 1983. [2] W. Krichene, A. Bayen and P. Bartlett. Accelerated Mirror Descent in Continuous and Discrete Time. NIPS 2015. [3] W. Su, S. Boyd and E. Candes. A differential equation for modeling Nesterov's accelerated gradient method: theory and insights. NI...

متن کامل

A Variational Perspective on Accelerated Methods in Optimization

Accelerated gradient methods play a central role in optimization, achieving optimal rates in many settings. Although many generalizations and extensions of Nesterov's original acceleration method have been proposed, it is not yet clear what is the natural scope of the acceleration concept. In this paper, we study accelerated methods from a continuous-time perspective. We show that there is a La...

متن کامل

A geometric alternative to Nesterov's accelerated gradient descent

We propose a new method for unconstrained optimization of a smooth and strongly convex function, which attains the optimal rate of convergence of Nesterov’s accelerated gradient descent. The new algorithm has a simple geometric interpretation, loosely inspired by the ellipsoid method. We provide some numerical evidence that the new method can be superior to Nesterov’s accelerated gradient descent.

متن کامل

The Log-Exponential Smoothing Technique and Nesterov's Accelerated Gradient Method for Generalized Sylvester Problems

The Sylvester smallest enclosing circle problem involves finding the smallest circle that encloses a finite number of points in the plane. We consider generalized versions of the Sylvester problem in which the points are replaced by sets. Based on the log-exponential smoothing technique and Nesterov’s accelerated gradient method, we present an effective numerical algorithm for solving these pro...

متن کامل

An algebraic calculation method for describing time-dependent processes in electrochemistry – Expansion of existing procedures

In this paper an alternative model allowing the extension of the Debye-Hückel Theory (DHT) considering time dependence explicitly is presented. From the Electro-Quasistatic approach (EQS) introduced in earlier studies time dependent potentials are suitable to describe several phenomena especially conducting media as well as the behaviour of charged particles (ions) in electrolytes. This leads t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2014