Newton-like method with diagonal correction for distributed optimization
نویسندگان
چکیده
We consider distributed optimization problems where networked nodes cooperatively minimize the sum of their locally known convex costs. A popular class of methods to solve these problems are the distributed gradient methods, which are attractive due to their inexpensive iterations, but have a drawback of slow convergence rates. This motivates the incorporation of second-order information in the distributed methods, but this task is challenging: although the Hessians which arise in the algorithm design respect the sparsity of the network, their inverses are dense, hence rendering distributed implementations difficult. We overcome this challenge and propose a class of distributed Newton-like methods, which we refer to as Distributed Quasi Newton (DQN). The DQN family approximates the Hessian inverse by: 1) splitting the Hessian into its diagonal and off-diagonal part, 2) inverting the diagonal part, and 3) approximating the inverse of the off-diagonal part through a weighted linear function. The approximation is parameterized by the tuning variables which correspond to different splittings of the Hessian and by different weightings of the off-diagonal Hessian part. Specific choices of the tuning variables give rise to different variants of the proposed general DQN method – dubbed DQN-0, DQN-1 and DQN-2 – which mutually trade-off communication and computational costs for convergence. Simulations illustrate that the proposed DQN methods compare favorably with existing alternatives.
منابع مشابه
Distributed Sensor Selection using a Truncated Newton Method
We propose a new distributed algorithm for computing a truncated Newton method, where the main diagonal of the Hessian is computed using belief propagation. As a case study for this approach, we examine the sensor selection problem, a Boolean convex optimization problem. We form two distributed algorithms. The first algorithm is a distributed version of the interior point method by Joshi and Bo...
متن کاملDynamic scaling based preconditioning for truncated Newton methods in large scale unconstrained optimization
This paper deals with the preconditioning of truncated Newton methods for the solution of large scale nonlinear unconstrained optimization problems. We focus on preconditioners which can be naturally embedded in the framework of truncated Newton methods, i.e. which can be built without storing the Hessian matrix of the function to be minimized, but only based upon information on the Hessian obt...
متن کاملA New Load-Flow Method in Distribution Networks based on an Approximation Voltage-Dependent Load model in Extensive Presence of Distributed Generation Sources
Power-flow (PF) solution is a basic and powerful tool in power system analysis. Distribution networks (DNs), compared to transmission systems, have many fundamental distinctions that cause the conventional PF to be ineffective on these networks. This paper presents a new fast and efficient PF method which provides all different models of Distributed Generations (DGs) and their operational modes...
متن کاملOptimization with the time-dependent Navier-Stokes equations as constraints
In this paper, optimal distributed control of the time-dependent Navier-Stokes equations is considered. The control problem involves the minimization of a measure of the distance between the velocity field and a given target velocity field. A mixed numerical method involving a quasi-Newton algorithm, a novel calculation of the gradients and an inhomogeneous Navier-Stokes solver, to find the opt...
متن کاملAn efficient improvement of the Newton method for solving nonconvex optimization problems
Newton method is one of the most famous numerical methods among the line search methods to minimize functions. It is well known that the search direction and step length play important roles in this class of methods to solve optimization problems. In this investigation, a new modification of the Newton method to solve unconstrained optimization problems is presented. The significant ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- SIAM Journal on Optimization
دوره 27 شماره
صفحات -
تاریخ انتشار 2017