Graphical Newton for Huge-Block Coordinate Descent on Sparse Graphs
نویسندگان
چکیده
Block coordinate descent (BCD) methods are often very effective for optimization problems where dependencies between variables are sparse. These methods can make a substantial amount of progress by applying Newton’s method to update a block of variables, but this leads to an iteration cost of O(|b|) in terms of the block size |b|. In this paper, we show how to use message-passing to compute the Newton step in O(|b|) when the block has a forest-structured dependency. Consequently, this allows us to update huge blocks for sparse problems, resulting in significant numerical improvements over existing approaches. We also present a greedy approach for selecting forest-structured blocks.
منابع مشابه
Let’s Make Block Coordinate Descent Go Fast: Faster Greedy Rules, Message-Passing, Active-Set Complexity, and Superlinear Convergence
Block coordinate descent (BCD) methods are widely-used for large-scale numerical optimization because of their cheap iteration costs, low memory requirements, amenability to parallelization, and ability to exploit problem structure. Three main algorithmic choices influence the performance of BCD methods: the block partitioning strategy, the block selection rule, and the block update rule. In th...
متن کاملLarge-Scale Optimization Algorithms for Sparse Conditional Gaussian Graphical Models
This paper addresses the problem of scalable optimization for l1-regularized conditional Gaussian graphical models. Conditional Gaussian graphical models generalize the well-known Gaussian graphical models to conditional distributions to model the output network influenced by conditioning input variables. While highly scalable optimization methods exist for sparse Gaussian graphical model estim...
متن کاملMaximum likelihood estimation of Gaussian graphical models: Numerical implementation and topology selection
We describe algorithms for maximum likelihood estimation of Gaussian graphical models with conditional independence constraints. It is well-known that this problem can be formulated as an unconstrained convex optimization problem, and that it has a closed-form solution if the underlying graph is chordal. The focus of this paper is on numerical algorithms for large problems with non-chordal grap...
متن کاملComplexity of Inexact Proximal Newton methods
Recently several, so-called, proximal Newton methods were proposed for sparse optimization [6, 11, 8, 3]. These methods construct a composite quadratic approximation using Hessian information, optimize this approximation using a first-order method, such as coordinate descent and employ a line search to ensure sufficient descent. Here we propose a general framework, which includes slightly modif...
متن کاملPascal Challenge: Linear Support Vector Machines
ξ(w; xi, yi) = max(1− yi(w xi), 0), We consider the LIBLINEAR package (available at http://www.csie.ntu.edu.tw/~cjlin/ liblinear), which can handle L1and L2-loss linear SVMs. The L1-SVM solver implemented in LIBLINEAR employes a coordinate descent method to solve the dual problem. Details are in [1]. This method is very useful for large sparse data with a huge number of instances and features. ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017