Scaling the Indian Buffet Process via Submodular Maximization
نویسندگان
چکیده
Inference for latent feature models is inherently difficult as the inference space grows exponentially with the size of the input data and number of latent features. In this work, we use Kurihara & Welling (2008)’s maximization-expectation framework to perform approximate MAP inference for linearGaussian latent feature models with an Indian Buffet Process (IBP) prior. This formulation yields a submodular function of the features that corresponds to a lower bound on the model evidence. By adding a constant to this function, we obtain a nonnegative submodular function that can be maximized via a greedy algorithm that obtains at least a 13 approximation to the optimal solution. Our inference method scales linearly with the size of the input data, and we show the efficacy of our method on the largest datasets currently analyzed using an IBP model.
منابع مشابه
Supplement to: Scaling the Indian Buffet Process via Submodular Maximization
Here we discuss the “shifted” equivalence class of binary matrices first proposed by Ding et al. (2010). For a given N ×K binary matrix Z, the equivalence class for this binary matrix [Z] is obtained by shifting allzero columns to the right of the non-zero columns while maintaining the non-zero column orderings, see Figure 1. Placing independent Beta( α K , 1) priors on the Bernoulli entries of...
متن کاملScaling Submodular Maximization via Pruned Submodularity Graphs
We propose a new randomized pruning method (called “submodular sparsification (SS)”) to reduce the cost of submodular maximization. The pruning is applied via a “submodularity graph” over the n ground elements, where each directed edge is associated with a pairwise dependency defined by the submodular function. In each step, SS prunes a 1 1/ p c (for c > 1) fraction of the nodes using weights o...
متن کاملMulti-document Summarization via Budgeted Maximization of Submodular Functions
We treat the text summarization problem as maximizing a submodular function under a budget constraint. We show, both theoretically and empirically, a modified greedy algorithm can efficiently solve the budgeted submodular maximization problem near-optimally, and we derive new approximation bounds in doing so. Experiments on DUC’04 task show that our approach is superior to the bestperforming me...
متن کاملLearning Sparse Combinatorial Representations via Two-stage Submodular Maximization
We consider the problem of learning sparse representations of data sets, where the goal is to reduce a data set in manner that optimizes multiple objectives. Motivated by applications of data summarization, we develop a new model which we refer to as the two-stage submodular maximization problem. This task can be viewed as a combinatorial analogue of representation learning problems such as dic...
متن کاملProbabilistic Submodular Maximization in Sub-Linear Time
In this paper, we consider optimizing submodular functions that are drawn from some unknown distribution. This setting arises, e.g., in recommender systems, where the utility of a subset of items may depend on a user-specific submodular utility function. In modern applications, the ground set of items is often so large that even the widely used (lazy) greedy algorithm is not efficient enough. A...
متن کامل