Equivalence of Entropy Regularization and Relative-Entropy Proximal Method

نویسنده

  • Jason K. Johnson
چکیده

We consider two entropy-based interior point methods that solve LP relaxations of MAP estimation in graphical models: (1) an entropy-regularization method and (2) a relative-entropy proximal method. Using the fact that relative-entropy is the Bregman distance induced by entropy, we show that the two approaches are actually equivalent. The purpose of this note is to show one connection between the two approaches described in [1, 2]. Another connection between these two works is that both use distributed iterative-scaling/Bregman-projections algorithms [3, 4] to solve the “inner-loop” optimizations required by the methods summarized below. This second connection, however, is not explored in this present note. Introduction For the sake of this note, we consider the exponential family of probability distributions on n binary variables x = (x1, . . . , xn) ∈ {0, 1} : P (x) = exp{θφ(x) − Φ(θ)} = 1 Z(θ) exp{ ∑

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Observational Modeling of the Kolmogorov-Sinai Entropy

In this paper, Kolmogorov-Sinai entropy is studied using mathematical modeling of an observer $ Theta $. The relative entropy of a sub-$ sigma_Theta $-algebra having finite atoms is defined and then   the ergodic properties of relative  semi-dynamical systems are investigated.  Also,  a relative version of Kolmogorov-Sinai theorem  is given. Finally, it is proved  that the relative entropy of a...

متن کامل

Some properties of the parametric relative operator entropy

The notion of entropy was introduced by Clausius in 1850, and some of the main steps towards the consolidation of the concept were taken by Boltzmann and Gibbs. Since then several extensions and reformulations have been developed in various disciplines with motivations and applications in different subjects, such as statistical mechanics, information theory, and dynamical systems. Fujii and Kam...

متن کامل

A short variational proof of equivalence between policy gradients and soft Q learning

Two main families of reinforcement learning algorithms, Q-learning and policy gradients, have recently been proven to be equivalent when using a softmax relaxation on one part, and an entropic regularization on the other. We relate this result to the well-known convex duality of Shannon entropy and the softmax function. Such a result is also known as the Donsker-Varadhan formula. This provides ...

متن کامل

Relative Entropy Rate between a Markov Chain and Its Corresponding Hidden Markov Chain

 In this paper we study the relative entropy rate between a homogeneous Markov chain and a hidden Markov chain defined by observing the output of a discrete stochastic channel whose input is the finite state space homogeneous stationary Markov chain. For this purpose, we obtain the relative entropy between two finite subsequences of above mentioned chains with the help of the definition of...

متن کامل

A note on inequalities for Tsallis relative operator entropy

‎In this short note‎, ‎we present some inequalities for relative operator entropy which are generalizations of some results obtained by Zou [Operator inequalities associated with Tsallis relative operator entropy‎, ‎{em Math‎. ‎Inequal‎. ‎Appl.}‎ ‎{18} (2015)‎, ‎no‎. ‎2‎, ‎401--406]‎. ‎Meanwhile‎, ‎we also show some new lower and upper bounds for relative operator entropy and Tsallis relative o...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008