Improved bounds for discretization of Langevin diffusions: Near-optimal rates without convexity

نویسندگان

چکیده

Discretizations of the Langevin diffusion have been proven very useful for developing and analyzing algorithms sampling stochastic optimization. We present an improved non-asymptotic analysis Euler-Maruyama discretization diffusion. Our does not require global contractivity, yields polynomial dependence on time horizon. Compared to existing approaches, we make additional smoothness assumption, improve rate in step size from O(η) O(η2) terms KL divergence. This result matches correct order numerical SDEs, without suffering exponential dependence. When applied MCMC, this simultaneously improves analyses a range that are based Dalalyan’s approach.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Optimal scaling of discrete approximations to Langevin diffusions

We consider the optimal scaling problem for proposal distributions in Hastings-Metropolis algorithms derived from Langevin diffusions. We prove an asymptotic diffusion limit theorem and show that the relative efficiency of the algorithm can be characterised by its overall acceptance rate, independently of the target distribution. The asymptotically optimal acceptance rate is 0.574. We show that...

متن کامل

Tempered Langevin diffusions and algorithms

We consider a class of Langevin diffusions with state-dependent volatility. The volatility of the diffusion is chosen so as to make the stationary distribution of the diffusion with respect to its natural clock, a heated version of the stationary density of interest. The motivation behind this construction is the desire to construct uniformly ergodic diffusions with required stationary densitie...

متن کامل

Near-optimal Regret Bounds for Reinforcement Learning Near-optimal Regret Bounds for Reinforcement Learning

This technical report is an extended version of [1]. For undiscounted reinforcement learning in Markov decision processes (MDPs) we consider the total regret of a learning algorithm with respect to an optimal policy. In order to describe the transition structure of an MDP we propose a new parameter: An MDP has diameter D if for any pair of states s, s there is a policy which moves from s to s i...

متن کامل

Optimal A Priori Discretization Error Bounds for Geodesic Finite Elements

We prove optimal bounds for the discretization error of geodesic finite elements for variational partial differential equations for functions that map into a nonlinear space. For this we first generalize the well-known Céa lemma to nonlinear function spaces. In a second step we prove optimal interpolation error estimates for pointwise interpolation by geodesic finite elements of arbitrary order...

متن کامل

Near-optimal Regret Bounds for Reinforcement Learning

For undiscounted reinforcement learning in Markov decision processes (MDPs) we consider the total regret of a learning algorithm with respect to an optimal policy. In order to describe the transition structure of an MDP we propose a new parameter: An MDP has diameter D if for any pair of states s, s′ there is a policy which moves from s to s′ in at most D steps (on average). We present a reinfo...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Bernoulli

سال: 2022

ISSN: ['1573-9759', '1350-7265']

DOI: https://doi.org/10.3150/21-bej1343