Minimum information divergence of Q-functions for dynamic treatment resumes

نویسندگان

چکیده

This paper aims at presenting a new application of information geometry to reinforcement learning focusing on dynamic treatment resumes. In standard framework learning, Q-function is defined as the conditional expectation reward given state and an action for single-stage situation. We introduce equivalence relation, called policy equivalence, in space all Q-functions. A class divergence every stage. The main objective propose estimator optimal function by method minimum based dataset trajectories. particular, we discuss $$\gamma $$ -power that shown have advantageous property such between policy-equivalent Q-functions vanishes. essentially works seek policy, which discussed semiparametric model Q-function. specific choices power index give interesting relationships value function, geometric harmonic means numerical experiment demonstrates performance context regimes.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Information geometry of divergence functions

Measures of divergence between two points play a key role in many engineering problems. One such measure is a distance function, but there are many important measures which do not satisfy the properties of the distance. The Bregman divergence, KullbackLeibler divergence and f -divergence are such measures. In the present article, we study the differential-geometrical structure of a manifold ind...

متن کامل

diagnostic and developmental potentials of dynamic assessment for writing skill

این پایان نامه بدنبال بررسی کاربرد ارزیابی مستمر در یک محیط یادگیری زبان دوم از طریق طرح چهار سوال تحقیق زیر بود: (1) درک توانایی های فراگیران زمانیکه که از طریق برآورد عملکرد مستقل آنها امکان پذیر نباشد اما در طول جلسات ارزیابی مستمر مشخص شوند; (2) امکان تقویت توانایی های فراگیران از طریق ارزیابی مستمر; (3) سودمندی ارزیابی مستمر در هدایت آموزش فردی به سمتی که به منطقه ی تقریبی رشد افراد حساس ا...

15 صفحه اول

Minimum Dynamic Discrimination Information Models

In this paper, we introduce the minimum dynamic discrimination information (MDDI) approach to probability modeling. The MDDI model relative to a given distributionG is that which has least Kullback–Leibler information discrepancy relative to G, among all distributions satisfying some information constraints given in terms of residual moment inequalities, residualmoment growth inequalities, or h...

متن کامل

Minimum Divergence

This paper studies the Minimum Divergence (MD) class of estimators for econometric models specified through moment restrictions. We show that MD estimators can be obtained as solutions to a computationally tractable optimization problem. This problem is similar to the one solved by the Generalized Empirical Likelihood estimators of Newey and Smith (2004), but it is equivalent to it only for a s...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Information geometry

سال: 2022

ISSN: ['2511-2481', '2511-249X']

DOI: https://doi.org/10.1007/s41884-022-00084-8