Experimental Design for Partially Observed Markov Decision Processes

نویسندگان
چکیده

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Partially observable Markov decision processes

For reinforcement learning in environments in which an agent has access to a reliable state signal, methods based on the Markov decision process (MDP) have had many successes. In many problem domains, however, an agent suffers from limited sensing capabilities that preclude it from recovering a Markovian state signal from its perceptions. Extending the MDP framework, partially observable Markov...

متن کامل

Introduction to Pomp: Inference for Partially-observed Markov Processes

1. Partially-observed Markov processes 1 2. A first example: a discrete-time bivariate autoregressive process. 3 3. Defining a partially observed Markov process in pomp. 3 4. Simulating the model 5 5. Computing likelihood using particle filtering 6 6. Interlude: utility functions for extracting and changing pieces of a pomp object 9 7. Estimating parameters using iterated filtering: mif 10 8. N...

متن کامل

A Partially Observed Markov Decision Process for Dynamic Pricing

In this paper, we develop a stylized partially observed Markov decision process (POMDP) framework, to study a dynamic pricing problem faced by sellers of fashion-like goods. We consider a retailer that plans to sell a given stock of items during a finite sales season. The objective of the retailer is to dynamically price the product in a way that maximizes expected revenues. Our model brings to...

متن کامل

Finite Model Approximations for Partially Observed Markov Decision Processes with Discounted Cost

We consider finite model approximations of discretetime partially observed Markov decision processes (POMDPs) under the discounted cost criterion. After converting the original partially observed stochastic control problem to a fully observed one on the belief space, the finite models are obtained through the uniform quantization of the state and action spaces of the belief space Markov decisio...

متن کامل

Bounded-Parameter Partially Observable Markov Decision Processes

The POMDP is considered as a powerful model for planning under uncertainty. However, it is usually impractical to employ a POMDP with exact parameters to model precisely the real-life situations, due to various reasons such as limited data for learning the model, etc. In this paper, assuming that the parameters of POMDPs are imprecise but bounded, we formulate the framework of bounded-parameter...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: SIAM/ASA Journal on Uncertainty Quantification

سال: 2018

ISSN: 2166-2525

DOI: 10.1137/16m1084924