نتایج جستجو برای: partially observable markov decision process

تعداد نتایج: 1776231  

1999
Omid Madani Steve Hanks Anne Condon

We investigate the computability of problems in probabilistic planning and partially observable innnite-horizon Markov decision processes. The undecidability of the string-existence problem for probabilistic nite automata is adapted to show that the following problem of plan existence in probabilistic planning is undecidable: given a probabilistic planning problem, determine whether there exist...

Journal: :Systems and Computers in Japan 2004
Yoshihide Yamashiro Atsushi Ueno Hideaki Takeda

Reinforcement learning often involves assuming Markov characteristics. However, the agent cannot always observe the environment completely, and in such cases, different states are observed as the same state. In this research, the authors develop a Delayed Reward-based Genetic Algorithm for POMDP (DRGA) as a means to solve a partially observable Markov decision problem (POMDP) which has such per...

1999
Omid Madani Steve Hanks Anne Condon

We investigate the computability of problems in probabilistic planning and partially observable innnite-horizon Markov decision processes. The undecidability of the string-existence problem for probabilistic nite automata is adapted to show that the following problem of plan existence in probabilistic planning is undecidable: given a probabilistic planning problem, determine whether there exist...

2003
Nicolas Meuleau David E. Smith

For a given problem, the optimal Markov policy over a finite horizon is a conditional plan containing a potentially large number of branches. However, there are applications where it is desirable to strictly limit the number of decision points and branches in a plan. This raises the question of how one goes about finding optimal plans containing only a limited number of branches. In this paper,...

2014
Gavin Rens Thomas Andreas Meyer Gerhard Lakemeyer

We present a logic inspired by partially observable Markov decision process (POMDP) theory for specifying agent domains where the agent’s actuators and sensors are noisy (causing uncertainty). The language features modalities for actions and predicates for observations. It includes a notion of probability to represent the uncertainties, and the expression of rewards and costs are also catered f...

Journal: :JCM 2014
Shibing Zhang Huijian Wang Xiaoge Zhang

—Prediction of spectrum sensing and access is one of the keys in cognitive radio (CR). It is necessary to know the channel state transition probabilities to predict the spectrum. By the use of the model of partially observable Markov decision process (POMDP), this paper addressed the spectrum sensing and access in cognitive radio and proposed an estimation algorithm of channel state transition...

1998
Niels PEEK

The planning of clinical treatment actions for children with congenital heart disease requires a subtle trade-o between their immediate and long-term consequences, where most of these consequences cannot be predicted with certainty. It is described how this problem can be cast as a nite-horizon, partially observable Markov decision process. The complexity of the resulting model is reduced by us...

2005
Michael Jonas James G. Schmolze

For over a decade, the Hidden Markov Model (HMM) has been the primary tool used for acoustic modeling in the field of speech recognition. In this paper we examine a more general approach using a Partially Observable Markov Decision Process (POMDP) to mo del the base phonetic unit. We introduce the concept of multiple phonetic context classes, one for each of the infinite possible contexts a pho...

Journal: :Revue d'Intelligence Artificielle 2007
Vincent Thomas Christine Bourjot Vincent Chevrier

RÉSUMÉ. Cet article se focalise sur des approches formelles pour la construction de systèmes multi-agents. Ce travail a cherché à proposer des apprentissages décentralisés pour construire les comportements d’agents sociaux. Cet article propose un formalisme original, l’interacDEC-POMDP inspiré des modèles markoviens au sein duquel les agents peuvent interagir directement et localement entre eux...

2007
Marc G. Bellemare Doina Precup

Markov models have been a keystone in Artificial Intelligence for many decades. However, they remain unsatisfactory when the environment modelled is partially observable. There are pathological examples where no history of fixed length is sufficient for accurate prediction or decision making. On the other hand, working with a hidden state (like in Hidden Markov Models or Partially Observable Ma...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید