نتایج جستجو برای: partially observable markov decision process

تعداد نتایج: 1776231  

2010
Christopher Amato Dan Bernstein Alan Carlin Marek Petrik Akshat Kumar Sven Seuken Siddharth Srivastava Raphen Becker Jiaying Shen

INCREASING SCALABILITY IN ALGORITHMS FOR CENTRALIZED AND DECENTRALIZED PARTIALLY OBSERVABLE MARKOV DECISION PROCESSES: EFFICIENT DECISION-MAKING AND COORDINATION IN UNCERTAIN ENVIRONMENTS

2000
Valentina Bayer Zubek

This paper introduces the even-odd POMDP, an approximation to POMDPs (Partially Observable Markov Decision Problems) in which the world is assumed to be fully observable every other time step. This approximation works well for problems with a delayed need to observe. The even-odd POMDP can be converted into an equivalent MDP, the 2MDP, whose value function, V 2MDP , can be combined online with ...

2008
Debajyoti Ray Brooks King-Casas P. Read Montague Peter Dayan

Classical game theoretic approaches that make strong rationality assumptions have difficulty modeling human behaviour in economic games. We investigate the role of finite levels of iterated reasoning and non-selfish utility functions in a Partially Observable Markov Decision Process model that incorporates game theoretic notions of interactivity. Our generative model captures a broad class of c...

2016
Ping Hou William Yeoh Pradeep Varakantham

Partially Observable Markov Decision Processes (POMDPs) are often used to model planning problems under uncertainty. The goal in Risk-Sensitive POMDPs (RS-POMDPs) is to find a policy that maximizes the probability that the cumulative cost is within some user-defined cost threshold. In this paper, unlike existing POMDP literature, we distinguish between the two cases of whether costs can or cann...

2006
Trung H. BUI Mannes POEL Anton NIJHOLT

We propose a novel approach to developing a dialogue model that is able to take into account some aspects of the user's affective state and to act appropriately. Our dialogue model uses a Partially Observable Markov Decision Process approach with observations composed of the observed user's affective state and action. A simple example of route navigation is explained to clarify our approach. Th...

2006
Zinovi Rabinovich Jeffrey S. Rosenschein

In this paper we introduce a novel approach to continual planning and control, called Dynamics Based Control (DBC). The approach is similar in spirit to the Actor-Critic [6] approach to learning and estimation-based differential regulators of classical control theory [13]. However, DBC is not a learning algorithm, nor can it be subsumed within models of standard control theory. We provide a gen...

Journal: :J. Artif. Intell. Res. 2001
Nevin Lianwen Zhang Weihong Zhang

Partially observable Markov decision processes (POMDPs) have recently become popular among many AI researchers because they serve as a natural model for planning under uncertainty. Value iteration is a well-known algorithm for nding optimal policies for POMDPs. It typically takes a large number of iterations to converge. This paper proposes a method for accelerating the convergence of value ite...

2008
James Henderson Oliver Lemon

In spoken dialogue systems, Partially Observable Markov Decision Processes (POMDPs) provide a formal framework for making dialogue management decisions under uncertainty, but efficiency and interpretability considerations mean that most current statistical dialogue managers are only MDPs. These MDP systems encode uncertainty explicitly in a single state representation. We formalise such MDP sta...

2003
Amalia F. Foka Panos E. Trahanias

This paper introduces a methodology for avoiding obstacles by controlling the robot’s velocity. Contemporary approaches to obstacle avoidance usually dictate a detour from the originally planned trajectory to its goal position. In our previous work, we presented a method for predicting the motion of obstacles, and how to make use of this prediction when planning the robot trajectory to its goal...

2014
Milos Hauskrecht Hamish Fraser

Diagnosis of a disease and its treatment are not separate, oneshot activities. Instead they are very often dependent and interleaved over time, mostly due to uncertainty about the underlying disease, uncertainty associated with the response of a patient to the treatment and varying cost of different treatment and diagnostic (investigative) procedures. The framework particularly suitable for mod...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید