نتایج جستجو برای: sequential decision making

تعداد نتایج: 618036  

2013
João V. Messias Matthijs T. J. Spaan Pedro U. Lima

Markov Decision Processes (MDPs) provide an extensive theoretical background for problems of decision-making under uncertainty. In order to maintain computational tractability, however, real-world problems are typically discretized in states and actions as well as in time. Assuming synchronous state transitions and actions at fixed rates may result in models which are not strictly Markovian, or...

2012
Thomas J. Walsh Sergiu Goschin

We describe theoretical bounds and a practical algorithm for teaching a model by demonstration in a sequential decision making environment. Unlike previous efforts that have optimized learners that watch a teacher demonstrate a static policy, we focus on the teacher as a decision maker who can dynamically choose different policies to teach different parts of the environment. We develop several ...

2004
Gangshu Cai

CAI, GANGSHU. Flexible Decision-Making in Sequential Auctions. (Under the direction of Assistant Professor Peter R. Wurman). Because sequential auctions have permeated society more than ever, it is desirable for participants to have the optimal strategies beforehand. However, finding closed-form solutions to various sequential auction games is challenging. Current literature provides some answe...

Journal: :CoRR 2018
Meik Dörpinghaus Izaak Neri Édgar Roldán Heinrich Meyr Frank Jülicher

This paper provides a statistical method to test whether a system that performs a binary sequential hypothesis test is optimal in the sense of minimizing the average decision times while taking decisions with given reliabilities. The proposed method requires samples of the decision times, the decision outcomes, and the true hypotheses, but does not require knowledge on the statistics of the obs...

2008
Daniel E. Acuña Paul R. Schrater

Studies of sequential decision-making in humans frequently find suboptimal performance relative to an ideal actor that has perfect knowledge of the model of how rewards and events are generated in the environment. Rather than being suboptimal, we argue that the learning problem humans face is more complex, in that it also involves learning the structure of reward generation in the environment. ...

Journal: :Neuron 2016
Michael N. Shadlen Daphna Shohamy

Decisions take time, and as a rule more difficult decisions take more time. But this only raises the question of what consumes the time. For decisions informed by a sequence of samples of evidence, the answer is straightforward: more samples are available with more time. Indeed, the speed and accuracy of such decisions are explained by the accumulation of evidence to a threshold or bound. Howev...

Journal: :Artif. Intell. 2011
Daniel Kikuti Fábio Gagliardi Cozman Ricardo Shirota Filho

This paper presents new insights and novel algorithms for strategy selection in sequential decision making with partially ordered preferences; that is, where some strategies may be incomparable with respect to expected utility. We assume that incomparability amongst strategies is caused by indeterminacy/imprecision in probability values. We investigate six criteria for consequentialist strategy...

2016
Markus Schöbel Jörg Rieskamp Rafael Huber Stephen C. Pratt

People often make decisions in a social environment. The present work examines social influence on people's decisions in a sequential decision-making situation. In the first experimental study, we implemented an information cascade paradigm, illustrating that people infer information from decisions of others and use this information to make their own decisions. We followed a cognitive modeling ...

2017
Duc Thien Nguyen Akshat Kumar Hoong Chuin Lau

Multiagent sequential decision making has seen rapid progress with formal models such as decentralized MDPs and POMDPs. However, scalability to large multiagent systems and applicability to real world problems remain limited. To address these challenges, we study multiagent planning problems where the collective behavior of a population of agents affects the joint-reward and environment dynamic...

2010
Sridhar Mahadevan

Automatically constructing novel representations of tasks from analysis of state spaces is a longstanding fundamental challenge in AI. I review recent progress on this problem for sequential decision making tasks modeled as Markov decision processes. Specifically, I discuss three classes of representation discovery problems: finding functional, state, and temporal abstractions. I describe solut...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید