نتایج جستجو برای: online decision problem
تعداد نتایج: 1388513 فیلتر نتایج به سال:
We consider a Markov decision process (MDP) setting in which the reward function is allowed to change after each time step (possibly in an adversarial manner), yet the dynamics remain fixed. Similar to the experts setting, we address the question of how well an agent can do when compared to the reward achieved under the best stationary policy over time. We provide efficient algorithms, which ha...
optimal adaptive leader-follower consensus of linear multi-agent systems: known and unknown dynamics
in this paper, the optimal adaptive leader-follower consensus of linear continuous time multi-agent systems is considered. the error dynamics of each player depends on its neighbors’ information. detailed analysis of online optimal leader-follower consensus under known and unknown dynamics is presented. the introduced reinforcement learning-based algorithms learn online the approximate solution...
This study develops a Web-based collaborative system framework based on a multiple perspective approach. This framework is a recent decision support system (DSS) paradigm proposed by Courtney [Decis. Support Syst. 31 (2001) 17] for knowledge management of and decision making about a special organizational problem. It consists of four main components. The first component is a group decision-maki...
Users of E-Sales platforms typically face the problem of choosing the most suitable product or service from large and potentially complex assortments. Whereas the problem of finding and presenting suitable items fulfilling the user’s requirements can be tackled by providing additional support in the form of recommenderand configuration systems, the control of psychological side e↵ects resulting...
Online mechanism design (MD) considers the problem of providing incentives to implement desired system-wide outcomes in systems with self-interested agents that arrive and depart dynamically. Agents can choose to misrepresent their arrival and departure times, in addition to information about their value for different outcomes. We consider the problem of maximizing the total longterm value of t...
In this paper, we propose a Markov Decision Process model for an empty repositioning problem in a two-port system. We consider two cases. The first case is the offline case, where demand information is assumed as a random variable with known distribution. The second case is online case where demand information is partially known. In both cases, we figure out the optimal controlling policies. T...
This study provides an exploratory model to understand the factors that influence consumers to adopt the internet instead of traditional channels for information search and product purchase. The authors reviewed previous established theories on consumer decision making in offline environments and research findings regarding consumer behaviour in an online environment. The authors embraced the c...
We consider an online decision problem over a discrete space in which the loss function is submodular. We give algorithms which are computationally efficient and are Hannan-consistent in both the full information and bandit settings.
Bandit algorithms are widely used in sequential decision problems to maximize the cumulative reward. One potential application is mobile health, where goal promote user's health through personalized interventions based on user specific information acquired wearable devices. Important considerations include type of, and frequency with which data collected (e.g. GPS, or continuous monitoring), as...
With the technology advances, online shopping has experienced a phenomenal growth. In line with such phenomenon and its relevancy, considerable number of studies have shown an interest within this area. Although recent researches particularly addressed consumer’s behavior, findings were inconsistent. Thereby, further been called for. The present study aims to investigate effect few variables de...
نمودار تعداد نتایج جستجو در هر سال
با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید