Distributed deep reinforcement learning for simulation control

نویسندگان

چکیده

Abstract Several applications in the scientific simulation of physical systems can be formulated as control/optimization problems. The computational models for such generally contain hyperparameters, which control solution fidelity and expense. tuning these parameters is non-trivial general approach to manually ‘spot-check’ good combinations. This because optimal hyperparameter configuration search becomes intractable when parameter space large they may vary dynamically. To address this issue, we present a framework based on deep reinforcement learning (RL) train neural network agent that controls model solve by varying First, validate our RL problem controlling chaos chaotic dynamically changing system. Subsequently, illustrate capabilities accelerating convergence steady-state fluid dynamics solver automatically adjusting relaxation factors discretized Navier–Stokes equations during run-time. results indicate run-time learned policy leads significant reduction number iterations compared random selection factors. Our point potential benefits adaptive strategies across different geometries boundary conditions with implications reduced campaign expenses 4 Data codes available at https://github.com/Romit-Maulik/PAR-RL .

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Benchmarking Deep Reinforcement Learning for Continuous Control

Recently, researchers have made significant progress combining the advances in deep learning for learning feature representations with reinforcement learning. Some notable examples include training agents to play Atari games based on raw pixel data and to acquire advanced manipulation skills using raw sensory inputs. However, it has been difficult to quantify progress in the domain of continuou...

متن کامل

Structured Control Nets for Deep Reinforcement Learning

In recent years, Deep Reinforcement Learning has made impressive advances in solving several important benchmark problems for sequential decision making. Many control applications use a generic multilayer perceptron (MLP) for non-vision parts of the policy network. In this work, we propose a new neural network architecture for the policy network representation that is simple yet effective. The ...

متن کامل

Diff-DAC: Distributed Actor-Critic for Multitask Deep Reinforcement Learning

We propose a multiagent distributed actor-critic algorithm for multitask reinforcement learning (MRL), named Diff-DAC. The agents are connected, forming a (possibly sparse) network. Each agent is assigned a task and has access to data from this local task only. During the learning process, the agents are able to communicate some parameters to their neighbors. Since the agents incorporate their ...

متن کامل

Continuous control with deep reinforcement learning

We adapt the ideas underlying the success of Deep Q-Learning to the continuous action domain. We present an actor-critic, model-free algorithm based on the deterministic policy gradient that can operate over continuous action spaces. Using the same learning algorithm, network architecture and hyper-parameters, our algorithm robustly solves more than 20 simulated physics tasks, including classic...

متن کامل

Model-free Control for Distributed Stream Data Processing using Deep Reinforcement Learning

In this paper, we focus on general-purposeDistributed Stream Data Processing Systems (DSDPSs), which deal with processing of unbounded streams of continuous data at scale distributedly in real or near-real time. A fundamental problem in a DSDPS is the scheduling problem (i.e., assigning workload to workers/machines) with the objective of minimizing average end-to-end tuple processing time. A wi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Machine learning: science and technology

سال: 2021

ISSN: ['2632-2153']

DOI: https://doi.org/10.1088/2632-2153/abdaf8