Chance-Constrained Control With Lexicographic Deep Reinforcement Learning
نویسندگان
چکیده
منابع مشابه
Continuous control with deep reinforcement learning
We adapt the ideas underlying the success of Deep Q-Learning to the continuous action domain. We present an actor-critic, model-free algorithm based on the deterministic policy gradient that can operate over continuous action spaces. Using the same learning algorithm, network architecture and hyper-parameters, our algorithm robustly solves more than 20 simulated physics tasks, including classic...
متن کاملChance Constrained Quality Control
Consideration of quality implicitly introduces the need to adjust inputs in order to obtain desired output. Output evaluation must consider quality as well as cost. Real decisions may involve other objectives as well. Production problems often involve a dynamic situation where the relationship between cost and quality must be experimentally developed. The proposed method is to use regression as...
متن کاملBenchmarking Deep Reinforcement Learning for Continuous Control
Recently, researchers have made significant progress combining the advances in deep learning for learning feature representations with reinforcement learning. Some notable examples include training agents to play Atari games based on raw pixel data and to acquire advanced manipulation skills using raw sensory inputs. However, it has been difficult to quantify progress in the domain of continuou...
متن کاملStructured Control Nets for Deep Reinforcement Learning
In recent years, Deep Reinforcement Learning has made impressive advances in solving several important benchmark problems for sequential decision making. Many control applications use a generic multilayer perceptron (MLP) for non-vision parts of the policy network. In this work, we propose a new neural network architecture for the policy network representation that is simple yet effective. The ...
متن کاملChance Constrained Model Predictive Control
This work focuses on robustness of model predictive control (MPC) with respect to satisfaction of process output constraints. A method of improving such robustness is presented. The method relies on formulating output constraints as chance constraints using the uncertainty description of the process model. The resulting on-line optimization problem is convex. The proposed approach is illustrate...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Control Systems Letters
سال: 2020
ISSN: 2475-1456
DOI: 10.1109/lcsys.2020.2979635