In this paper, we present an online reinforcement learning algorithm for constrained Markov decision processes with a safety constraint. Despite the necessary attention of the scientific community, considering stochas...
详细信息
With the growing presence of semiconductor devices in healthcare, automotive, and consumer electronics, Automatic Test Equipment (ATE) systems play an increasingly vital role in ensuring quality and reliability during...
详细信息
The Multiport Autonomous Reconfigurable Solar Power Plant (MARS) is an integrated photovoltaic (PV) power generation and energy storage system (ESS), that is designed to connect to both alternating current (AC) transm...
详细信息
We address the problem of determining the least improbable deviations leading to an unsafe rare event in a weakly perturbed mechanical system with probabilistic initial conditions. These deviations are obtained as the...
详细信息
In this paper, we consider the problem of safety assessment for Markov decision processes without explicit knowledge of the model. We aim to learn probabilistic safety specifications associated with a given policy wit...
详细信息
ISBN:
(数字)9783907144107
ISBN:
(纸本)9798331540920
In this paper, we consider the problem of safety assessment for Markov decision processes without explicit knowledge of the model. We aim to learn probabilistic safety specifications associated with a given policy without compromising the safety of the process. To accomplish our goal, we characterize a subset of the state-space namely proxy set, which contains the states that are near in a probabilistic sense to the forbidden set consisting of all unsafe states. We compute the safety function using the single-step temporal difference method. To thi
s
end, we relate the safety function computation to that of the value function estimation using temporal difference learning. Since the given control policy could be unsafe, we use a safe baseline sub-policy to generate data for learning. We then use an off-policy temporal difference learning method with importance sampling to learn the safety function corresponding to the given policy. Finally, we demonstrate our results using a numerical example.
In this paper, we present an online reinforcement learning algorithm for constrained Markov decision processes with a safety constraint. Despite the necessary attention of the scientific community, considering stochas...
详细信息
ISBN:
(数字)9798350316339
ISBN:
(纸本)9798350316346
In this paper, we present an online reinforcement learning algorithm for constrained Markov decision processes with a safety constraint. Despite the necessary attention of the scientific community, considering stochastic stopping time, the problem of learning optimal policy without violating safety constraints during the learning phase is yet to be addressed. To this end, we propose an algorithm based on linear programming that does not require a process model. We show that the learned policy is safe with high confidence. We also propose a method to compute a safe baseline policy, which is central in developing algorithms that do not violate the safety constraints. Finally, we provide simulation results to show the efficacy of the proposed algorithm. Further, we demonstrate that efficient exploration can be achieved by defining a subset of the state-space called proxy set.
In this paper, we present an online reinforcement learning algorithm for constrained Markov decision processes with a safety constraint. Despite the necessary attention of the scientific community, considering stochas...
详细信息
In this paper, we aim to study safety specifications for a Markov decision process with stochastic stopping time in an almost model-free setting. Our approach involves characterizing a proxy set of the states that are...
In this paper, we aim to study safety specifications for a Markov decision process with stochastic stopping time in an almost model-free setting. Our approach involves characterizing a proxy set of the states that are near in a probabilistic sense to the set of unsafe states - forbidden set. We also provide results that relate safety function with reinforcement learning. Consequently, we develop an online algorithm based on the temporal difference method to compute the safety function. Finally, we provide simulation results that demonstrate our work in a simple example.
We propose a compositional framework for the stochastic safety of distributed Markov Decision Processes (MDPs) and Partially Observable Markov Decision Processes (POMDPs). We use MDP and POMDPs and their distributed v...
We propose a compositional framework for the stochastic safety of distributed Markov Decision Processes (MDPs) and Partially Observable Markov Decision Processes (POMDPs). We use MDP and POMDPs and their distributed versions as an appropriate modelling paradigm for computational ecosystems, understood in the context of distributed systems. We extend our work on stochastic safety from MDPs to POMDPs, and then to networked MDP/POMDPs. We propose a unifying mathematical framework for stochastic safety for MDPs, their partially observable version and their composition.
In this paper, we consider the problem of safety assessment for Markov decision processes without explicit knowledge of the model. We aim to learn probabilistic safety specifications associated with a given policy wit...
详细信息
暂无评论