In this paper, a novel discrete-time iterative zero-sum adaptive dynamic programming(ADP) algorithm is developed for solving the optimal control problems of nonlinear systems. Two iteration processes, which are lower ...
详细信息
ISBN:
(纸本)9781509054626
In this paper, a novel discrete-time iterative zero-sum adaptive dynamic programming(ADP) algorithm is developed for solving the optimal control problems of nonlinear systems. Two iteration processes, which are lower and upper iterations, are employed to solve the lower and upper value functions, respectively. Arbitrary positive semi-definite functions are acceptable to initialize the upper and lower iterations of the iterative zero-sum ADP algorithm. It is proven that the upper and lower value functions converge to the optimal performance index function if the optimal performance index function exists, where the existence criterion of the optimal performance index function is unnecessary. Simulation examples are given to illustrate the effective performance of the present method.
An online adaptive dynamic programming (ADP) design is proposed for the control of urban open-channel flow systems, whose topographic parameters are not assumed to be accessible. According to the Saint-Venant continui...
详细信息
ISBN:
(纸本)9781509054015
An online adaptive dynamic programming (ADP) design is proposed for the control of urban open-channel flow systems, whose topographic parameters are not assumed to be accessible. According to the Saint-Venant continuity equation, a simplified model is firstly built. Subsequently, an adaptive dynamic programming control scheme is implemented, whose purpose is to track the desired water level, as well as to decrease the control cost. The design contains two RBF neural networks (NN). One action NN is employed to generate the control signal. Another critic NN is designed to approximate the long-term cost function. The two NNs are coordinated to approach an optimal solution. Finally, the adaptive dynamic programming controller is validated in rainstorm situation in simulation environment. The results demonstrate that the designed scheme outperforms the traditional PID counterpart.
In this paper, the robust control for a class of continuous-time nonlinear system with unmatched uncertainties is investigated using an event-triggered adaptive dynamic programming method. First, the robust control pr...
详细信息
ISBN:
(数字)9789811052309
ISBN:
(纸本)9789811052309;9789811052293
In this paper, the robust control for a class of continuous-time nonlinear system with unmatched uncertainties is investigated using an event-triggered adaptive dynamic programming method. First, the robust control problem is solved using the optimal control method. Under the event-triggered mechanism, the solution of the optimal control problem can asymptotically stabilize the uncertain system with an designed triggering condition. That is, the designed event-triggered controller is robust to the original uncertain system. Then, a single critic network structure with experience replay technique is constructed to approach the optimal control policies. Finally, a simulation example is provided to demonstrate the effectiveness of the proposed control scheme.
In this paper, we propose a data-driven adaptive dynamic programming approach to solve the Hamilton-Jacobi(HJ) equations for the two-player nonzero-sum(NZS) game with completely unknown dynamics. First, the model-base...
详细信息
ISBN:
(纸本)9781509046584
In this paper, we propose a data-driven adaptive dynamic programming approach to solve the Hamilton-Jacobi(HJ) equations for the two-player nonzero-sum(NZS) game with completely unknown dynamics. First, the model-based policy iteration(PI) algorithm is given, where the knowledge of system dynamics is required. To relax this requirement,a data-driven adaptive dynamic programming(ADP) is proposed in this paper to solve the unknown nonlinear NZS game with only online data. Neural network approximators are constructed to approach the solution of the HJ equations. The online data is collected under the two initial admissible control policies. Then, the NN weights are updated based on the least-squares method using the collected online data repeatedly, which is a kind of the off-policy learning ***, a simulation example is provided to demonstrate the effectiveness of the proposed control scheme.
This paper presents a decentralized optimal control method for modular and reconflgurable robots(MRRs) based on adaptivedynamic ***,the dynamic model of MRRs is formulated by using the Newton-Euler iterative algori...
详细信息
ISBN:
(纸本)9781538629185
This paper presents a decentralized optimal control method for modular and reconflgurable robots(MRRs) based on adaptivedynamic ***,the dynamic model of MRRs is formulated by using the Newton-Euler iterative algorithm,and then the state space description is ***,the optimal control policy of the MRRs system is obtained based on the policy iteration algorithm,which is used to solve the Hamilton-Jacobi-Bellman(HJB) equation via the critic neural ***,the stability of the closed-loop system is proved by using the Lyapunov ***,simulations are conducted to illustrate the effectiveness for the 2-DOF MRRs.
In this study, a nonquadratic performance function is introduced to overcome the saturation nonlinearity in actuators. Then a novel solution, generalized policy iteration adaptive dynamic programming algorithm, is app...
详细信息
ISBN:
(纸本)9783319590813;9783319590806
In this study, a nonquadratic performance function is introduced to overcome the saturation nonlinearity in actuators. Then a novel solution, generalized policy iteration adaptive dynamic programming algorithm, is applied to deal with the problem of optimal control. To achieve this goal, we use two neural networks to approximate control vectors and performance index function. Finally, this paper focuses on an example simulated on Matlab, which verifies the excellent convergence of the mentioned algorithm and feasibility of this scheme.
This paper presents a hybrid adaptive dynamic programming (hybrid-ADP) approach for determining the optimal continuous and discrete control laws of a switched system online, solely from state observations. The new hyb...
详细信息
This paper presents a hybrid adaptive dynamic programming (hybrid-ADP) approach for determining the optimal continuous and discrete control laws of a switched system online, solely from state observations. The new hybrid-ADP recurrence relationships presented are applicable to model-free control of switched hybrid systems that are possibly nonlinear. The computational complexity and convergence of the hybrid-ADP approach are analyzed, and the method is validated numerically showing that the optimal controller and value function can be learned iteratively online from state observations.
作者:
Bian, TaoJiang, Zhong-PingNYU
Tandon Sch Engn Dept Elect & Comp Engn Control & Networks LabMetrotech Ctr 5 Brooklyn NY 11201 USA
This paper presents a novel non-model-based, data-driven adaptive optimal controller design for linear continuous-time systems with completely unknown dynamics. Inspired by the stochastic approximation theory, a conti...
详细信息
This paper presents a novel non-model-based, data-driven adaptive optimal controller design for linear continuous-time systems with completely unknown dynamics. Inspired by the stochastic approximation theory, a continuous-time version of the traditional value iteration (VI) algorithm is presented with rigorous convergence analysis. This VI method is crucial for developing new adaptive dynamic programming methods to solve the adaptive optimal control problem and the stochastic robust optimal control problem for linear continuous-time systems. Fundamentally different from existing results, the a priori knowledge of an initial admissible control policy is no longer required. The efficacy of the proposed methodology is illustrated by two examples and a brief comparative study between VI and earlier policy iteration methods. (C) 2016 Elsevier Ltd. All rights reserved.
In this paper, the problem of H-infinity control design for affine nonlinear discrete-time systems is addressed by using adaptive dynamic programming (ADP). First, the nonlinear H-infinity control problem is transform...
详细信息
In this paper, the problem of H-infinity control design for affine nonlinear discrete-time systems is addressed by using adaptive dynamic programming (ADP). First, the nonlinear H-infinity control problem is transformed into solving the two-player zero-sum differential game problem of the nonlinear system. Then, the critic, action and disturbance networks are designed by using neural networks to solve online the Hamilton-Jacobi-Isaacs (HJI) equation associating with the two-player zero-sum differential game. When novel weight update laws for the critic, action and disturbance networks are tuned online by using data generated in real-time along the system trajectories, it is shown that the system states, all neural networks weight estimation errors are uniformly ultimately bounded by using Lyapunov techniques. Further, it is shown that the output of the action network approaches the optimal control input with small bounded error and the output of the disturbance network approaches the worst disturbance with small bounded error. At last, simulation results are presented to demonstrate the effectiveness of the new ADP based method. (C) 2016 Elsevier B.V. All rights reserved.
In this paper, a novel value iteration adaptive dynamic programming (ADP) algorithm, called "generalized value iteration ADP" algorithm, is developed to solve infinite horizon optimal tracking control proble...
详细信息
In this paper, a novel value iteration adaptive dynamic programming (ADP) algorithm, called "generalized value iteration ADP" algorithm, is developed to solve infinite horizon optimal tracking control problems for a class of discrete-time nonlinear systems. The developed generalized value iteration ADP algorithm permits an arbitrary positive semi-definite function to initialize it, which overcomes the disadvantage of traditional value iteration algorithms. Convergence property is developed to guarantee that the iterative performance index function will converge to the optimum. Neural networks are used to approximate the iterative performance index function and compute the iterative control policy, respectively, to implement the iterative ADP algorithm. Finally, a simulation example is given to illustrate the performance of the developed algorithm.
暂无评论