This paper concerns two-player zero-sum stochastic differential games with nonanticipative strategies against closed-loop controls in the case where the coefficients of mean-field stochastic differential equations and...
详细信息
This paper concerns two-player zero-sum stochastic differential games with nonanticipative strategies against closed-loop controls in the case where the coefficients of mean-field stochastic differential equations and cost functional depend on the joint distribution of the state and the *** our game,both the(lower and upper)value functions and the(lower and upper)second-order Bellman–Isaacs equations are defined on the Wasserstein space P_(2)(R^(n))which is an infinite dimensional *** dynamic programming principle for the value functions is *** the(upper and lower)value functions are smooth enough,we show that they are the classical solutions to the second-order Bellman–Isaacs *** the other hand,the classical solutions to the(upper and lower)Bellman–Isaacs equations are unique and coincide with the(upper and lower)value *** an illustrative application,the linear quadratic case is *** the Isaacs condition,the explicit expressions of optimal closed-loop controls for both players are ***,we introduce the intrinsic notion of viscosity solution of our second-order Bellman–Isaacs equations,and characterize the(upper and lower)value functions as their viscosity solutions.
In this paper, we are concerned with game-theoretic interpretations to the following oblique derivative boundary value problem Delta pNu=0in Omega,+gamma u=gamma Gon partial derivative Omega,\documentclass[12pt]{minim...
详细信息
In this paper, we are concerned with game-theoretic interpretations to the following oblique derivative boundary value problem Delta pNu=0in Omega,+gamma u=gamma Gon partial derivative Omega,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\begin{aligned} \left\{ \begin{array}{ll} \Delta _{p}<^>{N}u=0 & \text {in} \,\,\, \Omega ,\\ \langle \beta , Du \rangle + \gamma u = \gamma G & \text {on} \,\,\, \partial \Omega ,\\ \end{array} \right. \end{aligned}$$\end{document}where Delta pN\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Delta _{p}<^>{N}$$\end{document} is the normalized p-Laplacian. This problem can be regarded as a generalized version of the Robin boundary value problem for the Laplace equations. We construct several types of stochastic games associated with this problem by using 'shrinking tug-of-war'. For the value functions of such games, we investigate the properties such as existence, uniqueness, regularity and convergence.
This paper introduces the formalism required to analyze a certain class of stochastic control problems that involve a super diffusion as the underlying controlled system. To establish the existence of these processes,...
详细信息
This paper introduces the formalism required to analyze a certain class of stochastic control problems that involve a super diffusion as the underlying controlled system. To establish the existence of these processes, we show that they are weak scaling limits of controlled branching processes. First, we prove a generalized It & ocirc;'s formula for this dynamics in the space of finite measures, using the differentiation in the space of finite positive measures. This lays the groundwork for a PDE characterization of the value function of a control problem, which leads to a verification theorem. Finally, focusing on an exponential-type value function, we show how a regular solution to a finite-dimensional HJB equation can be used to construct a smooth solution to the HJB equation in the space of finite measures, via the so-called branching property technique. (c) 2025 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY license (http://***/licenses/by/4.0/).
The COVID-19 pandemic has seen the development of several mathematical models. In recent years, the very topical issue of re- susceptibility has led to the proposal of more complex models to address this issue. The pa...
详细信息
The COVID-19 pandemic has seen the development of several mathematical models. In recent years, the very topical issue of re- susceptibility has led to the proposal of more complex models to address this issue. The paper deals with an optimal control problem applied to COVID-19. The Pontryagin maximum principle and the dynamic programming principle are used to solve the problem. A compartmental Ordinary Differential Equation (ODE) model is proposed to study the evolution of the pandemic by controlling the effectiveness of the detection campaign and the treatment. We prove the global stability of the Disease-Free Equilibrium (DFE) and the existence of optimal control and trajectories of the model. In the optimal control problem, we bring the system back to the DFE. Numerical simulations based on COVID-19 data in Senegal show possibilities to reduce the disease evolution, sometimes by emphasizing the detection campaign and/or the treatment proposed to patients.
We study deterministic optimal control problems for differential games with finite horizon. We propose new approximations of the strategies in feedback form and show error estimates and a convergence result of the val...
详细信息
We study deterministic optimal control problems for differential games with finite horizon. We propose new approximations of the strategies in feedback form and show error estimates and a convergence result of the value in some weak sense for one of the formulations. This result applies in particular to neural network approximations. This work follows some ideas introduced in Bokanowski, Prost and Warin (PDEA, 2023) for deterministic optimal control problems, yet with a simplified approach for the error estimates, which allows to consider a global optimization scheme instead of a time-marching scheme. We also give a new approximation result between the continuous and the semi-discrete optimal control value in the game setting, improving the classical convergence order O(Delta t1/2)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$O({\Delta t}<^>{1/2})$$\end{document} to O(Delta t)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$O({\Delta t})$$\end{document}, under some assumptions on the dynamical system. Numerical examples are performed on elementary academic problems related to backward reachability, with exact analytic solutions given, as well as a two-player game in the presence of state constraints, using stochastic gradient-type algorithms to deal with the minimax problem.
This paper aims at studying a class of discontinuous deterministic control problems under state constraints using a linear programming approach. As for classical control problems (Gaitsgory and Quincampoix (2009) [16]...
详细信息
This paper aims at studying a class of discontinuous deterministic control problems under state constraints using a linear programming approach. As for classical control problems (Gaitsgory and Quincampoix (2009) [16], Goreac and Serea (2011) [19]), the primal linear problem is stated on some appropriate space of probability measures. Naturally, the support of these measures is contained in the set of constraints. This linearized value function and its dual can, alternatively, be seen as the limit of standard penalized problems. Second, we provide a semigroup property for this set of probability measures leading to dynamic programming principles for control problems under state constraints. An abstract principle is provided for general bounded cost. Linearized versions are obtained under further (semi)continuity assumptions. (c) 2012 Elsevier Inc. All rights reserved.
We use optimal control via a distributed exterior field to steer the dynamics of an ensemble of N interacting ferromagnetic particles which are immersed into a heat bath by minimizing a quadratic functional. Using the...
详细信息
We use optimal control via a distributed exterior field to steer the dynamics of an ensemble of N interacting ferromagnetic particles which are immersed into a heat bath by minimizing a quadratic functional. Using the dynamic programming principle, we show the existence of a unique strong solution of the optimal control problem. By the Hopf-Cole transformation, the associated Hamilton-Jacobi-Bellman equation of the dynamic programming principle may be re-cast into a linear PDE on the manifold M=(S2)N, whose classical solution may be represented via Feynman-Kac formula. We use this probabilistic representation for Monte-Carlo simulations to illustrate optimal switching dynamics.
Negative dynamicprogramming for risk-sensitive control is studied. Under some compactness and semicontinuity assumptions the following results are proved: the convergence of the value iteration algorithm to the optim...
详细信息
Negative dynamicprogramming for risk-sensitive control is studied. Under some compactness and semicontinuity assumptions the following results are proved: the convergence of the value iteration algorithm to the optimal expected total reward, the Borel measurability or upper semicontinuity of the optimal value functions, and the existence of an optimal stationary policy. (c) 2008 Elsevier B.V. All rights reserved.
The dynamic programming principle for a multidimensional singular stochastic control problem is established in this paper. When assuming Lipschitz continuity on the data, it is shown that the value function is continu...
详细信息
The dynamic programming principle for a multidimensional singular stochastic control problem is established in this paper. When assuming Lipschitz continuity on the data, it is shown that the value function is continuous and is the unique viscosity solution of the corresponding Hamilton-Jacobi-Bellman equation.
We study optimal control of the general stochastic McKean-Vlasov equation. Such a problem is motivated originally from the asymptotic formulation of cooperative equilibrium for a large population of particles (players...
详细信息
We study optimal control of the general stochastic McKean-Vlasov equation. Such a problem is motivated originally from the asymptotic formulation of cooperative equilibrium for a large population of particles (players) in mean-field interaction under common noise. Our first main result is to state a dynamic programming principle for the value function in the Wasserstein space of probability measures, which is proved from a flow property of the conditional law of the controlled state process. Next, by relying on the notion of differentiability with respect to probability measures due to [P.L. Lions, Cours au College de France : Theorie des jeux a champ moyens, (2012), pp. 2006-2012] and Ito's formula along a flow of conditional measures, we derive the dynamicprogramming Hamilton-Jacobi-Bellman equation and prove the viscosity property together with a uniqueness result for the value function. Finally, we solve explicitly the linear-quadratic stochastic McKean-Vlasov control problem and give an application to an interbank systemic risk model with common noise.
暂无评论