How to compute (super) hedging costs in rather general financial market models with transaction costs in discrete-time? Despite the huge literature on this topic, most of results are characterizations of the super-hed...
详细信息
How to compute (super) hedging costs in rather general financial market models with transaction costs in discrete-time? Despite the huge literature on this topic, most of results are characterizations of the super-hedging prices while it remains difficult to deduce numerical procedure to estimate them. We establish here a dynamic programming principle and we prove that it is possible to implement it under some conditions on the conditional supports of the price and volume processes for a large class of market models including convex costs such as order books but also non convex costs, e.g. fixed cost models.(c) 2023 Elsevier Inc. All rights reserved.
In this paper, we investigate a backward doubly stochastic recursive optimal control problem wherein the cost function is expressed as the solution to a backward doubly stochastic differential equation. We present the...
详细信息
作者:
Dong, YuchaoMeng, QingxinZhang, QiTongji Univ
Sch Math Sci Key Lab Intelligent Comp & Applicat Minist Educ Shanghai 200092 Peoples R China Huzhou Univ
Dept Math Sci Huzhou 313000 Zhejiang Peoples R China Fudan Univ
Sch Math Sci Shanghai 200433 Peoples R China Fudan Univ
Lab Math Nonlinear Sci Shanghai 200433 Peoples R China
This paper aims to explore the relationship between maximum principle and dynamic programming principle for stochastic recursive control problem with random coefficients. Under certain regular conditions for the coeff...
详细信息
This paper aims to explore the relationship between maximum principle and dynamic programming principle for stochastic recursive control problem with random coefficients. Under certain regular conditions for the coefficients, the relationship between the Hamiltonian system with random coefficients and stochastic Hamilton-Jacobi-Bellman equation is obtained. It is very different from the deterministic coefficients case since stochastic Hamilton-Jacobi-Bellman equation is a backward stochastic partial differential equation with solution being a pair of random fields rather than a deterministic function. A linear quadratic recursive optimization problem is given as an explicitly illustrated example based on this kind of relationship.
Within the framework of viscosity solution, we study the relationship between the maximum principle (MP) from M. Hu, S. Ji and X. Xue [SIAM J. Control Optim. 56 (2018) 4309-4335] and the dynamic programming principle ...
详细信息
Within the framework of viscosity solution, we study the relationship between the maximum principle (MP) from M. Hu, S. Ji and X. Xue [SIAM J. Control Optim. 56 (2018) 4309-4335] and the dynamic programming principle (DPP) from M. Hu, S. Ji and X. Xue [SIAM J. Control Optim. 57 (2019) 3911-3938] for a fully coupled forward-backward stochastic controlled system (FBSCS) with a nonconvex control domain. For a fully coupled FBSCS, both the corresponding MP and the corresponding Hamilton-Jacobi-Bellman (HJB) equation combine an algebra equation respectively. With the help of a new decoupling technique, we obtain the desirable estimates for the fully coupled forward-backward variational equations and establish the relationship. Furthermore, for the smooth case, we discover the connection between the derivatives of the solution to the algebra equation and some terms in the first-order and second-order adjoint equations. Finally, we study the local case under the monotonicity conditions as from J. Li and Q. Wei [SIAM J. Control Optim. 52 (2014) 1622-1662] and Z. Wu [Syst. Sci. Math. Sci. 11 (1998) 249-259], and obtain the relationship between the MP from Z. Wu [Syst. Sci. Math. Sci. 11 (1998) 249-259] and the DPP from J. Li and Q. Wei [SIAM J. Control Optim. 52 (2014) 1622-1662].
This paper is devoted to studying an infinite time horizon stochastic recursive control problem with jumps, where an infinite time horizon stochastic differential equation and backward stochastic differential equation...
详细信息
This paper is devoted to studying an infinite time horizon stochastic recursive control problem with jumps, where an infinite time horizon stochastic differential equation and backward stochastic differential equation with jumps describe the state process and the cost functional, respectively. By establishing the dynamic programming principle, we shed light on the value function of the control problem with an integral-partial differential equation of HJB type in the sense of viscosity solutions. On the other hand, stochastic verification theorems are also studied to provide sufficient conditions to verify the optimality of the given admissible controls. Such a study is carried out within the framework of classical solutions as well as in that of viscosity solutions. Our work emphasizes important differences from the approach for finite time horizon problems. In particular, we have to work in an L-p-setting for p > 4 in order to study the verification theorem in viscosity sense.
We show that the well-known relationship between the dual extremal are in the maximum principle and the optimal value function (of dynamicprogramming), calculated on the optimal trajectory, is valid for the control o...
详细信息
We show that the well-known relationship between the dual extremal are in the maximum principle and the optimal value function (of dynamicprogramming), calculated on the optimal trajectory, is valid for the control of parabolic variational inequalities. It follows that every optimal control is given by a feedback law. In the case when the functions defining the performance index are convex also with respect to the state variable, a more specific result is obtained.
This paper investigates the relationship between the stochastic maximum principle and the dynamic programming principle for singular stochastic control problems. The state of the system under consideration is governed...
详细信息
This paper investigates the relationship between the stochastic maximum principle and the dynamic programming principle for singular stochastic control problems. The state of the system under consideration is governed by a stochastic differential equation, with nonlinear coefficients, allowing both classical control and singular control. We show that the necessary conditions for optimality, obtained earlier, are in fact sufficient provided some concavity conditions are fulfilled. In a second step, we prove a verification theorem and we show that the solution of the adjoint equation coincides with the derivative of the value function. Finally, using these results, we solve explicitly an example.
The dynamic programming principle (DPP) is fundamental for control and optimization, including Markov decision problems (MDPs), reinforcement learning (RL), and, more recently, mean-field controls (MFCs). However, in ...
详细信息
The dynamic programming principle (DPP) is fundamental for control and optimization, including Markov decision problems (MDPs), reinforcement learning (RL), and, more recently, mean-field controls (MFCs). However, in the learning framework of MFCs, the DPP has not been rigorously established, despite its critical importance for algorithm designs. In this paper, we first present a simple example in MFCs with learning where the DPP fails with a misspecified Q function and then propose the correct form of Q function in an appropriate space for MFCs with learning. This particular form of Q function is different from the classical one and is called the IQ function. In the special case when the transition probability and the reward are independent of the mean-field information, it integrates the classical Q function for single-agent RL over the state-action distribution. In other words, MFCs with learning can be viewed as lifting the classical RLs by replacing the state-action space with its probability distribution space. This identification of the IQ function enables us to establish precisely the DPP in the learning framework of MFCs. Finally, we illustrate through numerical experiments the time consistency of this IQ function.
This paper focuses on the McKean-Vlasov system's stochastic optimal control problem with Markov regime-switching. To this end, the authors establish a new It & ocirc;'s formula using the linear derivative ...
详细信息
This paper focuses on the McKean-Vlasov system's stochastic optimal control problem with Markov regime-switching. To this end, the authors establish a new It & ocirc;'s formula using the linear derivative on the Wasserstein space. This formula enables us to derive the Hamilton-Jacobi-Bellman equation and verification theorems for McKean-Vlasov optimal controls with regime-switching using dynamicprogramming. As concrete applications, the authors first study the McKean-Vlasov stochastic linear quadratic optimal control problem of the Markov regime-switching system, where all the coefficients can depend on the jump that switches among a finite number of states. Then, the authors represent the optimal control by four highly coupled Riccati equations. Besides, the authors revisit a continuous-time Markowitz mean-variance portfolio selection model (incomplete market) for a market consisting of one bank account and multiple stocks, in which the bank interest rate, the appreciation and volatility rates of the stocks are Markov-modulated. The mean-variance efficient portfolios can be derived explicitly in closed forms based on solutions of four Riccati equations.
The aim of the paper is to provide a linearization approach to the L-infinity-control problems. We begin by proving a semigroup-type behaviour of the set of constraints appearing in the linearized formulation of (stan...
详细信息
The aim of the paper is to provide a linearization approach to the L-infinity-control problems. We begin by proving a semigroup-type behaviour of the set of constraints appearing in the linearized formulation of (standard) control problems. As a byproduct we obtain a linear formulation of the dynamic programming principle. Then, we use the L-p approach and the associated linear formulations. This seems to be the most appropriate tool for treating L-infinity problems in continuous and lower semicontinuous setting.
暂无评论