April 22, 2021
Similar papers 5
November 22, 2015
In this paper we present a dynamic programing approach to stochastic optimal control problems with dynamic, time-consistent risk constraints. Constrained stochastic optimal control problems, which naturally arise when one has to consider multiple objectives, have been extensively investigated in the past 20 years, however, in most formulations, the constraints are formulated as either risk-neutral (i.e., by considering an expected cost), or by applying static, single-period r...
May 10, 2020
This paper studies an optimal stochastic impulse control problem in a finite horizon with a decision lag, by which we mean that after an impulse is made, a fixed number units of time has to be elapsed before the next impulse is allowed to be made. The continuity of the value function is proved. A suitable version of dynamic programming principle is established, which takes into account the dependence of state process on the elapsed time. The corresponding Hamilton-Jacobi-Bell...
September 11, 2021
Stochastic optimal principle leads to the resolution of a partial differential equation (PDE), namely the Hamilton-Jacobi-Bellman (HJB) equation. In general, this equation cannot be solved analytically, thus numerical algorithms are the only tools to provide accurate approximations. The aims of this paper is to introduce a novel fitted finite volume method to solve high dimensional degenerated HJB equation from stochastic optimal control problems in high dimension ($ n\geq 3$...
December 20, 2004
We consider a class of optimal control problems of stochastic delay differential equations (SDDE) that arise in connection with optimal advertising under uncertainty for the introduction of a new product to the market, generalizing classical work of Nerlove and Arrow (1962). In particular, we deal with controlled SDDE where the delay enters both the state and the control. Following ideas of Vinter and Kwong (1981) (which however hold only in the deterministic case), we reform...
July 22, 2018
This paper considers a non-Markov control problem arising in a financial market where asset returns depend on hidden factors. The problem is non-Markov because nonlinear filtering is required to make inference on these factors, and hence the associated dynamic program effectively takes the filtering distribution as one of its state variables. This is of significant difficulty because the filtering distribution is a stochastic probability measure of infinite dimension, and the...
May 3, 2019
The paper concerns the study of equilibrium points, or steady states, of economic systems arising in modeling optimal investment with \textit{vintage capital}, namely, systems where all key variables (capitals, investments, prices) are indexed not only by time $\tau$ but also by age $s$. Capital accumulation is hence described as a partial differential equation (briefly, PDE), and equilibrium points are in fact equilibrium distributions in the variable $s$ of ages. Investment...
August 14, 2024
This paper continues the study of controlled interacting particle systems with common noise started in [W. Gangbo, S. Mayorga and A. {\'{S}}wi{\k{e}}ch, \textit{SIAM J. Math. Anal.} 53 (2021), no. 2, 1320--1356] and [S. Mayorga and A. {\'{S}}wi{\k{e}}ch, \textit{SIAM J. Control Optim.} 61 (2023), no. 2, 820--851]. First, we extend the following results of the previously mentioned works to the case of multiplicative noise: (i) We generalize the convergence of the value functio...
December 29, 2021
This paper studies a mean-risk portfolio choice problem for log-returns in a continuous-time, complete market. This is a growth-optimal problem with risk control. The risk of log-returns is measured by weighted Value-at-Risk (WVaR), which is a generalization of Value-at-Risk (VaR) and Expected Shortfall (ES). We characterize the optimal terminal wealth up to the concave envelope of a certain function, and obtain analytical expressions for the optimal wealth and portfolio poli...
August 12, 2020
In this paper, we study backward doubly stochastic recursive optimal control problem where the cost function is described by the solution of a backward doubly stochastic differential equation. We give the dynamical programming principle for this kind of optimal control problem and show that the value function is the unique Sobolev weak solution for the corresponding stochastic Hamilton-Jacobi-Bellman equation.
May 14, 2021
This paper considers time-inconsistent problems when control and stopping strategies are required to be made simultaneously (called stopping control problems by us). We first formulate the timeinconsistent stopping control problems under general multi-dimensional controlled diffusion model and propose a formal definition of their equilibria. We show that an admissible pair $(\hat{u},C)$ of controlstopping policy is equilibrium if and only if the auxiliary function associated ...