Mathematical Control & Related Fields
September 2019 , Volume 9 , Issue 3
Select all articles
In this paper, we consider the exact synchronization of a kind of parabolic system and obtain Pontryagin's maximum principle for a related optimal control problem. The method relies on the properties of the null controllability for parabolic system and an observability estimate for a linear parabolic system.
Focusing on the problem arising from a stochastic model of risk control and dividend optimization techniques for a financial corporation, this work considers a parabolic variational inequality with gradient constraint
Suppose the company's performance index is the total discounted expected dividends, our objective is to choose a pair of control variables so as to maximize the company's performance index, which is the solution to the above variational inequality under certain initial-boundary conditions. The main effort is to analyse the properties of the solution and two free boundaries arising from the above variational inequality, which we call dividend boundary and reinsurance boundary.
We consider the problem of boundary control for a vibrating string with $N$ interior point masses. We assume the control of Dirichlet, or Neumann, or mixed type is at the left end, and the string is fixed at the right end. Singularities in waves are "smoothed" out to one order as they cross a point mass. We characterize the reachable set for an $L^2$ control. The control problem is reduced to a moment problem, which is then solved using the theory of exponential divided differences in tandem with unique shape and velocity controllability results. The results are sharp with respect to both the regularity of the solution and with respect to time. The eigenfunctions of the associated Sturm--Liouville problem are used to construct Riesz bases for a family of asymmetric spaces that include the sets of reachable positions and velocities.
This paper considers the stochastic linear quadratic optimal control problem in which the control domain is nonconvex. By the functional analysis and convex perturbation methods, we establish a novel maximum principle. The application of the proposed maximum principle is illustrated through a work-out example.
We show how to reconstruct the shape of a solid of revolution by measuring its temperature on the boundary. This inverse problem reduces to finding a coefficient of a parabolic equation from values of the trace of its solution on the boundary. This is achieved by using the inverse spectral theory of the string, as developed by M.G. Krein, which provides uniqueness and also a reconstruction algorithm.
Physical systems such as water and gas networks are usually operated in a state of equilibrium and feedback control is employed to damp small perturbations over time. We consider flow problems on networks, described by hyperbolic balance laws, and analyze the stability of the linearized systems. Sufficient conditions for exponential stability in the continuous and discretized setting are presented. The analysis is extended to arbitrary Sobolev norms. Computational experiments illustrate the theoretical findings.
In this paper, we study general time-inconsistent stochastic control models which are driven by a stochastic differential equation with random jumps. Specifically, the time-inconsistency arises from the presence of a non-exponential discount function in the objective functional. We consider equilibrium, instead of optimal, solution within the class of open-loop controls. We prove an equivalence relationship between our time-inconsistent problem and a time-consistent problem such that the equilibrium controls for the time-consistent problem coincide with the equilibrium controls for the time-inconsistent problem. We establish two general results which characterize the open-loop equilibrium controls. As special cases, a generalized Merton's portfolio problem and a linear-quadratic problem are discussed.
We establish existence of Markov chains of mean-field type with unbounded jump intensities by means of a fixed point argument using the total variation distance. We further show existence of nearly-optimal controls and, using a Markov chain backward SDE approach, we suggest conditions for existence of an optimal control and a saddle-point for respectively a control problem and a zero-sum differential game associated with payoff functionals of mean-field type, under dynamics driven by such Markov chains of mean-field type.
Add your name and e-mail address to receive news of forthcoming issues of this journal:
[Back to Top]