- Research
- Open access
- Published:
An optimal control problem for linear SDE of mean-field type with terminal constraint and partial information
Advances in Difference Equations volume 2019, Article number: 160 (2019)
Abstract
This paper is concerned with an optimal control problem for a linear stochastic differential equation (SDE) of mean-field type, where the drift coefficient of observation equation is linear with respect to the state, the control and their expectations, and the state is subject to a terminal constraint. The control problem cannot be solved by transforming it into a standard optimal control problem for an SDE without mean-field term. By virtue of a backward separation method with a decomposition technique, one optimality condition and one forward–backward filter are derived. Two linear-quadratic (LQ) optimal control problems and one cash management problem with terminal constraint and partial information are studied, and optimal feedback controls are explicitly obtained.
1 Introduction
One begins with a complete filtered probability space \((\varOmega , \mathcal{F}, (\mathcal{F}_{t})_{0\leq t\leq T}, \mathbb{P})\), on which are given an \(\mathcal{F}_{t}\)-adapted standard Brownian motion \((\omega _{t}, \tilde{\omega }_{t})\) with value in \(\mathbb{R}^{2}\) and a Gaussian random variable ξ with mean \(\mu _{0}\) and covariance \(\sigma _{0}\). \((\omega , \tilde{\omega })\) is independent of ξ. Let \(T>0\) be a fixed time horizon, let \(\mathbb{R}^{m}\) be the m-dimensional Euclidean space, and let \(f_{x}\) be the partial derivative of f with respect to x. If \(x: [0,T]\rightarrow \mathbb{R}\) is uniformly bounded, one writes \(x\in L^{\infty }(0,T;\mathbb{R})\). If \(x:[0,T]\rightarrow \mathbb{R}\) is square-integrable, one writes \(x\in L^{2}(0,T;\mathbb{R})\). If \(x:[0,T]\times \varOmega \rightarrow \mathbb{R}\) is an \(\mathcal{F}_{t}\)-adapted, square-integrable process, one writes \(x\in L^{2}_{\mathcal{F}}(0,T;\mathbb{R})\). One also adopts similar notations for other filtrations and Euclidean spaces.
Consider the linear SDE
where \(\mathbb{E}\) is expectation, and v is a control process. Since \(\mathbb{E}x^{v}_{t}\) and \(\mathbb{E}v_{t}\) appear in the equation, one calls it an SDE of mean-field type. Assume that the solution \(x^{v}\) to the equation is observed through
The cost functional is
Here \(v_{t}\) is required to be \(\sigma \{y^{v}_{s};0\leq s \leq t \}\)-adapted and to satisfy
and
the functions a, ā, b, b̄, c, c̃, f, f̄, g, ḡ, h, l, ϕ and φ will be specified in Sect. 2. This is a partially observable optimal control problem with terminal constraint. This problem can reduce to an optimal control problem with certain additional control domain constraint, but it cannot be studied by classical control theory for SDE without mean-field term. From this viewpoint, this problem extends some standard optimal control problems and covers a few financial models.
Classical variation provides an effective tool for studying optimal control problems. However, it is not always valid for partially observed optimal control problems. A main reason is there is a circular dependence between the control v and the observation \(y^{v}\). In 2008, Wang and Wu [1] originally proposed a backward separation method. In 2018, Wang coauthored their monograph [2], where the backward separation method was systematically introduced and was regarded as one of most important tools for studying partially observed optimal control problems. Combining the backward separation method with Girsanov’s measure transformation, the circular dependence between v and \(y^{v}\) was decoupled in Wang et al. [3], and then a necessary condition for optimality was derived. Along this line, Zhang [4], Ma and Liu [5] extended [3] to the case of correlated state and observation noises, and the case of risk-sensitive control, respectively. Buckdahn et al. [6] studied an optimal control problem for SDE of conditional mean-field type. One emphasizes that [3,4,5] and [6] are based on the assumption that the drift coefficients of observation equations are uniformly bounded with respect to their components, which is restricted in some applications. Using the backward separation method with an approximation technique, Wang et al. [7] generalized [3,4,5,6] in the sense that the drift coefficient of observation equation linearly grows with respect to the state, and for any \(\ell >0\), the control v satisfies \(\mathbb{E}\sup_{0\leq t\leq T}\vert v_{t}\vert ^{\ell }<+ \infty \). Note that [7] did not study the case of mean-field and terminal constraint.
Clearly, the control problem in this paper does not satisfy the assumptions above, and then the foregoing techniques are not valid. To overcome the difficulty caused, one will adopt a decomposition technique introduced in Wang et al. [8, 9], where a partially observable forward–backward stochastic control system without mean-field term and/or terminal constraint was considered. Combining the decomposition technique with the backward separation method, one solves the control problem. The contributions of this paper are as follows.
-
One new necessary condition for optimality is derived. The condition together with forward–backward filter provides an effective method for studying stochastic optimal control with terminal constraint and incomplete information.
-
Three LQ examples with terminal constraint and partial information are solved, and optimal feedback controls are obtained by accident.
-
An SDE of mean-field type naturally arises from the study of standard LQ optimal control driven by SDE without mean-field term. This interesting contribution can be found in Example 4.2 below.
The control problem is also related to those of Meyer-Brandis et al. [10], Elliott et al. [11], Yong [12], Hafayed and Abbas [13], Ni et al. [14] and Hafayed et al. [15]. Specifically, [10, 15], respectively, studied a mean-field type control problem with partial information, where neither noisy observation nor filter is studied. The other work investigated mean-field type controls with complete information.
The rest of this paper is organized as follows. In Sect. 2, one reformulates the control problem and provides preliminary results. Section 3 derives one optimality condition and one forward–backward filtering equation of mean-field type. In Sect. 4, one explicitly solves three LQ optimal control problems with terminal constraint and partial information. Finally, in Sect. 5, one gives some concluding remarks.
2 Problem formulation and preliminary
Define \(x^{0}\) and \(y^{0}\) by two SDEs
and
where \(a, \bar{a}, c, \tilde{c}, f, \bar{f}, h \in L^{\infty }(0,T; \mathbb{R})\). Assume that the control \(v\in L^{2}_{\mathcal{F}}(0,T;\mathbb{R})\). Define \(x^{v,1}\) and \(y^{v,1}\) by
and
where \(b, \bar{b}, g,\bar{g}\in L^{\infty }(0,T; \mathbb{R})\). It is clear that Eqs. (1), (2), (3) and (4) have unique solutions, respectively. Set
It follows from Itô’s formula that \(x^{v}\) and \(y^{v}\) satisfy
and
respectively. For any \(v\in L^{2}_{\mathcal{F}}(0,T;\mathbb{R})\), one introduces a constraint condition regarding the terminal state and its distribution
Let
and let U be a nonempty convex subset of \(\mathbb{R}\). Define three admissible control sets
and
It is easy to see that the inclusion relationship among them is
With (5) and the definition of \(\mathcal{U}_{\mathrm {ad}}\), one proves the equality
In fact, since \(v_{t}\) is \(\mathcal{F}^{y^{0}}_{t}\)-adapted, then it follows from (3) and (4) that \(y^{v,1}_{t}\) is \(\mathcal{F}^{y^{0}}_{t}\)-adapted, and thus \(y^{v}_{t}\) is also \(\mathcal{F}^{y^{0}}_{t}\)-adapted by using the second equality in (5). This implies that \(\mathcal{F}^{y^{v}}_{t}\subseteq \mathcal{F}^{y^{0}}_{t}\), \(v\in \mathcal{U}_{\mathrm {ad}}\). In the same way, one gets \(\mathcal{F}^{y^{v}}_{t}\supseteq \mathcal{F}^{y^{0}}_{ t}\), \(v\in \mathcal{U}_{\mathrm {ad}}\). Then one draws the desired conclusion.
The cost functional is in the form of
where \(l:[0,T]\times \mathbb{R}^{2}\times U\times \mathbb{R}\rightarrow \mathbb{R}\) and \(\phi , \varphi :\mathbb{R}^{2}\rightarrow \mathbb{R}\) are continuously differentiable with respect to \((x,\bar{x},v,\bar{v})\) and \((x, \bar{x})\), respectively, and there is a constant \(C>0\) such that
with \(\psi =\phi , \varphi \) and \(\chi =x, \bar{x}, v, \bar{v}\).
Then the optimal control problem with terminal constraint is restated as follows.
Problem (TC)
Find a \(u\in \mathcal{U}^{c}_{\mathrm {ad}}\) such that
subject to (6), (7), (8) and (9). Any u satisfying the equality is called an optimal control of Problem (TC), and \(x^{u}\) is called the optimal state corresponding to u.
One also introduces an auxiliary problem without terminal constraint.
Problem (A)
Find a \(u\in \mathcal{U}_{\mathrm {ad}}\) such that
In what follows, one provides two preliminary results, whose proofs can be found in the Appendix.
Proposition 2.1
For any \(\kappa \in \Delta \), one has
Proposition 2.2
Suppose that, for any \(\kappa \in \Delta \), \(u_{\kappa }\) is an optimal control of Problem (A). Moreover, suppose that there exists \(\kappa _{0}\in \Delta \) such that
Then \(u=u_{\kappa _{0}}\) is an optimal control of Problem (TC).
Proposition 2.1 reveals the equivalence between Problem (A) and the problem of minimizing \(J_{\kappa }[v]\) over \(\mathcal{U}^{0}_{\mathrm {ad}}\). Proposition 2.2 together with Proposition 2.1 shows that one can obtain an optimal control of Problem (TC) by the following procedures: (1) to derive all optimal controls \(u_{\kappa }\) of Problem (A); (2) to find \(u_{\kappa _{0}}\) satisfying \(\mathbb{E}\varphi (x^{u_{\kappa _{0}}}_{T}, \mathbb{E}x^{u_{\kappa _{0}}} _{T})=0\). Then such \(u_{\kappa _{0}}\) is exactly an optimal control of Problem (TC). Clearly, it is a more convenient approach in at least some detailed cases. See, e.g., Sect. 4 for more details.
This remark shows that the second procedure above can easily be finished in general, and thus it is enough to study Problem (A).
3 Optimality condition of Problem (A)
For any \(v,v_{j}\in \mathcal{U}_{\mathrm {ad}}\), let \(x^{v}\) and \(x^{v_{j}}\) be the solutions to (6) corresponding to v and \(v_{j}\), \(j=1,2,\ldots \) . For simplicity, we set
where \(\lambda =v\), \(u_{\kappa }\), \(v_{j}\), \(j=1,2,\ldots \) .
Theorem 3.1
If \(u_{\kappa }\) is an optimal control of Problem (A), then the backward stochastic differential equation of mean-field type
has a unique solution \((p_{\kappa }, q_{\kappa },\tilde{q}_{\kappa }) \in L^{2}_{\mathcal{F}}(0,T;\mathbb{R}^{3})\) such that for any \(\nu \in U\)
where the Hamiltonian function H is defined by
Proof
If \(u_{\kappa }\) is an optimal control of Problem (A), Proposition 2.1 implies that
For any \(v\in \mathcal{U}^{0}_{\mathrm {ad}}\), let \(x^{u_{\kappa }+\varepsilon v}\) be the solution to (6) corresponding to \(u_{\kappa }+ \varepsilon v\), where \(0\leq \varepsilon \leq 1\). Introduce the variational equation
which admits a unique solution \(x_{1}\in L^{2}_{\mathcal{F}}(0,T;\mathbb{R})\). It follows from Hölder’s inequality that
Combining the limit with the optimality of u, one derives the first-order variational inequality
On the other hand, once \(x^{u_{\kappa }}\) is determined by (6), (11) admits a unique solution \((p_{\kappa }, q_{\kappa }, q _{\kappa })\) in \(L^{2}_{\mathcal{F}}(0,T;\mathbb{R}^{3})\). Using Itô’s formula to \(x_{1}p_{\kappa }\) and inserting it into the variational inequality, one gets
Due to \({u_{\kappa }}\in \mathcal{U}^{0}_{\mathrm {ad}}\) and the arbitrariness of \(v_{t}\), one deduces
Recalling for any \({u_{\kappa }}\in \mathcal{U}_{\mathrm {ad}}\), \(\mathcal{F} ^{y^{u_{\kappa }}}_{t}=\mathcal{F}^{y^{0}}_{t}\), then one draws the desired conclusion. □
According to (12), one needs to compute the optimal filters of (11) and (6) depending on \(\mathcal{F}^{y^{v}}_{t}\) in order to compute \(u_{\kappa }\). For this purpose, one denotes by
and
the filters of \(\varPhi _{t}\) and \(\varPsi _{t}\), respectively. Moreover, one denotes by
the mean square error of \(\hat{x}^{v}_{t}\), \(v\in \mathcal{U}_{\mathrm {ad}}\). Using Theorems 2.1 and 2.2 in [2] to (6), (7) and (11), one derives the filters \(\hat{x}^{v}_{t}\) and \(\hat{p} _{\kappa ,t}\) of \(x^{v}_{t}\) and \(p_{\kappa ,t}\) with respect to \(\mathcal{F}^{y^{v}}_{t}\).
Theorem 3.2
For any \(v\in \mathcal{U}_{\mathrm {ad}}\), the filters \(\hat{x}^{v}_{t}\) and \(\hat{p}_{\kappa ,t}\) satisfy
and
respectively, where Σ is the unique solution to
is a standard Brownian motion with value in \(\mathbb{R}\), and
One emphasizes that (13) with (14) is a forward–backward stochastic differential filtering equation of mean-field type, which has a unique solution \((\hat{x}^{u_{\kappa }},\hat{p}_{\kappa }, Q)\in L ^{2}_{\mathcal{F}^{y^{u_{\kappa }}}}(0,T;\mathbb{R}^{3})\) for given \(u_{\kappa }\). It shows that Theorem 3.2 is different from the usual filtering theories. See, e.g., Xiong [16].
4 Three LQ cases of Problem (TC)
In this section, one aims at illustrating Theorems 3.1 and 3.2 by three examples. For convenience, one still adopts the state equation, the observation equation, and the corresponding assumptions introduced in Sects. 2 and 3 unless noted otherwise.
Example 4.1
Find an admissible control to minimize
over \(\mathcal{U}^{c}_{\mathrm {ad}}\) with \(U=\mathbb{R}\) and the terminal constraint
subject to
and
where \(A,\bar{A}, B, \bar{B}\in L^{\infty }(0,T;\mathbb{R})\), \(A_{t}>0\), \(A_{t}+\bar{A}_{t}\geq 0\), \(B_{t}>0\), \(B_{t}+\bar{B}_{t}>0\), \(D\geq 0\), \(D+\bar{D}\geq 0\), \(b\neq 0\) and \(b+\bar{b}\neq 0\).
Define an auxiliary cost functional without terminal constraint
with
Since both κ and γ are constants, it is enough to minimize (21) over \(\mathcal{U}_{\mathrm {ad}}\) subject to (19) and (20). One will use three steps to explicitly solve the example.
Step 1 Candidate optimal control of the auxiliary problem without terminal constraint.
With the data, the Hamiltonian function is
where \((p_{\kappa }, q_{\kappa },\tilde{q}_{\kappa })\) is determined by the Hamiltonian system
If \(u_{\kappa }\) is an optimal control of the auxiliary problem, then it follows from Theorem 3.1 that
Solving it, we get
Step 2 Feedback form of (23).
Inserting (23) into (22) and taking expectations, one gets an ordinary differential equation
Note that the first equation and the second equation in (24) are coupled. Since
the assumption condition of Theorem 2.6 in Peng and Wu [17] is satisfied, and hence (24) has a unique solution \((\mathbb{E}x^{u_{\kappa }}, \mathbb{E}p_{\kappa })\). Noticing the terminal condition of (24), one sets
where α and \(\beta _{\kappa }\) are deterministic and differential functions such that \(\alpha _{T}=D+ \bar{D}\) and \(\beta _{\kappa ,T}=\kappa \). Using the chain rule for computing the derivative of (25), one has
Comparing the equality with the second equation in (24), one deduces
and
It is easy to see (26) and (27) admit unique solutions, respectively. Inserting (25) into the first equation of (24), one derives
Using Theorem 3.2 to (22) with (23), one gets a forward–backward stochastic differential filtering equation of mean-field type,
where Σ and ω̂ satisfy (15) with (16), and \(\mathbb{E}x^{u_{\kappa }}\) and \(\mathbb{E}p_{\kappa }\) solve (28) and (25), respectively. Since \(b\neq 0\), (29) admits a unique solution \((\hat{x}^{u_{\kappa }},\hat{p} _{\kappa },Q)\in L^{2}_{\mathcal{F}^{y^{u_{\kappa }}}}(0,T;\mathbb{R}^{3})\) by Theorem 2.6 in [17] again. Similarly, let
where Γ, Γ̄ and \(\varLambda _{\kappa }\) are three deterministic and differential functions satisfying \(\varGamma _{T}=D\), \(\bar{\varGamma }_{T}=\bar{D}\) and \(\varLambda _{\kappa ,T}=\kappa \). It follows from Itô’s formula that
Comparing it with the second equation in (29), one deduces
and
which admit a unique solution, respectively. Plugging (30) into (23), one gets
where Γ, Γ̄, \(\varLambda _{\kappa }\) and \(\hat{x} ^{u_{\kappa }}\) solve (31), (32), (33) and
respectively.
Step 3 Optimal control of Example 4.1.
Solving (27) and (28), one gets
with
Recalling the terminal constraint (18), it yields
Then Proposition 2.2 implies the desired conclusion. The above deduction is summarized as follows.
Proposition 4.1
The optimal feedback control of Example 4.1 is given by (34) with κ being replaced by (35).
Example 4.2
In particular, if one lets the coefficients of (17), (19) and (20) \(\bar{A}=\bar{B}=\bar{D}= \bar{a}=\bar{b}=c=f=\bar{f}=g=\bar{g}=0\) on \([0, T]\), then Example 4.1 is reduced to an LQ optimal control with terminal constraint and complete information. Further, let \(v\in L^{2}(0, T; \mathbb{R})\). Since v is deterministic, Proposition 4.1 implies that the optimal feedback control is
where α, \(\beta _{\kappa _{0}}\) and \(x^{u_{\kappa _{0}}}\) satisfy
and
with
respectively.
Note that (36) is an SDE of mean-field type. It shows that one begins with a classical control system without mean-field term, but one ends up with a control system of mean-field type. This is a very interesting phenomenon indeed.
Example 4.3
One denotes by v the rate of capital withdrawal or injection of a firm, and by \(x^{v}\) the cash-balance process on \([0, T]\). Assume that the liability of the firm is governed by
where \(c_{t}\,d\omega _{t}\) and \(\tilde{c}_{t}\,d\tilde{\omega }_{t}\) describe the liability risk. Assume that the firm owns an initial investment ξ, and only invests in a money account with compounded interest rate a. Then the cash-balance is denoted by
whose differential form is the same as (6) with \(\bar{a}= \bar{b}=0\). Due to the discreteness of the account information, the firm partially observes the cash-balance by the corresponding stock price
Set
It follows from Itô’s formula that \(y^{v}\) is governed by (7) with \(\bar{f}_{t}=\bar{g}_{t}=0\). The firm hopes to find a suitable v such that
is minimized over \(\mathcal{U}^{c}_{\mathrm {ad}}\) with terminal constraint (18). The model implies that the firm wants to minimize the risk of \(x^{v}_{T}\) and v under a fixed terminal cash-balance level. Since
Example 4.3 is also a special case of Example 4.1. The following result is an immediate one of Proposition 4.1.
Corollary 4.1
The optimal rate of capital withdrawal or injection of the firm is
where Γ, Γ̄, Λ and \(\hat{x}^{u_{\kappa _{0}}}\) are the solutions to
and
with
One remarks that the model in Example 4.3 is inspired by Huang et al. [18], where the variance of v does not enter the performance functional.
5 Concluding remarks
This paper has studied an optimal control problem driven by SDE of mean-field tpye, where the drift coefficient of observation equation is linear with respect to the state, the control and their expectations, and the observation equation is explicitly dependent on the control. This framework covers many more financial models including [18], but it causes trouble in addressing the control problem. This trouble has been solved by the backward separation method with a decomposition technique. The results obtained here partially improve those of [3,4,5,6].
The traditional separation principle method is also applicable to study Problem (TC). However, the traditional method cannot offer a valid way to derive an optimality condition of Problem (TC). The reason is listed below. Let \(\varrho (t,x)\) be the conditional density of \(x^{v}_{t}\), given the observable filtration \(\mathcal{F}^{y^{v}}_{t}\). Similar to Theorem 2.3 in [2], one gets
where
is an \(\mathcal{F}^{y^{v}}_{t}\)-adapted and \(\mathbb{R}\)-valued standard Brownian motion, and hence, the cost functional (9) is rewritten as
This is an optimal control problem driven by stochastic partial differential equation with complete information. To obtain the optimality condition of the control problem, lots of stochastic calculuses on partial differential equation should be needed. Then it seems that the traditional method is not as effective as the case of Sects. 3 and 4 in this paper.
Abbreviations
- SDE:
-
stochastic differential equation
- LQ:
-
linear-quadratic
- TC:
-
terminal constraint
References
Wang, G.C., Wu, Z.: Kalman–Bucy filtering equations of forward and backward stochastic systems and applications to recursive optimal control problems. J. Math. Anal. Appl. 342, 1280–1296 (2008)
Wang, G.C., Wu, Z., Xiong, J.: An Introduction to Optimal Control of FBSDE with Incomplete Information. Springer, Cham (2018)
Wang, G.C., Zhang, C.H., Zhang, W.H.: Stochastic maximum principle for mean-field type optimal control with partial information. IEEE Trans. Autom. Control 59, 522–528 (2014)
Zhang, H.Y.: A necessary condition for mean-field type stochastic differential equations with correlated state and observation noises. J. Ind. Manag. Optim. 12, 1287–1301 (2016)
Ma, H.P., Liu, B.: Optimal control problem for risk-sensitive mean-field stochastic delay differential equation with partial information. Asian J. Control 19, 2097–2115 (2017)
Buckdahn, R., Li, J., Ma, J.: A mean-field stochastic control problem with partial observations. Ann. Appl. Probab. 27, 3201–3245 (2017)
Wang, G.C., Wu, Z., Xiong, J.: Maximum principles for forward–backward stochastic control systems with correlated state and observation noises. SIAM J. Control Optim. 51, 491–524 (2013)
Wang, G.C., Wu, Z., Xiong, J.: A linear-quadratic optimal control problem of forward–backward stochastic differential equations with partial information. IEEE Trans. Autom. Control 60, 2904–2916 (2015)
Wang, G.C., Xiao, H., Xing, G.J.: An optimal control problem for mean-field forward–backward stochastic differential equation with noisy observation. Automatica 86, 104–109 (2017)
Meyer-Brandis, T., Øksendal, B., Zhou, X.Y.: A mean-field stochastic maximum principle via Malliavin calculus. Stochastics 84, 643–666 (2012)
Elliott, R., Li, X., Ni, Y.H.: Discrete time mean-field stochastic linear quadratic optimal control problems. Automatica 49, 3222–3233 (2013)
Yong, J.M.: Linear-quadratic optimal control problems for mean-field stochastic differential equations. SIAM J. Control Optim. 51, 2809–2838 (2013)
Hafayed, H., Abbas, S.: On near-optimal mean-field stochastic singular controls: necessary and sufficient conditions for near-optimality. J. Optim. Theory Appl. 160, 778–808 (2014)
Ni, Y.H., Zhang, J.F., Li, X.: Indefinite mean-field stochastic linear-quadratic optimal control. IEEE Trans. Autom. Control 60, 1786–1800 (2015)
Hafayed, M., Abba, A., Abbas, S.: On partial-information optimal singular control problem for mean-field stochastic differential equations driven by teugels martingales measures. Int. J. Control 89, 397–410 (2016)
Xiong, J.: An Introduction to Stochastic Filtering Theory. Oxford University Press, London (2008)
Peng, S.G., Wu, Z.: Fully coupled forward–backward stochastic differential equations and applications to optimal control. SIAM J. Control Optim. 37, 825–843 (1999)
Huang, J.H., Wang, G.C., Wu, Z.: Optimal premium policy of an insurance firm: full and partial information. Insur. Math. Econ. 47, 208–215 (2010)
Øksendal, B.: Stochastic Differential Equations: An Introduction with Applications. Springer, New York (1998)
Availability of data and materials
Not applicable.
Funding
This work is supported in part by the NSF of China under Grant No. 61603219, and by the Shandong Jiaotong University Climbing Research Innovation Team Program.
Author information
Authors and Affiliations
Contributions
The author read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The author declares that there is no conflict of interests regarding the publication of this paper.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
One presents three lemmas first, and then one gives a proof of Proposition 2.1.
Lemma A1
For any \(v_{j}\in L^{2}_{\mathcal{F}}(0,T;\mathbb{R})\), \(j=1,2\), there is a constant \(C>0\) such that
Proof
The estimate is obtained by Itô’s formula and Hölder’s inequality. The details of the proof are omitted to save space. □
Lemma A2
For any \(v, v_{j}\in \mathcal{U}_{\mathrm {ad}}\), \(j=1,2,\ldots \) ,
Proof
Using Taylor’s expansion, Hölder’s inequality and Lemma A1, one gets
as \(j\rightarrow +\infty \). Here \(C>0\) is a constant which can be different from line to line. Similarly, one has
with \(j\rightarrow +\infty \). Then the proof is complete. □
Lemma A3
\(\mathcal{U}_{\mathrm {ad}}\) is dense in \(\mathcal{U} ^{0}_{\mathrm {ad}}\).
Proof
For any \(v\in \mathcal{U}^{0}_{\mathrm {ad}}\), one defines a family of controls by
where \(\nu \in U\), i, j are natural numbers, \(1\leq i\leq j-1\), and \(\delta _{j}=T/j\). Similar to [8], one proves that (i) \(v_{j}\in \mathcal{U}_{\mathrm {ad}}\) for any j, and (ii) \(v_{j}\rightarrow v\) as \(j\rightarrow +\infty \) in \(L^{2}_{\mathcal{F}^{y^{0}}}(0,T;U)\). Thus the proof is complete. □
Proof of Proposition 2.1
From the definition of decision set, it is easy to see
Then one only needs to prove the reverse inequality. Since \(v_{j}\) introduced in Lemma A3 is an element of \(\mathcal{U}_{\mathrm {ad}}\),
and, consequently, it follows from Lemma A2 that
Then the arbitrariness of v implies the desired result. □
Proof of Proposition 2.2
According to the definition of optimal control and value function, the desired conclusion is drawn by using the idea introduced in Chap. 11 of Øksendal [19]. One drops the details of the proof for saving space. □
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Zhang, H. An optimal control problem for linear SDE of mean-field type with terminal constraint and partial information. Adv Differ Equ 2019, 160 (2019). https://doi.org/10.1186/s13662-019-2029-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13662-019-2029-0