HAL Id: hal-03113182
https://hal.inria.fr/hal-03113182
Submitted on 18 Jan 2021
HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.
Method for Linear-Quadratic Optimal Control Problems
Tobias Breiten, Laurent Pfeiffer
To cite this version:
Tobias Breiten, Laurent Pfeiffer. On the Turnpike Property and the Receding-Horizon Method for
Linear-Quadratic Optimal Control Problems. SIAM Journal on Control and Optimization, Society
for Industrial and Applied Mathematics, 2020, 58 (2), pp.26. �10.1137/18M1225811�. �hal-03113182�
METHOD FOR LINEAR-QUADRATIC OPTIMAL CONTROL PROBLEMS ∗
TOBIAS BREITEN
†AND LAURENT PFEIFFER
‡Abstract. Optimal control problems with a very large time horizon can be tackled with the Receding Horizon Control (RHC) method, which consists in solving a sequence of optimal control problems with small prediction horizon. The main result of this article is the proof of the exponential convergence (with respect to the prediction horizon) of the control generated by the RHC method towards the exact solution of the problem. The result is established for a class of infinite-dimensional linear-quadratic optimal control problems with time-independent dynamics and integral cost. Such problems satisfy the turnpike property: the optimal trajectory remains most of the time very close to the solution to the associated static optimization problem. Specific terminal cost functions, derived from the Lagrange multiplier associated with the static optimization problem, are employed in the implementation of the RHC method.
Key words. Receding horizon control, model predictive control, value function, optimality systems, Riccati equation, turnpike property.
AMS subject classifications. 49J20, 49L20, 49Q12, 93D15.
1. Introduction.
1.1. Context. We consider in this article the following class of linear-quadratic optimal control problems:
(P)
inf
y∈W (0, T ¯ ) u∈L
2(0, T;U) ¯
J T ,Q,q ¯ (u, y) :=
Z T ¯ 0
`(y(t), u(t)) dt + 1 2 hy( ¯ T ), Qy( ¯ T )i + hq, y( ¯ T )i, subject to: ˙ y(t) = Ay(t) + Bu(t) + f , y(0) = y 0 ,
where the integral cost ` is defined by
`(y, u) = 1 2 kCyk 2 Z + hg , yi V
∗,V + α 2 kuk 2 U + hh , ui U .
Here V ⊂ Y ⊂ V ∗ is a Gelfand triple of real Hilbert spaces [28, page 147], where the embedding of V into Y is dense, V ∗ denotes the topological dual of V and U, Z denote further Hilbert spaces. The operator A : D(A) ⊂ Y → Y is the infinitesimal generator of an analytic C 0 -semigroup e At on Y , B ∈ L(U, V ∗ ), C ∈ L(Y, Z), α > 0, Q ∈ L(Y ) is self-adjoint positive semi-definite and D(A) denotes the domain of A. The pairs (A, B) and (A, C) are assumed to be stabilizable and detectable, respectively. The elements y 0 ∈ Y , f ∈ V ∗ , g ∈ V ∗ , h ∈ U , q ∈ Y are given.
The following problem, referred to as static optimization problem (or steady-state optimization problem), has a unique solution (y , u ) with unique associated Lagrange multiplier p :
(1.1) inf
(y,u)∈V ×U `(y, u), subject to: Ay + Bu + f = 0.
∗
Submitted to the editors on November 8, 2018.
†
Institute of Mathematics and Scientific Computing, University of Graz, Austria (tobias.breiten@uni-graz.at).
‡
Inria and CMAP (UMR 7641), CNRS, Ecole Polytechnique, Institut Polytechnique de Paris, Route de Saclay, 91128 Palaiseau, France (laurent.pfeiffer@uni-graz.at).
1
arXiv:1811.02421v3 [math.OC] 31 Jan 2020
A particularly important feature of (P ) is the exponential turnpike property. It states that there exist two constants M > 0 and λ > 0, independent of ¯ T , such that for all t ∈ [0, T ¯ ], k y(t)−y ¯ k Y ≤ M e −λt +e −λ( ¯ T −t)
, where ¯ y denotes the optimal trajectory.
The trajectory ¯ y is thus made of three arcs, the first and last one being transient short- time arcs and the middle one a long-time arc, where the trajectory remains close to y . We refer the reader to the books [31, 32], where different turnpike properties are established for different kinds of systems. We mention in particular the general characterization of the turnpike phenomenon for linear systems in [32, Section 5.34].
For linear-quadratic problems, we mention the articles [9, 12] for discrete-time systems and the articles [21] and [22] containing results for classes of infinite-dimensional systems. We also mention the early reference [1] dealing with a tracking problem.
Exponential turnpike properties have been established for non-linear systems in [26]
and [25].
The aim of this article is to analyze the efficiency of the Receding Horizon Control (RHC) method (also called Model Predictive Control method), that we briefly present here, a detailed description can be found in Section 5. We consider an implementation of the method with three parameters: a sampling time τ, a prediction horizon T , and a prescribed number of iterations N . The method generates in a recursive way a control u RH and its associated trajectory y RH . At the beginning of iteration n of the algorithm, u RH and y RH have already been computed on (0, nτ ). Then, an optimal control problem is solved on the interval (nτ, nτ + T), with initial condition y RH (nτ ), with the same integral cost as in (P ), but with the following terminal cost function:
(1.2) φ(y) = hp , yi Y .
The restriction of the solution to (nτ, (n + 1)τ) is then concatenated with (y RH , u RH ).
At iteration N, a last optimal control problem is solved on the interval (N τ, T ¯ ). The definition (1.2) is actually a particular choice of the terminal cost among a general class of linear-quadratic functions. For this specific definition, the main result of the article is the following estimate:
max ky RH − yk ¯ W (0, T ¯ ) , ku RH − uk ¯ L
2(0, T ¯ ;U)
≤ M e −λ(T−τ) e −λT ky 0 − y k Y + e −λ( ¯ T −(N τ+T )) k qk ˜ Y
, (1.3)
with ˜ q = q − p + Qy . The estimate is proven for sampling times and prediction horizons satisfying 0 < τ 0 ≤ τ ≤ T ≤ T ¯ . The constants τ 0 > 0, M > 0, and λ > 0 are independent of y 0 , f , g , h , q, N , τ , T , and ¯ T . Let us mention that the lower bound τ 0 cannot be chosen arbitrarily small. The idea of taking hp , yi as a terminal cost has been proposed in the recent article [29] in the context of discrete-time problems.
The choice of an appropriate terminal cost function is a key issue in the design of an appropriate RHC scheme. When φ is the exact value function, then the RHC method generates the exact solution to the problem, as a consequence of the dynamic programming principle. The article will give a (positive) answer to the following ques- tion: Does the RHC algorithm generate an efficient control if a good approximation of the value function is used as terminal cost function? The construction of such an approximation is here possible thanks to the turnpike property. We will see that the derivative of the value function (with respect to the initial condition), evaluated at y , converges to p as ¯ T − t increases. Roughly speaking, the definition (1.2) is a kind of first-order Taylor approximation of the value function, around y .
The RHC method is receiving a tremendous amount of attention and it is fre-
quently used in control engineering, in particular because it is computationally easier
to solve a problem with short horizon. Another reason is that the method can be used as a feedback mechanism: when the control is computed in real time with the RHC method, perturbations having arisen in the past can be taken into account. Let us point at some references from the large literature on receding horizon control. For finite-dimensional systems, we mention [13, 19], for infinite-dimensional systems, we mention [2, 3, 11], and for discrete-time systems the articles [10, 14].
In the current framework, the first-order optimality conditions take the usual form of a linear optimality system. The central idea for the derivation of estimate (1.3) is to compare the right-hand sides of the two optimality systems associated with the exact solution of (P ) (restricted to (nτ, nτ + T )) and with the solution to the optimal control problem with short prediction horizon T . This comparison is realized with the help of a priori bounds for linear optimality systems in specific weighted spaces. The analysis of the optimality systems is an important part of the present article. The a priori bounds that we have obtained are of general interest. A classical technique (used in particular in [21, 26]), allowing to decouple the optimality systems, plays an important role.
The article is structured as follows. In Section 2, we prove our error bound in weighted spaces for the optimality systems associated with (P). Some additional properties on linear optimality systems are provided in Section 3. We formulate then the class of linear-quadratic problems to be analyzed in Section 4. The turnpike property and some properties of the value function are then established. Section 5 deals with the RHC method and contains our main result (Theorem 5.2). An extension to infinite-horizon problems is realized in Section 6. Finally, we provide numerical results showing the tightness of our error estimate in Section 7.
1.2. Vector spaces. For T ∈ (0, ∞), we make use of the vector space W (0, T ) = y ∈ L 2 (0, T ; V ) | y ˙ ∈ L 2 (0, T ; V ∗ ) . As it is well-known, W (0, T ) is continuously embedded in C([0, T ], Y ). We can therefore equip it with the following norm:
kyk W (0,T) = max kyk L
2(0,T;V ) , k yk ˙ L
2(0,T;V
∗) , kyk L
∞(0,T;Y )
.
Weighted spaces. Let µ ∈ R be given, let T ∈ (0, ∞). We denote by L 2 µ (0, T ; U ) the space of measurable functions u: (0, T ) → U such that
kuk L
2µ
(0,T ;U) := ke µ· u(·)k L
2(0,T;U) = Z T 0
ke µt u(t)k 2 U dt 1/2
< ∞.
Observing that the mapping u ∈ L 2 µ (0, T ; U ) 7→ e µ· u ∈ L 2 (0, T ; U ) is an isometry, we deduce that L 2 µ (0, T ; U) is a Banach space. Since e µ· is bounded from above and from below by a positive constant, we have that for all measurable u : (0, T ) → U, u ∈ L 2 (0, T ; U ) if and only if u ∈ L 2 µ (0, T ; U ). The spaces L 2 (0, T ; U ) and L 2 µ (0, T ; U ) are therefore the same vector space, equipped with two different norms. We define in a similar way the space L 2 µ (0, T ; X ), for a given Hilbert space X . Similarly, we define the space L ∞ µ (0, T ; Y ) of measurable mappings from y : (0, T ) → Y such that
kyk L
∞µ
(0,T ;Y ) := ke µ· y(·)k L
∞(0,T;Y ) < ∞.
We finally define the Banach space W µ (0, T ) as the space of measurable mappings
y : (0, T ) → V such that e µ· y ∈ W (0, T ). One can check that for all measurable
mappings y : (0, T ) → V , y ∈ W (0, T ) if and only if y ∈ W µ (0, T ).
For T ∈ (0, ∞) and µ ∈ R , we introduce the space
(1.4) Λ T ,µ = W µ (0, T ) × L 2 µ (0, T ; U) × W µ (0, T ),
equipped with the norm k(y, u, p)k Λ
T ,µ= max kyk W
µ(0,T) , kuk L
2µ(0,T;U) , kpk W
µ(0,T)
. For T ∈ (0, ∞), we define the space
(1.5) Υ T ,µ = Y × L 2 µ (0, T ; V ∗ ) × L 2 µ (0, T ; V ∗ ) × L 2 µ (0, T ; U ) × Y that we equip with the norm
k(y 0 , f, g, h, q)k Υ
T ,µ= max ky 0 k Y , k(f, g, h)k L
2µ
(0,T;V
∗×V
∗×U) , e µT kqk Y . Let us emphasize the fact that the component q appears with a weight e µT in the above norm. The spaces Λ T ,0 and Λ T ,µ (resp. Υ T ,0 and Υ T ,µ ) are the same vector space, equipped with two different norms. In the following lemma, the equivalence between these two norms is quantified.
Lemma 1.1. For all µ 0 and µ 1 with µ 0 ≤ µ 1 , there exists a constant M > 0 such that for all T , for all (y, u, p) ∈ Λ T ,0 ,
k(y, u, p)k Λ
T ,µ0≤ M k(y, u, p)k Λ
T ,µ1, k(y, u, p)k Λ
T ,µ1
≤ M e (µ
1−µ
0)T k(y, u, p)k Λ
T ,µ0
, and such that, similarly, for all (y 0 , f, g, h, q) ∈ Υ T ,0 ,
k(y 0 , f, g, h, q)k Υ
T ,µ0≤ M k(y 0 , f, g, h, q)k Υ
T ,µ1,
k(y 0 , f, g, h, q)k Υ
T ,µ1≤ M e (µ
1−µ
0)T k(y 0 , f, g, h, q)k Υ
T ,µ0.
Proof of Lemma 1.1. Let y ∈ W (0, T ) and u ∈ L 2 (0, T ; U ). For proving the lemma, it suffices to prove the existence of M > 0, independent of T , y, and u, such that
(1.6) kuk L
2µ0
(0,T;U ) ≤ M kuk L
2µ1
(0,T ;U) , kuk L
2µ1
(0,T ;U) ≤ M e (µ
1−µ
0)T kuk L
2µ0
(0,T ;U) , and such that
(1.7) kyk W
µ0(0,T) ≤ M kyk W
µ1(0,T) , kyk W
µ1(0,T) ≤ M e (µ
1−µ
0)T kyk W
µ0(0,T ) . The inequalities (1.6) can be easily verified (with M = 1). One can also easily verify that
kyk L
2µ0
(0,T ;V ) ≤ M kyk L
2µ1
(0,T;V ) , kyk L
2µ1
(0,T ;V ) ≤ M e (µ
1−µ
0)T kyk L
2 µ0(0,T;V )
kyk L
∞µ0
(0,T ;Y ) ≤ M kyk L
∞µ1
(0,T;Y ) , kyk L
∞µ1
(0,T ;Y ) ≤ M e (µ
1−µ
0)T kyk L
∞µ0
(0,T;Y ) . Let z 0 (t) = e µ
0t y(t) and z 1 (t) = e µ
1t y(t). For proving (1.7), it remains to compare k z ˙ 0 k L
2(0,T;V
∗) and k z ˙ 1 k L
2(0,T;V
∗) . We have z 0 (t) = e (µ
0−µ
1)t z 1 (t) and thus ˙ z 0 (t) = (µ 0 − µ 1 )z 0 (t) + e (µ
0−µ
1)t z ˙ 1 (t). We deduce that
k z ˙ 0 k L
2(0,T;V
∗) ≤ M kz 0 k L
2(0,T ;V ) + k z ˙ 1 k L
2(0,T;V
∗) ≤ M kz 1 k W (0,T) = M kyk W
µ1
(0,T) .
Similarly, we have ˙ z 1 (t) = (µ 1 − µ 0 )z 1 (t) + e (µ
1−µ
0)t z ˙ 0 (t). We deduce that k z ˙ 1 k L
2(0,T;V
∗) ≤ M kz 1 k L
2(0,T;V ) + e (µ
1−µ
0)T k z ˙ 0 k L
2(0,T;V
∗)
≤ M e (µ
1−µ
0)T kz 0 k W (0,T)
= M e (µ
1−µ
0)T kyk W
µ0
(0,T ) . The inequalities (1.7) follow. This concludes the proof.
1.3. Assumptions. Throughout the article we assume that the following four assumptions hold true.
(A1) The operator −A can be associated with a V -Y coercive bilinear form a : V × V → R which is such that there exist λ 0 > 0 and δ ∈ R satisfying a(v, v) ≥ λ 0 kvk 2 V − δkvk 2 Y , for all v ∈ V .
(A2) [Stabilizability] There exists an operator F ∈ L(Y, U ) such that the semigroup e (A+BF )t is exponentially stable on Y .
(A3) [Detectability] There exists an operator K ∈ L(Z, Y ) such that the semigroup e (A−KC)t is exponentially stable on Y .
Assumptions (A2) and (A3) are well-known and analysed for infinite-dimensional systems, see e.g. [8]. Consider the algebraic Riccati equation: for all y 1 and y 2 ∈ D(A), (1.8) hA ∗ Πy 1 , y 2 i Y + hΠAy 1 , y 2 i Y + hCy 1 , Cy 2 i Z − α 1 hB ∗ Πy 1 , B ∗ Πy 2 i U = 0.
Due to the (exponential) stabilizability and detectability assumptions, it is well-known (see [8, Theorem 6.2.7] and [18, Theorem 2.2.1]) that (1.8) has a unique nonnegative self-adjoint solution Π ∈ L(Y, V ) ∩ L(V ∗ , Y ). Additionally, the semigroup generated by the operator A π := A − α 1 BB ∗ Π is exponentially stable on Y . We fix now, for the rest of the article, a real number λ such that
0 < λ < ¯ λ := −sup µ∈σ(A
π) Re(µ).
(1.9)
With (A1) holding the operator A associated with the form a generates an analytic semigroup that we denote by e At , see e.g. [24, Sections 3.6 and 5.4]. Let us set A 0 = A − λ 0 I. Then −A 0 has a bounded inverse in Y , see [24, page 75], and in particular it is maximal accretive, see [24]. We have D(A 0 ) = D(A) and the fractional powers of
−A 0 are well-defined. In particular, D((−A 0 )
12) = [D(−A 0 ), Y ]
12
:= (D(−A 0 ), Y )
1 2,2
the real interpolation space with indices 2 and 1 2 , see [5, Proposition 6.1, Part II, Chapter 1]. Assumption (A4) below will only be used in the proof Lemma 4.1, where the existence and uniqueness of a solution (y , u ) to the static problem is established.
It is not necessary for the analysis of optimality systems done in Sections 2 and 3.
(A4) It holds that [D(−A 0 ), Y ]
12
= [D(−A ∗ 0 ), Y ]
1 2= V .
2. Linear optimality systems. The section is dedicated to the analysis of the following optimality system:
(2.1)
y(0) = y 0 in Y
˙
y − (Ay + Bu) = f in L 2 µ (0, T ; V ∗ )
− p ˙ − A ∗ p − C ∗ Cy = g in L 2 µ (0, T ; V ∗ ) αu + B ∗ p = −h in L 2 µ (0, T ; U ) p(T ) − Qy(T ) = q in Y ,
where µ ∈ {−λ, 0, λ}, T > 0, Q ∈ L(Y ) is self-adjoint and positive semi-definite,
and (y 0 , f, g, h, q) ∈ Υ T ,µ . Given two times t 1 < t 2 , we introduce the operator
H: W (t 1 , t 2 ) × L 2 (t 1 , t 2 ; U) × W (t 1 , t 2 ) → L 2 (t 1 , t 2 ; V ∗ × V ∗ × U ), defined by H(y, u, p) = y ˙ − (Ay + Bu), − p ˙ − A ∗ p − C ∗ Cy, αu + B ∗ p
.
The dependence of H with respect to t 1 and t 2 is not indicated and the underlying values of t 1 and t 2 are always clear from the context. The operator H enables us to write the three intermediate equations of (2.1) in the compact form H(y, u, p) = (f, g, −h).
The main result of the section is the following theorem, which is proved in sub- section 2.2.
Theorem 2.1. Let Q ⊂ L(Y ) be a bounded set of self-adjoint and positive semi- definite operators. For all T > 0, for all Q ∈ Q, for all (y 0 , f, g, h, q) ∈ Υ T ,0 , there exists a unique solution (y, u, p) to system (2.1). Moreover, for all µ ∈ {−λ, 0, λ}, there exists a constant M independent of T, Q, and (y 0 , f, g, h, q) such that
(2.2) k(y, u, p)k Λ
T ,µ≤ M k(y 0 , f, g, h, q)k Υ
T ,µ.
Remark 2.2. The result of the theorem, for µ = 0, is rather classical in the litera- ture and can be established by analyzing the associated optimal control problem (see Lemma 2.8). The main novelty of our result is the estimate (2.2) in weighted spaces, with a constant M which is independent of T . Let us mention that a similar result has been obtained in [15, Theorem 3.1], for negative weights. The proof is based on a Neumann-series argument. Let us mention that the range of admissible weights in that reference is different from ours (compare in particular with [15, Corollary 3.16]).
2.1. Decouplable optimality systems. We prove in this subsection Theorem 2.1 in the case where Q = Π (Lemma 2.5). We begin with a useful result on forward and backward linear systems with a right-hand side in L 2 µ (0, T ; V ∗ ) (Lemma 2.4).
Lemma 2.3. For all µ ≤ λ, A π + µI generates an exponentially stable semigroup.
For all µ ≥ −λ, A ∗ π − µI generates an exponentially stable semigroup.
Proof. Let ˜ λ ∈ (λ, ¯ λ). Since the semigroup e A
πt is analytic, the spectrum de- termined growth condition is satisfied, see e.g. [27]. Hence, ke A
πt k L(Y ) ≤ M e − λt ˜ , where M does not depend on t. Therefore, ke (A
π+µI)t k L(Y ) ≤ M e (− ˜ λ+µ)t , which proves the exponential stability of A π + µI since − ˜ λ + µ < −λ + µ ≤ 0. Moreover, (e (A
π+µI)t ) ∗ = e (A
π+µI)
∗t (see [20, page 41]), thus the operator A ∗ π − µI generates a exponentially stable semigroup as well, for µ ≥ −λ.
Lemma 2.4. For all µ ≤ λ, for all T ∈ (0, ∞), for all y 0 ∈ Y , for all f ∈ L 2 µ (0, T ; V ∗ ), the following system:
(2.3) y ˙ = A π y + f, y(0) = y 0
has a unique solution in W µ (0, T ). Moreover, there exists a constant M > 0 indepen- dent of T , y 0 , and f such that kyk W
µ(0,T) ≤ M ky 0 k Y + kf k L
2µ
(0,T;V
∗)
.
For all µ ≥ −λ, for all T ∈ (0, ∞), for all q ∈ Y , for all Φ ∈ L 2 µ (0, T ; V ∗ ), the following system: − r ˙ = A ∗ π r + Φ, r(T) = q has a unique solution in W µ (0, T ).
Moreover, there exists a constant M > 0 independent of T , q, and Φ such that krk W
µ(0,T) ≤ M kΦk L
2µ(0,T;V
∗) + e µT kqk Y
.
Proof. Let us prove the first statement. Let y ∈ W (0, T ). Defining y µ := e µ· y ∈
W µ (0, T ) and f µ := e µ· f ∈ L 2 µ (0, T ; V ∗ ), we observe that y solves (2.3) if and only if
y µ is the solution to the following system:
(2.4) y ˙ µ = (A π + µI)y µ + f µ , y µ (0) = y 0 .
Since µ ≤ λ, the operator A π + µI generates an exponentially stable semigroup, by Lemma 2.3. Standard regularity results for analytic semigroups ensure the existence and uniqueness of a solution to (2.4), as well as the existence of a constant M > 0 independent of T , y 0 , and f such that ky µ k W (0,T) ≤ M ky 0 k + kf µ k L
2(0,∞)
, which is the estimate that was to be proved.
The second statement can be proved similarly with a time-reversal argument.
We are now ready to analyze (2.1) in the case where Q = Π. The key idea is to decouple the system with the help of the variable r = p − Πy. This variable is indeed the solution to a backward differential equation which is independent of y, u, and p.
Let us mention that this remarkable property only holds in the case Q = Π.
Lemma 2.5. For all µ ∈ [−λ, λ], for all T > 0, for all (y 0 , f, g, h, q) ∈ Υ T ,µ , there exists a unique (y, u, p) ∈ Λ T ,µ solution to (2.1) with Q = Π. Moreover, there exists a constant M > 0, independent of T and (y 0 , f, g, h, q) such that
(2.5) k(y, u, p)k Λ
T ,µ≤ M k(y 0 , f, g, h, q)k Υ
T ,µ.
Remark 2.6. All along the article, the variable M is a positive constant whose value may change from an inequality to the next one. When an estimate involving a constant M independent of some variables (for example T ) has to be proved, then all constants M used in the corresponding proof are also independent of these variables.
Proof of Lemma 2.5. Let Φ ∈ L 2 µ (0, T ; V ∗ ) be defined by Φ = Πf − α 1 ΠBh + g.
Let us denote by r ∈ W µ (0, T ) the unique solution to the system −˙ r = A ∗ π r + Φ, t ∈ [0, T ), r(T ) = q. By Lemma 2.4, there exists a constant M , independent of T and (y 0 , f, g, h, q) such that
(2.6) krk W
µ(0,T) ≤ M kΦk L
2µ
(0,T ;V
∗) + e µT kqk Y
≤ M k(y 0 , f, g, h, q)k Υ
T ,µ. By Lemma 2.4, the following system has a unique solution y ∈ W µ (0, T ):
(2.7) y ˙ = A π y − α 1 Bh + f − α 1 BB ∗ r, y(0) = y 0 . Since
− α 1 Bh + f − α 1 BB ∗ r L
2µ
(0,T;V
∗) ≤ M k(y 0 , f, g, h, q)k Υ
T ,µ, we have that (2.8) kyk W
µ(0,T) ≤ M k(y 0 , f, g, h, q)k Υ
T ,µ.
Let us set p = Πy + r. Since Π ∈ L(Y, V ) ∩ L(V ∗ , Y ), we have that Πy ∈ L 2 (0, T ; V ) ∩ H 1 (0, ∞; Y ). Therefore, using (2.6) and (2.8), we obtain that p ∈ W µ (0, T ) with kpk W
µ(0,T ) ≤ M k(y 0 , f, g, h, q)k Υ
T ,µ. We finally define u = − α 1 (h + B ∗ p). We deduce from the estimate on p that kuk L
2µ
(0,T;U ) ≤ M k(y 0 , f, g, h, q)k Υ
T ,µ. The bound (2.5) is proved.
Let us check that (y, u, p) is a solution to the linear system (2.1). It follows from the definition of u that αu + B ∗ p = −h. Using p = Πy + r and (2.7), we obtain that
Ay + Bu + f = Ay − α 1 Bh − α 1 BB ∗ p + f
= Ay − α 1 Bh − α 1 BB ∗ Πy − α 1 BB ∗ r + f = ˙ y.
It remains to verify that the adjoint equation is satisfied. We obtain with the defini- tions of p, y, Φ, and A π that p(T ) − Πy(T ) = q and that
˙
p = Π ˙ y + ˙ r = ΠA π y + − α 1 ΠBh + Πf
| {z }
=Φ−g
+ − α 1 ΠBB ∗
| {z }
=A
∗π−A
∗r + ˙ r.
Using ˙ r + A ∗ π r + Φ = 0 and (1.8), we obtain that
˙
p = ΠA π y + Φ − g + A ∗ π r − A ∗ r + ˙ r
= ΠA π y − A ∗ r − g = ΠA − 1 α ΠBB ∗ Π
y − A ∗ r − g
= − A ∗ Π + C ∗ C
y − A ∗ r − g = −A ∗ p − C ∗ Cy − g.
Therefore, the adjoint equation is satisfied and (y, u, p) is a solution to (2.2).
It remains to show uniqueness. To this end, it suffices to consider the case where (y 0 , f, g, h, q) = (0, 0, 0, 0, 0). Let (y, u, p) be a solution to (2.1). Let r = p − Πy. One can easily see that − r ˙ = A ∗ π r, r(T ) = 0, thus r = 0. Then, one has to check that y satisfies (2.7), with y 0 = 0, f = 0, h = 0, and r = 0. Therefore, y = 0. Finally, we obtain that p = r + Πy = 0 and that u = − 1 α (h + B ∗ p) = 0. Uniqueness is proved.
2.2. General case. We give a proof of Theorem 2.1 in this subsection. We consider successively the cases µ = 0, µ = −λ, and µ = λ.
2.2.1. Case without weight. Theorem 2.1, in the case where µ = 0, can be established by analyzing the optimal control problem associated with (2.1). This is the result of Lemma 2.8 below. The proof is classical and uses very similar arguments to the ones used in [6, Proposition 3.1].
We begin with a classical lemma, following from the detectability assumption.
Lemma 2.7. There exists a constant M > 0 such that for all T > 0, for all y 0 ∈ Y , for all u ∈ L 2 (0, T ; U ), for all f ∈ L 2 (0, T ; V ∗ ), the solution y ∈ W (0, T ) to the system
˙
y = Ay + Bu + f, y(0) = y 0 satisfies the following estimate:
kyk W (0,T) ≤ M ky 0 k Y + kuk L
2(0,T;U ) + kf k L
2(0,T;V
∗) + kCyk L
2(0,T ;Z)
. Proof. Let z ∈ W (0, T ) be the solution to
˙
z = Az + Bu + f + KC (y − z), z(0) = y 0 ,
where K is given by Assumption (A3). The above system can be re-written as follows:
˙
z = (A − KC)z + Bu + f + KCy, z(0) = y 0 .
Since (A − KC ) is exponentially stable, there exists a constant M , independent of T, y 0 , u, f , and y such that
kzk W(0,T) ≤ M ky 0 k Y + kBu + f + KCyk L
2(0,T;V
∗)
≤ M ky 0 k Y + kuk L
2(0,T;U) + kf k L
2(0,T;V
∗) + kCyk L
2(0,T;Z)
. (2.9)
Observing that e := z − y is the solution to ˙ e = (A − KC )e, e(0) = 0, we obtain that
e = 0 and that z = y. Thus y satisfies (2.9), as was to be proved.
Lemma 2.8. For all T > 0, for all Q ∈ Q, for all (y 0 , f, g, h, q) ∈ Υ T ,0 , the following optimal control problem
(LQ)
y∈W inf (0,T) u∈L
2(0,T;U)
h Z T 0
1
2 kCy(t)k 2 Z + hg(t), y(t)i + α 2 ku(t)k 2 U + hh(t), u(t)i U dt + 1 2 hy(T ), Qy(T )i Y + hq, y(T )i Y
i , subject to: y ˙ = Ay + Bu + f, y(0) = y 0 ,
has a unique solution (y, u). There exists a unique associated adjoint variable p, which is such that (y, u, p) is the unique solution to (2.1). Moreover, there exists a constant M , independent of T , Q, and (y 0 , f, g, h, q) such that
(2.10) k(y, u, p)k Λ
T ,0≤ M k(y 0 , f, g, h, q)k Υ
T ,0.
Proof. We follow the same lines as in [6, Lemma 3.2]. Let us first bound the value of the problem. Let y ∈ W (0, T ) be the solution to
˙
y = (A + BF )y + f, y(0) = y 0 ,
where F is given by Assumption (A2). Since (A + BF) is exponentially stable, there exists a constant M such that
kyk W(0,T) ≤ M max ky 0 k Y , kf k L
2(0,T;V
∗)
.
Let us set u = F y. We have kuk L
2(0,T;U) ≤ M max ky 0 k Y , kf k L
2(0,T;V
∗)
. Then, one can easily check the existence of a constant M such that
J 1 (u, y) ≤ M k(y 0 , f, g, h, q)k 2 Υ
T ,0.
Now, we prove the existence of a solution to the problem. Let (y n , u n ) n∈N ∈ W (0, T )×
L 2 (0, T ; U ) be a minimizing sequence such that for all n ∈ N , J 1 (y n , u n ) ≤ M k(y 0 , f, g, h, q)k 2 Υ
T ,0.
We now look for a lower bound for J 1 , so that we can further obtain a bound on (y n , u n ). We have
J 1 (y n , u n ) ≥ 1
2 kCy n k 2 L
2(0,T;Z) − kgk L
2(0,T;V
∗) ky n k W (0,T ) + α
2 ku n k 2 L
2(0,T;U) − khk L
2(0,T;U ) ku n k L
2(0,T;U) − kqk Y ky n (T )k Y
≥ 1
2 kCy n k 2 L
2(0,T;Z) + α 2
ku n k 2 L
2(0,T;U ) − khk L
2(0,T ;U)
α 2
− khk 2 L
2(0,T;U)
2α
− 1
2ε kgk L
2(0,T;V
∗) + kqk Y
2
− ε
2 ky n k 2 W(0,T) . Therefore, there exists a constant M such that
kCy n k ≤ M
k(y 0 , f, g, h, q)k Υ
T ,0+ 1
√ ε kgk L
2(0,T;V
∗) + kqk Y + √
εky n k W (0,T) , (2.11)
ku n k ≤ M
k(y 0 , f, g, h, q)k Υ
T ,0+ 1
√ ε kgk L
2(0,T;V
∗) + kqk Y + √
εky n k W (0,T)
.
(2.12)
Applying Lemma 2.7 and estimate (2.11), we obtain that
ky n k W (0,T) ≤ M ky 0 k Y + kf k L
2(0,T;V
∗) + ku n k L
2(0,T;U) + kCy n k L
2(0,T ;Z)
≤ M
k(y 0 , f, g, h, q)k Υ
T ,0+ √
εky n k W(0,T) + 1
√ ε kgk + kqk Y . Let us fix ε = (2M 1 )
2, where M is the constant obtained in the last inequality. It follows that there exists (another) constant M > 0 such that
(2.13) ky n k W (0,T ) ≤ M k(y 0 , f, g, h, q)k Υ
T ,0. Combined with (2.12), we obtain that
ku n k L
2(0,T;U) ≤ M k(y 0 , f, g, h, q)k Υ
T ,0.
The sequence (y n , u n ) n∈ N is therefore bounded in W (0, T ) × L 2 (0, T ; U ) and has a weak limit point (y, u) satisfying
(2.14) max kyk W (0,T ) , kuk L
2(0,T;U )
≤ M k(y 0 , f, g, h, q)k Υ
T ,0.
One can prove the optimality of (y, u) with the same techniques as those used for the proof of [7, Proposition 2].
Consider now the solution p to the adjoint system
(2.15) − p ˙ − A ∗ p − C ∗ Cy = g, p(T ) − Qy(T ) = q.
The optimality conditions for the problem yield αu + B ∗ p + h = 0, see e.g. [16]. It follows that (y, u, p) is a solution to (2.1).
Let us prove the uniqueness. If (y, u, p) is a solution to (2.1), then one can prove that (y, u) is a solution to problem (LQ) with associated costate p. Therefore, it suffices to prove the uniqueness of the solution to (2.1). To this end, it suffices to consider the case where (y 0 , f, g, h, q) = (0, 0, 0, 0, 0). Let (y, u, p) be a solution to (2.1). Then (y, u) is a solution to (LQ) and one can check that (2.14) holds. Thus, (y, u) = (0, 0) and then, p = 0, which proves the uniqueness.
It remains to prove the a priori bound. Observe that (y, u, p) is the solution to
(2.16)
y(0) = y 0 in Y
˙
y − (Ay + Bu) = f in L 2 (0, T ; V ∗ )
− p ˙ − A ∗ p − C ∗ Cy = g in L 2 (0, T ; V ∗ ) αu + B ∗ p = −h in L 2 (0, T ; U ) p(T ) − Πy(T ) = ˜ q in Y ,
where ˜ q = (Q − Π)y(T ) + q. By (2.14), we have k qk ˜ Y ≤ M k(y 0 , f, g, h, q)k Υ
T ,0. Thus by Lemma 2.5, k(y, u, p)k Λ
T ,0≤ M k(y 0 , f, g, h, q)k ˜ Υ
T ,0≤ M k(y 0 , f, g, h, q)k Υ
T ,0, which concludes the proof.
2.2.2. Case of a negative weight.
Proof of Theorem 2.1: the case µ = −λ. Let (y 0 , f, g, h, q) ∈ Υ T ,−λ . The fol- lowing inequality can be easily checked: k(f, g, h)k L
2(0,T) ≤ e λT k(f, g, h)k L
2−λ
(0,T) . Therefore, by Lemma 2.8, the system (2.1) has a unique solution (y, u, p), satisfying
k(y, u, p)k Λ
T ,0≤ M max ky 0 k Y , k(f, g, h)k L
2(0,T) , kqk Y
≤ M max ky 0 k Y , e λT k(f, g, h)k L
2−λ
(0,T ) , kqk Y
.
It follows that ky(T )k Y ≤ M max ky 0 k Y , e λT k(f, g, h)k L
2−λ
(0,T ) , kqk Y
and then that e −λT ky(T)k Y ≤ M max e −λT ky 0 k Y , k(f, g, h)k L
2−λ
(0,T ) , e −λT kqk Y
≤ M k(y 0 , f, g, h, q)k Υ
T ,−λ, (2.17)
since e −λT ≤ 1. The key idea now is to observe that y(0) = y 0 , H(y, u, p) = (f, g, −h), and p(T ) − Πy(T ) = ˜ q, where ˜ q = (Q − Π)y(T ) + q. Thus, by Lemma 2.5,
k(y, u, p)k Λ
T ,−λ≤ M k(y 0 , f, g, h, q)k ˜ Υ
T ,−λ≤ M k(y 0 , f, g, h, q)k Υ
T ,−λ+ e −λT k(Q − Π)k L(Y ) ky(T )k Y
≤ M k(y 0 , f, g, h, q)k Υ
T ,−λ+ M e −λT ky(T)k Y , (2.18)
since Q is bounded. Estimate (2.2) follows, combining (2.17) and (2.18).
2.2.3. Case of positive weight. The approach that we propose for dealing with the case µ = λ requires some more advanced tools, that we introduce now. For a given θ ∈ (0, T ), we make use of the following mixed weighted space:
kuk L
2λ,−λ
(0,T;U ) = ke ρ(·) u(·)k L
2(0,T;U) , where
ρ(t) = λt, for t ∈ [0, T − θ], ρ(t) = 2λ(T − θ) − λt, for t ∈ [T − θ, T ].
Observe that ρ is continuous and piecewise affine, with ˙ ρ(t) = λ for t ∈ [0, T − θ) and
˙
ρ(t) = −λ for t ∈ (T −θ, T ]. In a nutshell: We use a positive weight on (0, T −θ) and a negative weight on (T −θ, T ). We define similarly the space L 2 −λ,λ (0, T ; V ∗ ×V ∗ × U )
— that we often denote by L 2 −λ,λ (0, T ) — and the space W λ,−λ (0, T ). The spaces Λ λ,−λ and Υ λ,−λ are defined in a similar way as before, with the corresponding norms
k(y, u, p)k Λ
λ,−λ= max kyk W
λ,−λ(0,T ) , kuk L
2λ,−λ
(0,T;U ) , kpk W
λ,−λ(0,T)
, k(y 0 , f, g, h, q)k Υ
λ,−λ= max ky 0 k Y , k(f, g, h)k L
2λ,−λ
(0,T) , e ρ(T) kqk Y
.
The following lemma is a generalization of Lemma 2.5 for mixed weighted spaces.
Lemma 2.9. For all T > 0, for all (y 0 , f, g, h, q) ∈ Υ λ,−λ , the unique solution (y, u, p) to (2.1) with Q = Π satisfies the following bound:
(2.19) k(y, u, p)k Λ
λ,−λ≤ M k(y 0 , f, g, h, q)k Υ
λ,−λ, where M is independent of T , θ, and (y 0 , f, g, h, q).
Proof. We only give the main lines of the proof. One can obtain estimate (2.19) with the same decoupling as the one introduced in Lemma 2.5. The decoupled vari- ables y and r can then be estimated in W λ,−λ (0, T ), after an adaptation of Lemma 2.4 for right-hand sides in L 2 λ,−λ (0, T ; V ∗ ).
Proof of Theorem 2.1: the case µ = λ. Let us first fix some constants. We denote
by M 1 the constant involved in estimate (2.10). We denote by M 2 the constant
involved in Lemma 2.9. Note that M 1 ≥ 1 and M 2 ≥ 1. Finally, M 3 denotes an
upper bound on kQ − Πk L(Y ) . Let us set M 0 = 2M 1 M 2 ≥ 1 and let us fix θ > 0
such that M 0 M 3 e −λθ ≤ 1. The first four steps of this proof deal with the case
where T ≥ θ. We will consider the case T < θ in Step 5. Take now T ≥ θ and (y 0 , f, g, h, q) ∈ Υ T ,λ . Since Υ T ,λ is embedded in Υ T ,0 , the existence of a solution to (2.1) in Λ T ,0 is guaranteed. Let us denote it by (¯ y, u, ¯ p). ¯
Step 1: construction of the mappings χ 1 and χ 2 .
The main idea of the proof consists in obtaining an estimate of ¯ y(T ) with a fixed- point argument. To this end, we introduce two affine mappings, χ 1 and χ 2 , defined as follows: χ 1 : y T ∈ Y 7→ y(T − θ) ∈ Y , where y is the solution to
(2.20)
y(0) = y 0 in Y
H(y, u, p) = (f, g, −h) in L 2 λ,−λ (0, T ; V ∗ × V ∗ × U) p(T ) − Πy(T) = (Q − Π)y T + q in Y .
The mapping χ 2 is defined as follows: χ 2 : y T −θ ∈ Y 7→ y(T ) ∈ Y , where y ∈ W (T − θ, T ) is the solution to
y(T − θ) = y T −θ in Y
H(y, u, p) = (f, g, −h) in L 2 (T − θ, T; V ∗ × V ∗ × U ) p(T ) − Qy(T) = q in Y .
The existence and uniqueness of a solution to the above system follows from Lemma 2.8, after a shifting of the time variable. Observe that ¯ y(T − θ) = χ 1 (¯ y(T)) and that
¯
y(T) = χ 2 (¯ y(T − θ)). It follows that ¯ y(T ) is a fixed point of χ 2 ◦ χ 1 . Step 2: on the Lipschitz-continuity of χ 1 and χ 2 .
Let y T and ˜ y T ∈ Y . We have χ 1 (˜ y T ) − χ 1 (y T ) = y(T − θ), where y is the solution to
y(0) = 0 in Y
H(y, u, p) = (0, 0, 0) in L 2 λ,−λ (0, T ; V ∗ × V ∗ × U ) p(T ) − Πy(T) = (Q − Π)(˜ y T − y T ) in Y .
By Lemma 2.9,
k(y, u, p)k Λ
λ,−λ≤ M 2 e ρ(T) kQ − Πk L(Y ) k˜ y T − y T k Y ≤ M 2 M 3 e ρ(T ) k y ˜ T − y T k Y . Thus, e ρ(T −θ) ky(T − θ)k Y ≤ kyk W
λ,−λ(0,T) ≤ M 2 M 3 e ρ(T) k˜ y T − y T k Y . Observing that e ρ(T)−ρ(T −θ) = e −λθ , we finally obtain that
kχ 1 (˜ y T ) − χ 1 (y T )k Y = ky(T − θ)k Y ≤ M 2 M 3 e −λθ k y ˜ T − y T k Y ,
which proves that χ 1 is Lipschitz-continuous. Now, let us take y T −θ and ˜ y T−θ in Y . We have χ 2 (˜ y T −θ ) − χ 2 (˜ y T −θ ) = y(T), where y ∈ W (T − θ, T ) is the solution to
y(T − θ) = ˜ y T −θ − y T −θ in Y
H(y, u, p) = (0, 0, 0) in L 2 (T − θ, T ; V ∗ × V ∗ × U ) p(T) − Qy(T ) = 0 in Y .
We obtain with Lemma 2.8 that kyk W (T −θ,T ) ≤ M 1 k y ˜ T −θ − y T −θ k Y and thus kχ 2 (˜ y T−θ ) − χ 1 (y T−θ )k Y = ky(T )k ≤ M 1 k y ˜ T −θ − y T−θ k Y ,
proving that χ 2 is Lipschitz-continuous. As a consequence, the mapping χ 2 ◦ χ 1 is
Lipschitz-continuous, with modulus M 1 M 2 M 3 e −λθ ≤ 1 2 M 0 M 3 e −λθ ≤ 1 2 .
Step 3: on the invariance of B Y R
, with R = M 0 e −λ(T−θ) k(y 0 , f, g, h, q)k Υ
T ,λ. Let y T ∈ B Y (R). Consider the solution y to system (2.20). By Lemma 2.9, we have (2.21) kyk W
λ,−λ(0,T) ≤ M 2 max ky 0 k Y , k(f, g, h)k L
2λ,−λ
(0,T ) , e ρ(T ) k(Q−Π)y T +qk Y
. Let us estimate the last term in the above expression. We have
e ρ(T) k(Q − Π)y T + qk Y
≤ e λT−2λθ M 3 ky T k Y + kqk Y
≤ e −λθ M 0 M 3 k(y 0 , f, g, h, q)k Υ
T ,λ+ e λT kqk Y
≤ k(y 0 , f, g, h, q)k Υ
T ,λ+ e λT kqk Y . (2.22)
Observe that k(f, g, h)k L
2λ,−λ
(0,T ) ≤ k(f, g, h)k L
2λ
(0,T ) . Combining (2.21), (2.22), and this last observation, we obtain that
kyk W
λ,−λ(0,T) ≤ M 2 max ky 0 k Y , k(f, g, h)k λ , k(y 0 , f, g, h, q)k Υ
T ,λ+ e λT kqk Y
≤ 2M 2 k(y 0 , f, g, h, q)k Υ
T ,λ. It follows then that
kχ 1 (y T )k Y = e −ρ(T −θ) ke ρ(T−θ) y(T − θ)k Y
≤ e −λ(T−θ) kyk W
λ,−λ(0,T)
≤ 2M 2 e −λ(T −θ) k(y 0 , f, g, h, q)k Υ
T ,λ. (2.23)
Applying now Lemma 2.8, we obtain that
kχ 2 ◦ χ 1 (y T )k Y ≤ M 1 max kχ 1 (y T )k Y , k(f, g, h) |(T −θ,T) k 0 , kqk Y . (2.24)
Observing that e λ(T −θ) k(f, g, h) |(T −θ,T) k L
2(T−θ,T) ≤ k(f, g, h)k L
2λ