• Aucun résultat trouvé

Non-coercive first order Mean Field Games

N/A
N/A
Protected

Academic year: 2021

Partager "Non-coercive first order Mean Field Games"

Copied!
31
0
0

Texte intégral

(1)

HAL Id: hal-01938798

https://hal.archives-ouvertes.fr/hal-01938798

Submitted on 29 Nov 2018

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Non-coercive first order Mean Field Games

Paola Mannucci, Claudio Marchi, Carlo Mariconda, Nicoletta Tchou

To cite this version:

Paola Mannucci, Claudio Marchi, Carlo Mariconda, Nicoletta Tchou. Non-coercive first order Mean Field Games. Journal of Differential Equations, Elsevier, 2020, 269 (5), pp.4503-4543.

�10.1016/j.jde.2020.03.035�. �hal-01938798�

(2)

Non-coercive first order Mean Field Games

Paola Mannucci, Claudio Marchi , Carlo Mariconda, Nicoletta Tchou§

Abstract

We study first order evolutive Mean Field Games where the Hamiltonian is non- coercive. This situation occurs, for instance, when some directions are “forbidden” to the generic player at some points. We establish the existence of a weak solution of the system via a vanishing viscosity method and, mainly, we prove that the evolution of the population’s density is the push-forward of the initial density through the flow characterized almost everywhere by the optimal trajectories of the control problem underlying the Hamilton-Jacobi equation. As preliminary steps, we need that the optimal trajectories for the control problem are unique (at least for a.e. starting points) and that the optimal controls can be expressed in terms of the horizontal gradient of the value function.

Keywords: Mean Field Games, first order Hamilton-Jacobi equations, continuity equa- tion, non-coercive Hamiltonian, degenerate optimal control problem.

2010 AMS Subject classification: 35F50, 35Q91, 49K20, 49L25.

1 Introduction

In this paper we study the following Mean Field Game (briefly, MFG)

(1.1)

(i) −tu+H(x, Du) =F(x, m) inR2×(0, T) (ii) tm−div(m ∂pH(x, Du)) = 0 inR2×(0, T) (iii) m(x,0) =m0(x), u(x, T) =G(x, m(T)) onR2,

where, if p= (p1, p2) and x= (x1, x2), the functionsH(x, p) is

(1.2) H(x, p) = 1

2(p21+h2(x1)p22)

whereh(x1) is a regular bounded functionpossibly vanishingand thatF andGare strongly regularizing (see assumptions (H1) – (H4) below).

These MFG systems arise when the dynamics of the generic player are deterministic and, when h vanishes, may have a “forbidden” direction; actually, if the evolution of

Dipartimento di Matematica “Tullio Levi-Civita”, Università di Padova, mannucci@math.unipd.it

Dipartimento di Ingegneria dell’Informazione, Università di Padova, claudio.marchi@unipd.it

Dipartimento di Matematica “Tullio Levi-Civita”, Università di Padova, carlo.mariconda@unipd.it

§Univ Rennes, CNRS, IRMAR - UMR 6625, F-35000 Rennes, France, nicoletta.tchou@univ-rennes1.fr

(3)

the whole population’s distribution m is given, each agent wants to choose the control α= (α1, α2) in L2([t, T];R2) in order to minimize the cost

(1.3)

Z T

t

1

2|α(τ)|2+F(x(τ), m)

+G(x(T), m(T)) where, in [t, T], its dynamics x(·) are governed by

(1.4)

( x1(s) =α1(s)

x2(s) =h(x1(s))α2(s)

with x1(t) = x1 and x2(t) = x2. We see that the direction along x2 is forbidden when h(x1) has zero value. This kind of problems are called of “Grushin type” (see [27] or Example 1.1 below). As a matter of fact the structure of this degenerate dynamics will play an essential rule in our results. Even though our techniques apply to a wider class of degenerate operators (see the forthcoming papers [28, 29]), in the present paper we restrict our attention to this class of problems because they already contain all the main technical issues.

Let us recall that the MFG theory studies Nash equilibria in games with a huge number of (“infinitely many”) rational and indistinguishable agents. This theory started with the pioneering papers by Lasry and Lions [23, 24, 25] and by Huang, Malhamé and Caines [19]. A detailed description of the achievements obtained in these years goes beyond the scope of this paper; we just refer the reader to the monographs [1, 9, 5, 17, 18].

As far as we know, degenerate MFG systems have been poorly investigated up to now. Dragoni and Feleqi [16] studied a second order (stationary) system where the principal part of the operator fulfills the Hörmander condition; moreover, Cardaliaguet, Graber, Porretta and Tonon [11] tackled degenerate second order systems with coercive (and convex as well) first order operators. Hence, these results cannot be directly applied to the non-coercive problem (1.1).

The aim of this paper is to prove the existence of a solution of (1.1). The main result is the interpretation of the evolution of the population’s density as the push-forward of the distribution at the initial time through a flow which is suitably defined in terms of the optimal control problem underlying Hamilton Jacobi equation.

In order to establish a representation formula for m, we shall follow some ideas of P-L Lions in the lectures at College de France (2012) (see [9]), some results proved in [12, 10] and the Ambrosio superposition principle [2]. Indeed the non-coercivity of H prevents from applying directly the arguments of [9, Sect. 4.3]. Actually we have to study carefully the behaviour of the optimal trajectories of the control problem associated to the Hamilton-Jacobi equation (1.1)-(i) especially their uniqueness. A crucial point will be the application of the Pontryagin maximum principle and the statement of Theorem 2.1 on the uniqueness of the optimal trajectory after the rest time. As far as we know this uniqueness property has never been tackled before for this kind of degenerate dynamics and in our opinion may have interest in itself.

We point out that our approach could be applied to other first order “degenerate”

MFG but it is essential to prove some uniqueness properties of optimal trajectories in a set of starting points of full measure. In general this set depends on the semiconcavity properties ofu, as in the classical setting, and on the degeneracy of the dynamics.

We now list our notations and the assumptions, we give the definition of (weak) solution to system (1.1) and we state the existence result for system (1.1).

(4)

Notations and Assumptions. For x = (x1, x2) ∈ R2, φ : R2 → R and Φ : R2 → R2 differentiable, we set: DGφ(x) := (∂x1φ(x), h(x1)∂x2φ(x)) and divGΦ(x) := x1Φ1(x) + h(x1)∂x2Φ2(x).We denote byP1 the space of Borel probability measures onRdwith finite first order moment, endowed with the Kantorovich-Rubinstein distance d1. We denote C2(R2) the space of functions with continuous second order derivatives endowed with the norm kfkC2 := supx∈R2[|f(x)|+|Df(x)|+|D2f(x)|]. Throughout this paper, we shall require the following hypotheses:

(H1) The functions F,·) and G(·,·) are real-valued function, continuous on R2× P1;

(H2) The map F(x,·) is Lipschitz continuous from P1 to C2(R2) uniformly for x ∈ R2; moreover, there existsC∈Rsuch that

kF(·, m)kC2,kG(·, m)kC2C,m∈ P1, (H3) the function h:R→R isC2(R) withkhkC2C;

(H4) the initial distribution m0 has a compactly supported density (that we still denote by m0, with a slight abuse of notation), m0C2,δ(R2), for aδ∈(0,1).

Example 1.1 Easy examples ofh areh(x1) = sin(x1) or h(x1) = √x1

1+x21, (see [27] where the term h(x1) = √x1

1+x21 is introduced as a degenerate diffusion term).

We now introduce our definition of solution of the MFG system (1.1) and state the main result concerning its existence.

Definition 1.1 The pair (u, m) is a solution of system (1.1) if:

1) (u, m)∈W1,∞(R2×[0, T])×C0([0, T];P1(R2));

2) Equation (1.1)-(i) is satisfied byu in the viscosity sense;

3) Equation (1.1)-(ii) is satisfied bym in the sense of distributions.

Here below we state the main result of this paper.

Theorem 1.1 Under the above assumptions:

1. System (1.1) has a solution (u, m) in the sense of Definition 1.1, 2. m is the push-forward of m0 through the characteristic flow

(1.5)

( x1(s) =−ux1(x(s), s), x1(0) =x1, x2(s) =−h2(x1(s))ux2(x(s), s), x2(0) =x2 .

Remark 1.1 Uniqueness holds under classical hypothesis on the monotonicity of F and G as in [9].

(5)

This paper is organized as follows. In Section 2 we will find some properties of the solution u of the Hamilton-Jacobi equation (1.1)-(i) with fixedm: we will prove that u is Lipschitz continuous and semiconcave inx. Moreover still in this section we will establish a crucial point of the paper: the uniqueness of the optimal trajectory of the associated control problem. In Section 3 we study the continuity equation (1.1)-(ii) whereuis the solution of a Hamilton-Jacobi equation found in the previous section. Section 4 is devoted to the proof of the Theorem 1.1. Finally, the Appendix splits into two parts: in the former one, we give some results on the concatenation of optimal trajectories and the Dynamic Programming Principle while in the latter part we introduce the notion of theG-differentiability and we prove the main properties on theG-differentials which will be used along the paper.

2 Formulation of the optimal control problem

For every 0 ≤ t < T and x := (x1, x2) ∈ R2 we consider the following optimal control problem, where the functionsf, g, h satisfy the Hypothesis 2.1 here below.

Definition 2.1 (Optimal Control Problem (OC)) (2.1) MinimizeJt(x(·), α) :=

Z T

t

1

2|α(s)|2+f(x(s), s)ds+g(x(T)) subject to (x(·), α)∈ A(x, t), where

(2.2)

A(x, t) :=n(x(·), α)∈AC([t, T];R2L2([t, T];R2) : (1.4) holds a.e. with x(t) =xo. A pair (x(·), α) in A(x, t) is said to be admissible. We say thatx is an optimal trajectory if there is a control α such that (x, α) is optimal for (OC). Also, we shall refer to the system (1.4) as to the dynamics of the optimal control problem (OC).

In what follows, the functions f, g and hsatisfy the following conditions.

Hypothesis 2.1 fC0([0, T], C2(R2)) and there exists a constantC such that kf, t)kC2(R2)+kgkC2(R2)+khkC2(R)C,t∈[0, T].

The set Z ={z∈R : h(z) = 0} is totally disconnected.

Remark 2.1 Our Hypothesis 2.1 is very strong and we clearly do not need it all over the paper. We much prefer to state a more restrictive result without the need to add further assumptions step by step.

Remark 2.2 Thanks to the boundedness of functions in our Hypothesis 2.1, it is equiva- lent to choose A(x, t) or

(2.3)

A˜(x, t) :=n(x(·), α) ∈AC([t, T];R2L1([t, T];R2) : x satisfies (1.4) a.e., x(t) =xo, in the control problem.

Remark 2.3 Notice that, given a control lawαL2([t, T];R2), the above Hypothesis 2.1 on h implies that, given the initial point x, there is a unique trajectory x(·) such that (x(·), α) ∈ A(x, t).

Remark 2.4 (Existence of optimal solutions) Hypothesis 2.1, together with [14, The- orem 23.11], ensure that the optimal control problem (OC) admits a solution (x, α).

(6)

2.1 The Hamilton-Jacobi equation and the value function of the optimal control problem

The aim of this section is to study the Hamilton-Jacobi equation (1.1)-(i) with m fixed, namely

(2.4)

(tu+12|DGu|2 =f(x, t) inR2×(0, T),

u(x, T) =g(x) onR2

where f(x, t) := F(x, m) and g(x) :=G(x, m(T)). Under Hypothesis 2.1, we shall prove several regularity properties of the solution (especially Lipschitz continuity and semicon- cavity) and mainly the uniqueness of optimal trajectories for the associated optimal control problem.

Definition 2.2 The value function for the costJt defined in (2.1) is (2.5) u(x, t) := inf{Jt(x(·), α) : (x(·), α)∈ A(x, t)}.

An optimal pair (x(·), α) for the control problem (OC) in Definition 2.1 is also said to be optimal for u(x, t).

In the next lemma we show that the solution u of (2.4) can be represented as the value function of the control problem (OC) defined in (2.1).

Lemma 2.1 Under Hypothesis 2.1, the value function u, defined in (2.5), is the unique bounded uniformly continuous viscosity solution to problem (2.4).

Proof. The Dynamic Programming Principle (stated in Proposition 5.1 in Appendix be- low) yields that the value function is a solution to problem (2.4). Moreover applying clas- sical results uniqueness (see, for example, [8, eq. (7.40) and Thm. 7.4.14]), we obtain the statement. Moreover, taking as admissible control the law α= 0, from the representation formula (2.5), using the boundedness of f and g, we have |u(x, t)| ≤CT. ✷ Lemma 2.2 (Lipschitz continuity) Under Hypothesis 2.1, there hold:

1. u(x, t) is Lipschitz continuous with respect to the spatial variable x, 2. u(x, t) is Lipschitz continuous with respect to the time variable t.

Proof. In this proof, CT will denote a constant which may change from line to line but it always depends only on the constants in the assumptions (especially the Lipschitz constants off and g) and on T.

1. Let t be fixed. We follow the proof of [9, Lemma 4.7]. Let αε be an ε-optimal control foru(x, t) i.e.,

(2.6) u(x1, x2, t) +εZ T

t

1

2|αε(s)|2+f(x(s), s)ds+g(x(T))

where x(·) obeys to the dynamics (1.4) with α = αε. From the boundedness of u (es- tablished in Lemma 2.1) and our assumptions, there exists a constant CT such that

(7)

kαεkL2(t,T)CT.

We consider the pathx(s) starting fromy= (y1, y2), with control αε(·). Hence x1(s) = y1+

Z s

t

αε1(τ) =y1x1+x1(s) x2(s) = y2+

Z s

t h(y1x1+x1(τ))αε2(τ)

= y2x2+x2(s) + Z s

t

h(y1x1+x1(τ))αε2(τ)−h(x1(τ))αε2(τ)dτ.

Using the Lipschitz continuity off and h and the boundedness of hwe get f(x(s), s) =

f(x1(s), x2(s), s) +L|y1x1|+ +L

y2x2+ Z s

t

h(y1x1+x1(τ))αε2(τ)−h(x1(τ))αε2(τ)

f(x1(s), x2(s), s) +L|y1x1|+L|y2x2|+L|y1x1| Z s

t |αε2(τ)|

f(x(s), s) +L|y1x1|+L|y2x2|+L|y1x1|T12 Z s

t

ε2(s))2ds 12

. By the same calculations forg and substituting inequality (2.6) in

u(y1, y2, t)Z T

t

1

2|αε(s)|2+f(x(s), s)ds+g(x(T)), we get

u(y1, y2, t)u(x1, x2, t) +CT(|y2x2|+|y1x1|).

Reversing the role ofx and y we get the result.

2. We follow the same arguments as those in the proof of [9, Lemma 4.7]; to this end, we recall thath is bounded and we observe that there holds

|x(s)x| ≤C(st)kαk

sinceα is bounded as proved in Corollary 2.1 in the next section. ✷ In the following lemma we establish the semiconcavity ofu(x, t) taking advantage of the regularity hypothesis 2.1. This property will be needed in the study of the relationship between the regularity of the value function and the uniqueness of the optimal trajectories.

It is worth to remark that it is possible to prove thatu(x, t) is also semiconcave with respect to theχ-lines associated to the Grushin dynamics, as introduced in [4, Example 2.4], but this does not seem to be useful to our results.

Lemma 2.3 (Semi-concavity) Under Hypothesis 2.1, the value function u, defined in (2.5), is semiconcave with respect to the variable x.

Proof. For anyx, y∈R2andλ∈[0,1], considerxλ :=λx+(1λ)y. Letαbe anε-optimal control foru(xλ, t); we set

xλ(s) = (xλ,1(s), xλ,2(s)) :=

xλ,1+

Z s

t

α1(τ)dτ, xλ,2+ Z s

t

h(xλ,1(τ))α2(τ)

.

(8)

Let x(s) and y(s) satisfy (1.4) with initial condition respectively x and y still with the same controlα,ε-optimal foru(xλ, t). We have to estimateλu(x, t)+(1λ)u(y, t) in terms ofu(xλ, t). To this end, arguing as in the proof of [9, Lemma 4.7], we have to estimate the termsλf(x(s), s) + (1−λ)f(y(s), s) andλg(x(T)) + (1−λ)g(y(T)).

We explicitly provide the calculations for the second componentx2(s) since the calculations forx1(s) are the same as in [9]. We have

x2(s) = x2+ Z s

t h(x1(τ))α2(τ)

= x2xλ,2+xλ,2(s) + Z s

t

(h(x1(τ))−h(xλ,1(τ)))α2(τ)dτ,

and analogously fory2(s). For the sake of brevity we provide the explicit calculations only forf and we omit the analogous ones forg; we writef(x1, x2) :=f(x1, x2, s). We have

λf(x(s)) + (1λ)f(y(s)) = λf

x1(s), xλ,2(s) +x2xλ,2+ Z s

t

(h(x1(τ))−h(xλ,1(τ)))α2(τ) +(1−λ)f

y1(s), xλ,2(s) +y2xλ,2+ Z s

t

(h(y1(τ))−h(xλ,1(τ)))α2(τ)

. In the Taylor expansion of f centered in xλ(s) the contribution of the first variable can be dealt with as in [9]. Assuming without any loss of generalityx1=y1, the contribution of the second variable gives

λf(x(s)) + (1λ)f(y(s)) =f(xλ(s)) +x2f(xλ(s))

λ(x2xλ,2) + (1−λ)(y2xλ,2) +λ

Z s

t (h(x1(τ))−h(xλ,1(τ)))α2(τ)) + (1−λ)

Z s

t (h(y1(τ))−h(xλ,1(τ)))α2(τ)dτ)

+R, whereR is the error term of the expansion, namely

(2.7) R=λ∂x22,x2f1) 2

x2xλ,2+ Z s

t (h(x1(τ))−h(xλ,1(τ)))α2(τ) 2

+ (1−λ)∂x22,x2f2) 2

y2xλ,2+ Z s

t

(h(y1(τ))−h(xλ,1(τ)))α2(τ) 2

, for suitableξ1, ξ2∈R2.

Since λ(x2xλ,2) + (1−λ)(y2xλ,2) = 0, we get

(2.8) λf(x(s)) + (1λ)f(y(s)) =f(xλ(s)) +x2f(xλ(s)) Z s

t I(τ2(τ)+R, withI(τ) :=−h(xλ,1(τ)) +λh(x1(τ)) + (1−λ)h(y1(τ)).Now, our aim is to estimateI(τ).

Sincexλ,1(τ) =λx1(τ)+(1−λ)y1(τ),x1(τ)−xλ,1(τ) = (1−λ)(x1y1) andy1(τ)−xλ,1(τ) = λ(y1x1), the Taylor expansion for h centered inxλ,1(τ) yields

I(τ) = 1

2(1−λ)λ(y1x1)2[(1−λ)h′′(ξ) +λh′′(ξ)],e

(9)

for suitableξ,ξe∈R. Our Hypothesis 2.1 entails

|I(τ)| ≤(1−λ)λC(y1x1)2. Replacing the inequality above in (2.8), we obtain

(2.9) λf(x2(s)) + (1−λ)f(y2(s))≤f(xλ,2(s)) +C2T(1−λ)λ(y1x1)2+R.

Let us now estimate the error termR in (2.7). We have

x2xλ,2+ Z s

t (h(x1(τ))−h(xλ,1(τ)))α2(τ) 2

≤2(x2xλ,2)2+ 2 Z s

t

(h(x1(τ))−h(xλ,1(τ)))α2(τ) 2

≤2(1−λ)2(x2y2)2+ 2C(1−λ)2(x1y1)2C(1λ)2|xy|2 and, analogously

y2xλ,2+ Z s

t

(h(y1(τ))−h(xλ,1(τ)))α2(τ) 2

2|xy|2 Then, replacing these two inequalities in (2.7), we infer

(2.10) RC(1λ)λ|xy|2.

Taking into account (2.10) and (2.9), we get the semiconcavity of u.2.2 Necessary conditions and regularity for the optimal trajectories The application of the Maximum Principle yields the following necessary conditions.

Proposition 2.1 (Necessary conditions for optimality) Let (x, α) be optimal for (OC). There exists an arc pAC([t, T];R2), hereafter called the costate, such that

1. The pair, p) satisfies the adjoint equations: for a. e. s∈[t, T] p1=−p2h(x12+fx1(x, s) (2.11)

p2=fx2(x, s), (2.12)

the transversality condition

(2.13) −p(T) =Dg(x(T))

together with the maximum condition

(2.14) max

α=(α12)∈R2p1(s)α1+p2(s)h(x1(s))α2−|α|2

2 =

=p1(s)α1(s) +p2(s)h(x1(s))α2(s)−|α(s)|2

2 a. e.s∈[t, T].

(10)

2. The optimal control α is given by (2.15)

(α1 =p1

α2 =p2h(x1) a.e on [t, T].

3. The pair (x, p) satisfies the system of differential equations: for a. e. s∈[t, T] x1=p1

(2.16)

x2=h2(x1)p2 (2.17)

p1=−p22h(x1)h(x1) +fx1(x, s) (2.18)

p2=fx2(x, s) (2.19)

with the mixed boundary conditions x(t) =x, p(T) =−Dg(x(T)).

Proof. 1. Hypothesis 2.1 ensures the validity of the assumptions of the Maximum Principle [14, Theorem 22.17]. Since the endpoint is free, [14, Corollary 22.3] implies that the deduced necessary conditions hold in normal form: the claim follows directly.

2. The maximum condition (2.14) implies that Dα p1(s)α1+p2(s)h(x1(s))α2−|α|2

2

!

α=α

= 0 a. e.s∈[t, T] from which we get (2.15).

3. Conditions (2.16) – (2.17) follow directly from the dynamics (1.4) replacingα1, α2 by means of (2.15). Condition (2.18) follows similarly from (2.11), whereas (2.19) coincides

with (2.12). ✷

Remark 2.5 In the case where one prescribes the endpoint x(T) in Definition 2.1, the proof of Proposition 2.1 shows that the claim still holds true without the endpoint condition for p(T) in Point 3.

Corollary 2.1 (Feedback control and regularity) Let (x, α) be optimal for u(x, t) and p be the related costate as in Proposition 2.1. Then:

1. The costate p= (p1, p2) is uniquely expressed in terms of x for every s∈[t, T] by

(2.20)

p1(s) =−gx1(x(T))− Z T

s

fx1(x, τ)−p22h(x1)h(x1)dτ, p2(s) =−gx2(x(T))−

Z T

s fx2(x, τ)dτ.

2. The optimal control α = (α1, α2) is a feedback control (i.e., a function of x), uniquely expressed in terms of x for a. e.s∈[t, T] by

(2.21)

α1(s) =−gx1(x(T)) + Z s

T

fx1(x, τ)−p22h(x1)h(x1)dτ, α2(s) =p2(s)h(x1(s)).

3. The optimal trajectory x and the optimal control α are of class C1. In particular the equalities (2.15) (2.21) do hold for every s∈[t, T].

(11)

4. Assume that, for some k ∈ N, hCk+1 and Dxf(x, s) is of class Ck. Then x, p and α are of classCk+1.

Proof. Point 1 is an immediate consequence of (2.18) – (2.19) together with the endpoint condition p(T) =−Dg(x(T)). Point 2 follows then directly from (2.15).

3. Sincex is continuous, the continuity ofα follows from (2.21). The dynamics (1.4) then imply that xC1. Relations (2.20) and (2.21) imply, respectively, that p and α are of classC1.

4. The relations (2.20) and theC1-regularity ofx andpimply that, actually,pC2. Therefore, (2.21) gives the C2-regularity of α and, finally, the dynamics (1.4) yield the C2-regularity of x. Further regularity of x, α and p follows by a standard bootstrap

inductive argument. ✷

2.3 Uniqueness of the trajectories after the initial time

Next Theorem 2.1 implies that the optimal trajectories foru(x, t), do not bifurcate at any time r > t whenever h(x1) 6= 0 (see Corollary 2.2), otherwise they may rest at x in an interval from the initial time tbut they do not bifurcate as soon as they leavex.

Theorem 2.1 (Uniqueness of the optimal trajectory after the rest time) Under Hypothesis 2.1, let x be an optimal trajectory for u(x, t).

1. Assume that h(x1(τ)) 6= 0 for some t < τ < T. For every τr < T there are no other optimal trajectories foru(x(r), r), other than x, restricted to [r, T].

2. Assume that h(x1) = 0. Let tx be the rest time for x defined by tx:= sup{r∈[t, T] : xx on [t, r]}.

For every r > tx there are no optimal trajectories for u(x(r), r), other than x restricted to [r, T].

Remark 2.6 We point out that the rest time may be positive only when h(x1) vanishes.

Notice also that tx =T if and only if x is constant on [t, T].

The next Lemma 2.4 relates the initial constancy of an optimal trajectory to a stationary condition and is a key argument of the proof of part (2) of Theorem 2.1.

Lemma 2.4 (A stationary condition) Assume thath(x1) = 0. Letx = (x1, x2) be an optimal trajectory for u(x, t), and r∈[t, T]. Then

(2.22) xx on [t, r]⇔h(x1)≡0 on [t, r].

Proof. If h(x1) = 0 on [t, r] then x1 belongs to set of the zeros of h, which is totally disconnected by Hypothesis 2.1. It follows thatx1x1 on [t, r]. Moreover, the dynamics (1.4) imply that x2x2 on [t, r], so that xx on [t, r]. The opposite implication is

trivial, sinceh(x1) = 0. ✷

(12)

Remark 2.7 The two conditions in(2.22) are both equivalent to the fact that the (unique) control α such that Jt(x, α) =u(x, t) is singular on [t, r] in the sense of [7, Def. 2.3], i.e., that there exists an absolutely continuous arc π: [t, r]→R2\ {(0,0)} satisfying (2.23) π1 = 0, π2h(x1) = 0, π1=−π22h(x1)h(x1), π2 = 0.

Indeed, if xx on [t, r] for some t < rT then any arc π := (0, c) with 06=c ∈R is such that (x, π) satisfies (2.23) on [t, r]. Conversely, if (x, π) fulfills (2.23) on [t, r] for somet < rT, then π2=cfor some constant c. Sinceπ1 = 0then c6= 0. It follows from the second equation in (2.23) that h(x1) = 0 on [t, r].

Notice that (2.23) are the conditions satisfied by the costatep in (2.15)and (2.18) (2.19) when α = (0,0) and f ≡0.

Proof of Theorem 2.1. 1. Let y be optimal for u(x(r), r). Point 1 of Proposition 5.1 in the Appendix ensures that the concatenation z of x with y at r is optimal for u(x, t).

Letp:= (p1, p2), q:= (q1, q2) be the costates associated tox := (x1, x2) and, respectively, toz := (z1, z2). Both (x, p) and (z, q) satisfy (2.16) – (2.19) on [t, T]. Now, Corollary 2.1 shows thatx and z are of classC1. Sincex=z on [t, τ], the fact thatτ > t, together with (2.16) imply

p1(τ) = (x1)(τ) = lim

s→τ(x1)(s) = lim

s→τ(z1)(s) = (z1)(τ) =q1(τ), whereas (2.17), and the fact thath(x1(τ))6= 0 analogously yield

p2(τ) = (x2)(τ)

h2(x1(τ)) = (z2)(τ)

h2(z1(τ)) =q2(τ).

Therefore, both (x, p) and (z, q) are absolutely continuous solutions to the same Cauchy problem on [t, T], with Cauchy data at τ, for the first order differential system (2.16)- (2.19). The regularity assumptions on f and h and Caratheodory’s Theorem guarantee the uniqueness of the solution. Thusx =z on [t, T], from which we obtained the desired equalityx=y on [r, T].

2. We assume thattx< T, otherwise the claim is trivial. We deduce from Lemma 2.4 that there is τ ∈ [tx, r] satisfying h(x1(τ)) 6= 0. If y is optimal for u(x(r), r), then Point 1 of Proposition 5.1 shows that the concatenation z of x withy at r is optimal foru(x, t). Moreover, Point 2 of Proposition 5.1 imply that both x and z, restricted to [τ, T], are optimal foru(x(τ), τ). Point 1 of Theorem 2.1 implies that x =y on [τ, T],

proving the desired result. ✷

Corollary 2.2 Let x be an optimal trajectory foru(x, t). Ifh(x1)6= 0, for every 0< r <

T there are no other optimal trajectories for u(x(r), r), other than x, restricted to [r, T].

3 The continuity equation

In this section we want to study equation (1.1)-(ii). Sinceh is independent of x2, taking account of (1.2), this partial differential equation can be rewritten as

(3.1) tmx1(m∂x1u)h2(x1)∂x2(m∂x2u) =∂tm−divG(mDGu) = 0.

(13)

Hence our aim is to study the well posedness of the problem (3.2)

( tm−divG(m DGu) = 0, inR2×(0, T), m(x,0) =m0(x), on R2,

whereu is a solution to problem (3.3)

(tu+ 12|DGu|2 =F(x, m) inR2×(0, T), u(x, T) =G(x, m(T)), onR2,

where the functionm is fixed and fulfills (3.4) mC1/2([0, T],P1),

Z

R2|x|2dm(t)(x)K, t∈[0, T].

Note that this problem is equivalent to (2.4) with a fixedm.

Observe that, by Lemma 2.2-(1), in (3.2) the drift v = DGu is only bounded; this lack of regularity prevents to apply the standard results (uniqueness, existence and repre- sentation formula ofmas the push-forward ofm0 through the characteristic flow; e.g., see [2, Proposition 8.1.8]) for drifts which are Lipschitz continuous in x. We shall overcome this difficulty applying Ambrosio superposition principle [2, Theorem 8.2.1] and proving several results on the uniqueness of the optimal trajectory for the control problem stated in Section 2. The Ambrosio superposition principle yields a representation formula ofmas the push-forward of some measure onC0([0, T],R2) through the evaluation mapet. In the following theorem, we shall also recover uniqueness, existence and some regularity result for the solution to (3.2).

Theorem 3.1 Under assumptions (H1) – (H4), for anym as in (3.4), Problem (3.2) has a unique bounded solutionm in the sense of Definition 1.1. Moreover m(t,·) is absolutely continuous with supt∈[0,T]km(t,·)kC and it is a Lipschitz continuous map from [0, T] to P1 with a Lipschitz constant bounded by kDukkh2k. Moreover, the function m satisfies:

(3.5)

Z

R2φ dm(t) = Z

R2φ(γx(t))m0(x)dxφC00(R2),∀t∈[0, T] where, for a.e. x∈R2, γx is the solution to (1.5).

The proof of Theorem 3.1 is given in the next two subsections which are devoted respectively to the existence result (see Proposition 3.1), to the uniqueness result and to the representation formula (see Proposition 3.2) and to the Lipschitz regularity (see Corollary 3.1).

3.1 Existence of the solution

As in [10, Appendix] (see also [9, Section 4.4]), we now want to establish the existence of a solution to the continuity equation via a vanishing viscosity method, applied on the whole MFG system.

Proposition 3.1 Under assumptions (H1) – (H4), problem (3.2) has a bounded solution m in the sense of Definition 1.1.

(14)

We consider the solution (uσ, mσ) to the following problem (3.6)

(i) −tuσ∆u+12|DGu|2=F(x, m) inR2×(0, T) (ii) tmσ∆m−divG(mDGu) = 0 inR2×(0, T) (iii) m(x,0) =m0(x), u(x, T) =G(x, m(T)) onR2.

Let us recall that equation (3.6)-(ii) has a standard interpretation in terms of a suitable stochastic process (see relation (3.9) below). Our aim is to find a solution to problem (3.2) letting σ → 0+. To this end some estimates are needed; as a first step, we establish the well-posedness of system (3.6).

Lemma 3.1 Under assumptions(H1) – (H4), for anymas in (3.4), there exists a unique bounded classical solution (uσ, mσ) to problem (3.6). Moreover, mσ >0.

Proof. The proof uses standard regularity results for quasilinear parabolic equations; from Lemma 3.2 here below, the solution uσ of (3.6)-(i) is bounded in R2 ×[0, T]. Hence we can apply [22, Theorem 8.1, p.495], obtaining the existence and uniqueness of a classical solution uσ in all R2×[0, T]. Now mσ is the classical solution of the linear equation

tmσ∆m+b·Dm+c0m= 0, m(0) =m0

with b and c0 Hölder continuous coefficients. Hence still applying classical results (see [22, Theorem 5.1, p.320]) we get the existence and uniqueness of a classical solution mσ of (3.6)-(ii). From assumptions on m0 and the maximum principle (see for example [22,

Theorem 2.1, p.13]) we get thatmσ >0. ✷

Let us now prove that the functions uσ are Lipschitz continuous and semiconcave uniformly in σ.

Lemma 3.2 Under the same assumptions of Lemma 3.1, there exists a constant C >0, independent of σ such that

kuσkC, kDuσkC and D2uσCσ >0.

Proof. The L-estimate easily follows from the Comparison Principle and assumption (H2) because the functionsw±:=C±C(Tt) are respectively a super- and a subsolution for (3.6)-(i) if C is sufficiently large.

We refer to [9] for the proof of the uniform Lipschitz continuity of the functions uσ. The proof is similar to the deterministic one proved in Lemma 2.2 and it uses the representation formula by means a stochastic optimal control problem:

uσ(x, t) = min Z T

t

1

2|α(τ)|2+f(Y(τ), τ)

+g(Y(T)) where, in [t, T], Y(·) is governed by a stochastic differential equation (3.7)

( dY1 =α1(t)dt+√

2σdB1,t dY2 =h(Y1(t))α2(t)dt+√

2σdB2,t ,

where Y(t) = x and Bt is a standard 2-dimensional Brownian motion. (For an analytic proof see also [26, Chapter XI])

(15)

Let us now prove the part of the statement concerning the semiconcavity. We shall adapt the methods of [10, Lemma 5.2]. We fix a direction v = (α1, α2) with |v|= 1 and compute the derivative of equation (3.6)-(i) twice with respect tov obtaining

tvvuσ∆∂vvuvv(F(x, m(x, t)) =−vvh12 (∂1u)2+h(x1)2(∂2u)2i

=−(DGvu)2DGu·DGvvu12vv(h2)(∂2u)2−4hhα12u∂2vu

≤ −(DGvu)2DGu·DGvvu+C(1 +|DGvu|)

(the last inequality is due to our assumptions and to the first part of the statement). Since

−(DGvu)2+C(1 +|DGvu|) is bounded above by a constant, we deduce

tvvuσ∆∂vvu+DGu·DGvvuC;

on the other hand, we have kvvu(T,·)kC by assumption (H2) and we can conclude by comparison that vvuC for a constant C independent ofσ.

Let us now prove some useful properties of the functionsmσ.

Lemma 3.3 Under the same assumptions of Lemma 3.1, there exists a constant K >0, independent of σ, such that:

1. kmσkK,

2. d1(mσ(t1)−mσ(t2))≤K(t2t1)1/2t1, t2∈(0, T), 3.

Z

R2|x|2dmσ(t)(x)≤K Z

R2|x|2dm0(x) + 1

t∈(0, T).

Proof. 1. In order to prove this L estimate, we shall argue as in [10, Appendix]; for simplicity, we drop theσ’s. We note that

divG(mDGu) =DGm·DGu+m(∂11u+h222u)DGm·DGu+Cm

because of the semiconcavity of u established in Lemma 3.2 yields iiuC for i = 1,2 (see [8, Proposition1.1.3-(e)]) and m≥0. Therefore, by assumption (H2) the functionm satisfies

tmσ∆mDGm·DGu+Cm, m(x,0)≤C;

usingw = CeCt as supersolution (recall that C is independent of σ), we infer: kmkw=CeCT.

To prove Points 2 and 3 as in the proof of [9, Lemma 3.4 and 3.5], it is expedient to introduce the stochastic differential equation

(3.8) dXt=b(Xt, t)dt+√

2σdBt, X0=Z0 whereb= (∂u∂xσ

1, h2∂u∂xσ

2),Btis a standard 2-dimensional Brownian motion, andL(Z0) =m0. By standard arguments, (see [21] and [20, Chapter 5])

(3.9) m(t) :=L(Xt)

is a weak solution to (3.6)-(ii).

The rest of the proof of Points 2 and 3 follows the same arguments of [9, Lemma 3.4]

and, respectively, of [9, Lemma 3.5]; therefore, we shall omit it and we refer to [9] for the

detailed proof. ✷

Références

Documents relatifs

First introduced by Faddeev and Kashaev [7, 9], the quantum dilogarithm G b (x) and its variants S b (x) and g b (x) play a crucial role in the study of positive representations

Using the weak solution of Inverse mean curvature flow, we prove the sharp Minkowskitype inequality for outward minimizing hypersurfaces in Schwarzschild space... where Ric is the

10, the resulting hydrodynamics (later on referred to as hydrodynamic model for self-alignment with precession) consists of a conservation law for the density and a

For the blow-up result in the slow diffusion case, Ishida and Yokota [13] proved that every radially symmetric energy solution with large negative initial energy blows up in

• Our compactness results for the approximate solutions are valid for d = 2, 3, 4 which lead to the existence of global weak solutions for the general superlinear Hookean model in d

In this paper, we propose a nonlocal model for linear steady Stokes equation with no-slip boundary condition.. The main idea is to use volume constraint to enforce the no-slip

This paper is organized as follows: Section 2 describes time and space discretiza- tions using gauge formulation, Section 3 provides error analysis and estimates for the

The (first order) Godunov scheme is based on exact solutions to (1.1), and is a good prototypical example to lead us to defining a class of (high order) schemes consistent with