• Aucun résultat trouvé

Bifurcations in neural masses

N/A
N/A
Protected

Academic year: 2021

Partager "Bifurcations in neural masses"

Copied!
6
0
0

Texte intégral

(1)

HAL Id: hal-00331618

https://hal.archives-ouvertes.fr/hal-00331618

Submitted on 17 Oct 2008

HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.

Bifurcations in neural masses

Romain Veltz, Olivier Faugeras

To cite this version:

Romain Veltz, Olivier Faugeras. Bifurcations in neural masses. Deuxième conférence française de

Neurosciences Computationnelles, ”Neurocomp08”, Oct 2008, Marseille, France. �hal-00331618�

(2)

Bifurcations in neural masses

Romain VELTZ

+

, Oliver FAUGERAS

+

romain.veltz@sophia.inria.fr

Introduction

Neural continuum networks are an important aspect of the modeling of macroscopic parts of the cortex.

They have been first studied by Amari[6]. These net- works have then been the basis to model the visual cor- tex by Bresslov[4]. From a computational viewpoint, the neural masses could be used to perform image processing like segmentation, contour detection... Thus, there is a need to develop tools (theoretical and numerical) allow- ing the study of the dynamical and stationary properties of the neural masses equations.

In this paper, we look at the dependency of neural masses stationary solutions with respect to the stiffness of the nonlinearity. This choice is motivated by two rea- sons. First, it shows the differences with the case of an infinite stiffness (which corresponds formally to a heav- iside nonlinear function) as used for example in Ermen- trout[2]. Second, the study is generic and any other parameter could have been used with similar tools.

The study is done by using bifurcation theory in infi- nite dimensions. We provide a useful approximation of the connectivity matrix and give numerical examples of bifurcated branches which had not been yet fully com- puted in the literature. The analysis relies on the study of a simple model thought generic in the sense it has the properties that any neural mass system should possess generically.

1 General framework

We consider the following neural mass equation defined over a bounded piece of cortex Ω ⊂ R

d

, d = 1, 2, 3

1

:

V

t

(r, t) = −L.V(r, t) + [W · S(σV − θ)] (r, t) + I

ext

(r) V(., t) ∈ L

2

(Ω, R

p

) = H, a real vector space, where p is the number of populations of neurons and we endow H with the inner product

hV

1

, V

2

i

H

= Z

dr hV

1

(r), V

2

(r)i

Rp

S : R

p

→ R

p

is defined by S(x) = [S(x

1

), · · · , S(x

p

)]

T

. σ = diag(σ

1

, · · · , σ

p

) determines the slope of each of the p sigmoids

2

at the origin.

1We can formally have Ω⊂Rd×(0,2π)r, see Bresslov

2iethe funtionx→(1 +e−x)−1

θ ∈ R

p

determines the threshold of each of the p sig- moids.

W(t) is a linear operator from H to itself:

[W · V](r, t) = Z

W(¯ r, r)V(¯ r) d¯ r,

where W(r, r

0

, t) is a p × p matrix.

L = diag(

τ1

1

, · · · ,

τ1

p

) where the positive numbers τ

i

, i = 1, · · · , p determines the dynamics of each neural pop- ulation. Par changing the variables, we can assume θ = 0 and = Id.

In this article we focus on the influence of the slopes σ. We make the assumption that all slopes are equal to λ, σ = λId

p

. We can therefore rename V, W and I

ext

to restrict our study to:

V

t

(r, t) = −V(r, t) + [W · S(λV)] (r, t) + I

ext

(r) (1) which we rewrite :

0 = −V

f

+ W · S(λV

f

) + I

ext

def

= −F(V

f

, λ) (2) For λ given, let B

λ

= {V/F (V, λ) = 0}

In [1] we proved :

Proposition 1.1. ∀λ ∈ R , B

λ

6= ∅ In fact we can prove much better :

Proposition 1.2. Given a < b there exits a continuous curve s → (V

fs

, λ(s)) such that λ([0, 1]) = [a, b] and

∀s V

sf

∈ B

λ(s)

2 Bifurcation theory

Once we have a solution V

fλ

of F(V, λ) = 0 over an interval J = [a, b], we can study the steady states bifur- cation i.e. find the solutions, if any, that are different from V

fλ

. We start with a definition :

Definition 2.1. We say that the equation F (V, λ) = 0 bifurcates from (V

fλ

0

, λ

0

) a solution (U

λ

, λ) if there exists a sequence of solutions (U

n

, λ

n

) which satisfies

n→∞

lim λ

n

= λ

0

and lim

n→∞

kU

n

− V

λ0

k = 0

(3)

From F (V, λ) = V − W · S(λV) − I

ext

Writing V = V

fλ

+ U, we now consider (up to the identification ˜ F = F) :

F(U, λ) = L

λ

· U + ˜ G(U, λ), where

L

λ

= Id − W · (λDS(V

fλ

)) G(U, λ) = ˜ P

n≥2

D

(n)

G(V

fλ

, λ) · U

(n)

≡ P

n≥2

G

n

(U, λ) Then (Implicit functions theorem) :

Proposition 2.2. A necessary condition for λ

0

to be a bifurcation point is L

λ0

non invertible.

These points λ

0

are possible bifurcation points (ac- cording to the previous definition). The bifurcation the- ory aims at telling what happens around such point λ

0

, if there are other branches of solutions and their num- ber. This is interesting because it gives the stationary states of the neural masses equations.

3 PG-kernel approximation of the connectivity matrix

We focus on a class of connectivity functions which is very rich and can represent economically all kinds of bi- ologically plausible cortical connectivity kernels. More- over, such kernels allows to write the neural masses equa- tions as finite dimensional ODEs. We therefore assume that W is a PG-kernel.

Definition 3.1 (Pincherle-Goursat Kernels). The con- nectivity W(r, r

0

) is a PG-kernel if

W(r, r

0

) =

N−1

X

k,l=0

a

kl

X

k

(r) ⊗ Y

l

(r

0

)

where (X

k

, k = 1..N ) are linearly independent functions of H.

We make the following assumptions : Ω = [−1, 1]

2

, p = 2 and also :

S(0) = 0, I

ext

= 0

Thus, the persistent state V

fλ

= 0 is a trivial solution of the equations 2.

We are interested in excitatory-inhibitory neural fields, thus the sign of the connectivity functions has to be like this :

sign(W) =

+ −

± ±

3.1 Description of the connectivity ma- trix component

It is often seen in the literature that W

ij

is gaussian and the cortex is chosen infinite. For example one find in [2,5]

W

11

(r, r

0

) = exp−kr−¯ rk

2

/2 = e

−krk2/2

e

−kr0k2/2

e

hr,r0i

≈ e

−krk2/2

e

−kr0k2/2

1 + hr, r

0

i + 1

2 hr, r

0

i

2

+ ...

We notice two things :

• 1 + hr, r

0

i +

12

hr, r

0

i

2

+ ... is a polynomial

• e

−krk2/2

is bell-shaped, we will approximate it with

1 − krk

2

a

with a > 0. This choice is also moti- vated by the fact that e

−krk2/2

tends to zero at the edges of the infinite cortex, so we choose to keep it ie

1 − krk

2

a

= 0 for r ∈ ∂[−1, 1]

2

.

Finally we will have the following simplified (N = 2) connectivity functions :

W

ij

(r, r

0

) = C

ij

+

1−krk

2

a

1−kr

0

k

2

a

P

ij

(r, r

0

) (3) where P is a polynomial in r and r

0

and C is a constant matrix.

From now, we will consider such an orthonormal family (e

n

)

n≥0

of L

2

(Ω, R ) with e

0

being the constant function 1/2, and ∀(x, y) ∈ Ω e

n

(x, y) =

1 − x

2

− y

2

a

P

n

(x, y).

We deduce an orthonormal basis (B

n

)

n≥0

of L

2

(Ω, R

2

) which will be used in the next section :

B

2n

= e

n

0

, B

2n+1

= 0

e

n

3.2 Toy model

We now choose the following connection matrix:

W(r, r

0

) =

α

1

+ α

2

e

n

(r)e

n

(r

0

) −ρ

1

+ ρ

2

e

q

(r)e

r

(r

0

) ρ

1

+ ρ

2

e

q

(r)e

r

(r

0

) α

1

+ α

2

e

m

(r)e

m

(r

0

) ,

which can expressed as a PG-kernel. Adjusting the con- stants α

i

, ρ

i

allows to satisfy the condition on sign(W).

This connectivity matrix can be seen as the first order approximation of a given connectivity matrix W

0

.

It is easy to check that σ(W) = {α

1

± Iρ

1

, ±ρ

2

, α

2

}

with α

2

6= ρ

2

.

(4)

We write V =

6

P

i=1

v

i

X

i

and obtain the following form for the neural masses equation by projecting on the fam- ily (X

i

)

i

associated to e

p

, e

q

... :

˙

v

k

= v

k

+

*

W.S X

6

i=1

v

i

X

i

, X

k

+

H

/kX

k

k

2H

(4)

3.3 Bifurcation of the persistent states

We have prop.2.2 σ(L

λ

) = −1 + λs

1

σ

n

where σ

n

∈ σ(W). Thus the possible bifurcation points are

λ

n

= 1 s

1

σ

n

3.3.1 Case of the simple eigenvalues

Let’s examine the case of the simple eigenvalues ±ρ

2

as- sociated with the eigenvectors f

±

= (±B

2r+1

+B

2q

)/ √

2.

We find for α

±

G

2

(f

±

, λ

±1

), f

±

H

6= 0

that λ

+1

is a transcritical bifurcation point and there is a unique bifurcated branch on each side of λ

±1

given (for example near λ

+1

) by

U

+λ

≈ −2

−1+λs1ρ2

+1)2s22ρ2

[

Re3r+e3q

]

B

2r+1

+ B

2q

≈ ρ

2

v

5f

(λ) ≈ ρ

2

v

6f

(λ)

Thus, for λ close to λ

1

, the bifurcated solution looks like the eigenvector B

2r+1

+ B

2q

.

3.3.2 Case of the double eigenvalue

σ

2

= α

2

is a double eigenvalue of W with associated eigenvectors f

1

= B

2n

, f

2

= B

2m+1

.

The last possible steady state bifurcation parameter is

λ

2

= 1 s

1

α

2

Following [3], we find that (2) reduces (to the first order) to find (x, y) ∈ R

2

such that

(−1 + λs

1

α

2

)x + λ

2

s

2

α

2

x

2

R

e

3n

= 0 (−1 + λs

1

α

2

)y + λ

2

s

2

α

2

y

2

R

e

3m

= 0 which leads to

( x

f

=

λ21−λss 1α2

2α2R

e3n

y

f

=

λ21−λss 1α2

2α2R

e3m

Thus, from (0, λ

2

) bifurcates one branch of solution on each side of λ

2

.

3.3.3 Global bifurcation analysis of steady states

The bifurcation analysis gives us insight on the local vicinity of the bifurcation point, but what can we tell if we want information far away from a given bifurcation point. In that case, one has to use global bifurcation (Cf [3]). We first study the case of the bifurcation point (0, λ

+1

).

Theorem 3.2 (Global Bifurcation). Under some veri- fied assumptions, any bifurcated branch can be either

• unbounded in F × R

• contains an odd number of points (0, λ

i

) 6= (0, λ

0

) such that λ

−1i

are eigenvalues of L

λ

with odd alge- braic multiplicities.

Using Cauchy Schwarz inequality one find that :

V

fλ

F

≤ Cλ, with C ∈ R (recall that p = 2) Case of the bifurcation point (0, λ

+1

) :

Using these results, one find the following diagram for the persistent states on fig.1.

Figure 1: Predicted shape of the bifurcated branches.

The cone represents the domain where the steady states cannot be.

3.3.4 Case of the complex eigenvalue Recall that −1 + λs

1

α

1

± Iρ

1

are complex eigenvalues of L

λ

. The singular value

λ

H

= 1

s

1

α

1

(5)

and the eigenvalue is

β(λ

H

) = I ρ

1

α

1

Recall that the neural masses equation is

dVdt

=

−F(V, λ). To have an Hopf bifurcation, we need to check if Reβ

0

H

) 6= 0. Here, we find

Re dβ

dλ (λ

H

) = 2s

1

α

1

> 0

In order to know if we are in the supercritical or the subcritical case, we compute the Floquet exponent :

¨

µ

2

(0) = 2 α

12

+ ρ

12

2

s

22

− s

1

s

3

α

12

s

14

, s

1

= S

0

(0), s

2

= S

(2)

(0)/2 The Floquet exponent tells if the bifurcated periodic

solution is attractive (µ

2

> 0) or repulsive (µ

2

< 0).

⇒ We have an Hopf bifurcation and the appear- ance of a periodic solution to equation (1), it is called Breathers in [5].

3.4 Numerical example

In this section, we do a simulation of the evolution equa- tion (??). A simple way to ensure the assumptions on the non linearity is to choose :

S (z) = Sig(z − 2) − Sig(−2)

where Sig is the sigmoid function define on page 1. Note that any sigmoidal function would work.

We use the Gegenbauer polynomials G

q

(Cf. eq.3) as a basis of L

2

((−1, 1), R ). They are associated to the dot product :

hf, gi = Z

1

−1

(1−x

2

)

a−1/2

f (x)g(x)dx, a > 0 or −1

2 < a < 0 For a the polynomials, we use the functions P

n

(x, y) = G

n1

(x)G

n2

(y). For the numerical applications, we take the following parameters :

n = [4, 2], m = [6, 6], q = [2, 2], r = [4, 4], a = 4 α

1

= 6, α

2

= 3, ρ

1

= 6, ρ

2

= 4

From this family added with the constant e

0

= 1/2, we build an orthonormal family (e

i

)

i=0..4

.

Figure 2: Representation of e

2

used for the orthonormal family.

We use the ODE representation of the neural masses equation to compute the evolution of a given initial state V(0). This reduces to a 6-dimensional ODE system, easily solved using MATLAB or SCILAB.

For example, the following picture shows the peri- odic orbit arising at the Hopf bifurcation point, the plot shows the coordinates used in the definition of eq.4.

Figure 3: The periodic orbit

As we know the bifurcation type around λ

+1

, we can use the pseudo - arc length continuation method in order to follow the steady states numerically when the slope varies. We find for the first bifurcation point, the predicted turning point by the global analysis.

This is shown on fig.4.

(6)

Figure 4: Representation of the different branch of solu- tions (λ, ~ v) near the bifurcation point λ

+1

.The red line represent the bifurcation point λ

+1

.

Conclusion

We provided theoretical tool to study the dynamics of the neural masses. Once we have a global understanding of the dynamics, we can perform numerical simulation.

To this hand, a convenient method of approximation have been presented allowing to follow the bifurcated branches even if there are dynamically unstable. We think that this will prove to be useful when considering computational applications of the neural masses.

There still lack a study of the asymptotic behavior when the nonlinearity stiffness goes to infinity.

References

[1] O. Faugeras, R. Veltz, F. Grimbert, Persitent neural states: stationary localized activity patterns in nonlin- ear continuous n-population, q-dimensional neural net- works, Neural Computation 2008 (accepted for publica- tion).

[2] B. Ermentrout, Neural networks as spatiotempo- ral pattern-forming systems, Reports on Progress in Physics, 61, 1998, 353-430.

[3] Ma, Wang, Bifurcation theory and applications, World Scientific, 2005.

[4] P. Bressloff, J. Cowan, M. Golubitsky, P. Thomas, M. Wiener, Geometric visual hallucinations, euclidean symmetry and the functional architecture of striate cor- tex, Phil. Trans. Roy. Soc. (Lond.) B (2001) 356, 1-32.

[5] S.E.Folias, P.C.Bressloff, Breathers in Two- Dimensional Neural Media, Phys. Rev. Lett. 95, 208107 (2005)

[6] Amari, S.-I. Dynamics of pattern formation in

lateral-inhibition type neural fields, Biological Cybernet- ics, 27:7787, 1977.

[7] Gutkin, B., Ermentrout, G., and OSullivan, J.

(2000). Layer 3 patchy recurrent excitatory connections may determine the spatial organization of sustained ac- tivity in the primate prefrontal cortex, Neurocomputing 3233 (2000) 391-400.

[8] Krasnoselskii, M., Vainikko, G., Zabreiko, P., and Stetsenko, V. (1972). Approximate solutions of oper- ator equations. Wolteers-Noordhoff. Translated by D.

Louvish.

Références

Documents relatifs

The study of the di eren es between experimental binding energies and those al ulated with ma ros opi mass formulas, plotted on the plane N-Z over the entire nu lear lands

deduce the atomic masses, the nuclear reaction and separation energies, the adjusted values for the input data, the influences of data on the primary masses described in next

The history of the evaluation of the nuclear masses (actually atomic masses) is only slightly younger than that of the mass measurements themselves.. In their modern form,

In the following sections we present the mathematical model that describes a windshear situation while the aircraft is landing (Section 2), we state the necesarry optimality

L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des

Indeed, the masses of all elementary fermions follow the formula 2 n .p 2 /2 where n is an integer [3, 4] calculated for the electron, electron neutrino and quark up and p is

Abstract In this paper, we consider neural field equations with space-dependent de- lays. Neural fields are continuous assemblies of mesoscopic models arising when modeling

We then use an extension of the contraction analysis for nonlinear systems [3] and of the analysis of concurrent synchronization in dynamic system networks [4] to obtain