HAL Id: jpa-00247083
https://hal.archives-ouvertes.fr/jpa-00247083
Submitted on 1 Jan 1995
HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés.
Dynamics of a Neural Network Composed by two Hopfield Subnetworks Interconnected Unidirectionally
L. Viana, C. Martínez
To cite this version:
L. Viana, C. Martínez. Dynamics of a Neural Network Composed by two Hopfield Subnetworks Interconnected Unidirectionally. Journal de Physique I, EDP Sciences, 1995, 5 (5), pp.573-580.
�10.1051/jp1:1995147�. �jpa-00247083�
Classification Physics Abstracts
05.50q 05.90m 75.10Hk
Dynamics of
afleurai Network Composed by
two Hopfield
Subnetworks Interconnected Unidirectionally
L. Viana
(~)
and C. Martinez (~)(~ Laboratorio de Ensenada, Instituto de Fisica, UNAM, A. Postal 2681, 22800 Ensenada, B.C., México
(~) Centro de Investigaciôn Cientifica y de Educaciôn Superior de Ensenada, Programa de
Posgrado en Fisica de Materiales, A. Postal 2681, 22800 Ensenada, B-C-, México
(Received
9 November 1994, accepted 30 January1995)
Abstract. By means of numerical simulations, trie dynamical behaviour of a Neural Network composed of two Hopfield Subnetworks interconnected unidirectionally and updated synchron- ically
(at
zero temperature T
= 0) is studied. These connections are such that each of the N
neurons on the first subnet sends information to exclusively one of the N neurons on the second.
A set of p patterns, composed by a subpattern for each of the subnets, is stored according to a Hebb-like rule. The recoverability of one particular subpattern by the second subnet is studied
as a function of N and the load parameter cY =
p/N
in the case when the first subnetwork has already recovered its corresponding subpattern.l. Introduction
During
the last few years, there has been agrowing
interest in thestudy
ofIsing-like
models of Neural Networks(NN),
that is, thestudy
of the collective behaviour of alarge
number N ofsimple interacting
twc-state elements called neurons. These systems have become theprototype model for the storage and retrieval of information in a content-addressable way, as stored information acts as attractors of the
dynamics
of the system. One of the most studied models is theHopfield
model[Ii,
which consists of N elements connected to each other withsymmetrical
interactions between them asgiven by
Hebb'slearning
rule [2] for p patterns. It has been found that in thethermodynarnical
limit(N
-cc),
for a=
(p/N)
- 0, ail p patternscon be stored
exactly
[3]; on the otherhand,
for o m 0, the information is stored with a fewmistakes,
the number of which increases with a, until for o= oc m 0.138, there is a sudden
catastrophe
and information can nolonger
be retrieved [4].In this paper we
study
thedynamical
behaviour at zero temperature(T
=
0)
of a finite NNcomposed by
twosubnetworks,
each withintemalsymmetric
connections between everypair
of neurons
given by
the Hebb fuie. These subnetworks areunidirectionally
interconnected toeach other,
by
means of strong Hebb-like connections in such a way that each of the N neurons in the first net(subnetwork
1) sends information to one, andonly
one, of the N neurons in the@
Les Editions de Physique 1995574 JOURNAL DE PHYSIQUE I N°5
Subnetwnrkl Subnetwark2
Fig. 1. We considered two finite coupled subnetworks interconnected unidirectionally, in such a
way that each of the N elements
in subnetwork 1 sends information to one, and only one, of the N elements in subnetwork 2.
second one
(subnetwork 2) (see Fig. 1).
In this system, we store p random unbiased patternsor states
(memories) (($)
(1 = 1,..,
2N),
eachcomposed by
twosubpattems
or submemories and stored in itsrespective
subnet (1 = 1,..., N and= N +1,..., 2N,
respectively).
Anappropriate
value for the connections between the two subnetworks isexpected
to allow for the recovery of acomplete
pattern,provided
the initial state of the first subnet is "closeenough"
to thecorresponding subpattem.
This is the issue we address in this paper:by dring
computersimulations v~e will
study
therecoverability
of acomplete
pattern when subnetwork 1 hasalready
recovered itscorresponding subpattem,
as a function of the number of stored patterns.Studying
this kind of models is of interest as we know thatcomplex
behaviour is constitutedby
a serres ofsimpler operations.
Forexample,
inqrder
toread,
it isimportant
torecognize
letters,
but we need to know that astring
of letters represents a word. In order toexemplify
the
problem
to bestudied,
we could make ananalogy consisting
instoring
2-letter words(patterns),
formedby
letters(subpatterns).
In this way, letters would be stored in the subnets while words would be stored in the whole network. Forexample,
we could store the word"Si"
(pattern numberl) consisting
of letters "s"(subpattern()
and "1"(subpattern[),
where upper index refers to pattern number and the lower one to subnetwork number. In a similar way, we could aise store other 2-letterwords,
such as "no"(pattern
number2).
If we intendedto recover the word "si"
by putting
the initial state of the first subnetwork within the basisof attraction of the letter
"s",
we could retrieve the word"Si",
butperhaps,
we could aise retrieve the word "so" or the letter "s"together
with any nominated(another letter)
or netilominated
state(a spurious state)
[Si. For the sake ofclanty
and withoutlosing
anygenerality, throughout
this paper we willalways
refer to theretrievability
of patternlby
subnetwork 2(although
tue resultsapply
to any of tuepatterns),
thereforeeliminating
the need of a scriptto indicate pattern number in the variables notation.
In this
problem,
subnetwork 1 is atypical Hopfield
NeuralNetwork,
which evolves inde-pendently
of anydynamical
processtaking place
m subnetwork 2. The sizes of the basins of attraction for such a NN have been calculatedby using severa1techniques,
SO we will net re-peat such calculations.
By dring
computersimulations,
Forrest [6] evaluated the mean fractionf(q~)
of states which are recalled with less thanN/16
errons as a function of q~ and p= a
IN,
where q~ is the
overlap
between tue initial state of tue network and tue related nominated pattern. This criterion isequivalent
todemanding
a finaloverlap
qf >7/8,
or a fraction ofmistakes lower than 0.0625, to consider tue information as
correctly
recalled. Another group of methods used to calculate the size of the basins of attraction for theHopfield NN,
in thethermodynarnicallimit,
consists initerating numencally
the fluxequations
for theoverlaps
be- tween the state of tue system and each of tue p stored patterns until a fixedpoint
is reached [7];following
thisapproach,
Viana and Coolen [8]calculated,
for tue Hebbmodel,
tue fraction of ail initial states which are within any of tue(p
= 1,..,18)
basins of attraction related to tuenominated patterns.
By dividing
such fractionby
p, we con obtain tue mean fraction of ail initial states which lie within tue basm of attraction of one(any)
of tue patterns.Knowing
theseprevious results,
we can concentrate herein onstudying
tue behaviour of subnetwork 2provided
subnetwork 1 basalready
recovered tue information related to it. We would like to mention that there is another groupworking
oncoupled
networks from a dilferentperspective: they
aremodelling
semantic memoryby considenng
tue case of biased patterns stored on severalcoupled
subnetworks with diluted connections between them [9].2. Trie Model
We consider a Neural Network
composed by
twosubnetworks,
each constitutedby
Ninteracting
twc-state neurons with
symmetric
interactions. Bath nets are connectedunidirectionally
insuch a way that i~~ neuron in subnetwork 1 sends information
exclusively
to tue(1+ N)~~
neuron m subnetwork 2. Tue
dynarnical
evolution is assumed to takeplace synchronically
atzero
(T
=
0)
noise;therefore,
these two subnetworks evolve in timeaccording
to tueupdating
rule
Si(t +1)
=
Sign(ht(t)), (1)
where
S~(t)
= +1 denotes tuedynamical
state of tue i~~ neuron at time t with= 1,..
,
N
corresponding
to subnetwork 1 and= N + 1,.
,
2N to subnetwork 2;
h)
is tue field at siteon the k~~
subnetwork,
due to the interaction with the rest of the neurons. For the first network this field isgiven by
N
hi (t)
=
L Àjsj, (2a)
with = 1,.
,
N,
and for the second networkhÎ(t)- f
j ~
~
v
j+£w~
s ~~ ~~~ ~~~
~ ~ ~~~~~
(2b)
with
= N +1,. ,2N. The interactions J~j =
Jj~
within each subnetwork are such thatthe first N components of each of p random unbiased patterns
(($), (($
= +1 with
equal probability),
bave been stored in subnetwork 1, and theremaining
N components in subnetwork 2; these connections aregiven by
j
f
(M(M
(3~)
lJ Jf i j'
p=i for1
# j
and within tue samesubnet,
and Jm= 0. W~j is an
asymmetrica1Hebb-like
connectionsending
informationunidirectionally
fromspin j
in subnetwork tospm1
in subnetwork 2, and isgiven by
P
Wq
"£ ($()ôi,j+N, (3b)
p=1
576 JOURNAL DE PHYSIQUE I N°5
for
j
= 1,..., N and= N +1,..., 2N. In tue last
expression
tue normalization was chosen in a way that subnetworks do notget
disconnected as N grows.In order to characterize how close tue system is, at any time, to tue pattern to be
recalled,
we define m~ to be tue
overlap
between tuedynamical
state of subnetwork k(S~)
and itscorresponding subpattern( (()), given by
~k jl
~(~)
N 1 ~'
i=(k-1)N+1
with k = 1,2 for the first and second subnets. Tue use of tue
subscripts (f)
and(o),
inm)
and
ml,
will account for final and initialvalues, respectively.
Notice that the maximum value for theoverlaps
has been normalized to 1 within thesubnets,
so the fraction of correct bits recoveredby
each of the subnetworks isgiven by ) (1+ m~). Therefore,
theoverlaps (m~)
are related to the total
overlap
q, between thecomplete patternl
and the state of the whole system, as follows~
11
~
~~ ~~ ~~~ ~
~~Î
3. Numerical Simulations
We
study
thedynamical
behaviour of the network forsynchronous updating
of thespin
statesas
given by equations (1-2).
As we mentionedbefore,
we are interested inevaluating
the "size"of the basin of attraction of a
subpattern, provided
subnetwork 1 hasalready
recovered itscorresponding
part. We considered values of Nranging
between 32 and1024,
and several values of p, thehighest
of thembeing
144 for thelargest
network(N
=
1024).
For each set of values p and Nconsidered,
we studied a collection of 100 dilferent neural networks where the patterns were storedaccording
to thelearning
rulesgiven by
equations(3).
In ail cases, the initial state of subnetwork 1 was set to be Si=
(),
for1 = 1,.,N (ml
+1),
whilevarious initial states for subnetwork 2 were
considered;
so we knew subnetwork 1 was at, or at least very close to, tue minimum related to thesubpattern.
After tue whole network reacheda
steady
state, the criterion used toidentify
whether or net it had recovered it, washaving
afinal overlap
mi
with this memory of order 1, while the absolute value of the otheroverlaps being significantly
lower(of
order1/@).
The first
thing
we noticed was that the basm of attraction related to thesubpattern,
if subnetwork hasalready
recovered its part, ishuge (much bigger
thon the size it would have for aHopfield
network with the same N and a and without anexterna1signal);
except for asmall
a-dependent
fraction of cases, shown mFigure
2, thefollowing happens:
If subnetwork is put in an initial state
given by
S~=
fi
for1= 1,.
..,
N,
som(
= 1, then the
persistent signal coming
from subnetwork 1 drives subnetwork 2 towards the minimum related to thesubpattern, independently
of the value ofm(. Furthermore,
the value of the finaloverlap mi
isgenerally
the same for any initial state of subnetwork 2(having
dilferent values for dilferent realizations(J~j)).
Crosses xx inFigure
3 show a typical relationship between tue initial and finaloverlaps
for anarbitrary
network with N= 1024 and p
= 127. Notice that
even for an initial state with
m(
= -1 tue system evolves towards tue same final state as it does when values
m(
m +1 are used. Similar results were found for most N and pconsidered,
even if the initial state of subnetwork 2 was identical to other nominated pattern;
however,
wefound a few isolated cases difficult to
quantify (always
for even p, and the lowest values of Nioo
go ,
* so
lu
8o
# 50
40 30
20
tu o
o ce
dilferent (J~j) for each p value,
x x x x x x x x x x x x x
x
x x x x
o o o
o
o o o o o"-
o o oE
~ o o o o o o o
o.2 -1
,~
Fig.
cases: (xx)
represent the
with
ecovered its depends on m(, is by iamonds (oo) for a case with N = 64 and
p =
4.
578 JOURNAL DE PHYSIQUE I N°5
l~Ù1 li=6
2
o
8
<00z
6 p=<2
?
o
8
992
6 p=<8
;
2
o
0
972
6 p=24
?
~0 0.2 o-Î o-G 0 8
~>
J
Fig. 4. Statistical study over 100 realizations in each case, of the shift in the positions of the attractors, for N
= 1024 and various
even values of p. The figure in the right upper corner indicates the percentage of cases in which subnetwork 2 did not recover the pattern. For odd p, there is
a similar behaviour, but the shift is bigger,
as con be seen in Figure 5.
considered)
where, even when the information was stillrecovered,
itsquality depended
on the initialoverlap m(.
This is illustrated for aparticular
case with N= 64 and p = 4
by
diamonds(c<)
inFigure
3.These above-mentioned results are
interesting,
asthey imply that,
ifml
= 1, a
particular
network is able to recall a
complete
pattern, or it is trot, with ana-dependent probability.
But if thishappens,
then the basin of attraction covers the whole space of states, since any initialstate for subnetwork 2 will evolve towards a minimum related to the
subpattern. However,
as can be seen in
Figure
4, thepositions
of these minima areconsiderably displaced
from the nominated patterns(as m)
< 1) which reflects a reduction inthe'quality
of the information recovered.Figure
4 shows a statisticalstudy
of the shift in the positions of the minima(mi),
for N =1024,
and various values of p(o).
Thesegraphs
show the percentage of times i~ in whichm)
was obtained as the finaloverlap,
with respect to the total of cases in which thesubpattern
was recalled. The
figure
in theright
upper corner indicates the percentage of cases in which thesubpattern
was notrecovered;
in ail those cases, subnetwork 2 never reached asteady
state butkept switching,
due to frustration, among several states with the same valuemi.
It isinteresting
to note that the behaviour of this modeldepends
on the parity of p: for even values the percentage of times the information is recovered isslightly
smaller thon it is for odd values;however the
quahty
of theinformation,
whenrecovered,
is better for even p thon it is for odd p(Fig. 5).
As a increases, thequality
of the recalled information diminishes and the shift in themean value <
mi
> grows; this elfect is stronger for small pvalues,
and becomes smoother as p grows. It is notpossible
togive
a cntical value for o, for finiteN,
smce these elfects increase09
os
oi
08 A
~, ~~
V 04
~~
02
_,~~
01
0
o 002 004 008 oo8 O.l o.12 o.14
7.
5.-Mean value of the final overlapce <m)
> with bars representing a =<
(m))2
> <(m))
>~ for N = 1024 and p taking values between 3 and 144. A dotted(solid)
bueconuects even
(odd)
values for p.gradually.
4. Discussion
It is very
important
toemphasize
that if the initial state of subnetwork is within the basin of attraction of one pattern, subnetwork 2 will evolve towards the attractor related to that same pattern with ahigh a-dependent probability [loi. Therefore,
in these cases, neither other memory nominated or net willplay
a rote in the recovery of informationby
subnetwork 2,as tue basin of attraction covers tue whole space of states. This is very dilferent from what
happens
in a usualHopfield model,
where tue average size of tue basins of attraction of each of tue p stored memories is tue same, so tue fraction of statesleading
to each of them diminishesas p grows.
In these two
coupled networks,
tuestrength
of tueincoming signais
in subnetwork 2 is soimportant,
that it induces tue latter into tue minimumrequired,
no matter its initial state.This
finding
is consistent with tue observation that nervous systems show a dilferent set ofactivity
patterns when observed infreely moving
animais than when observed in vitro, where manyimportant
alferentinputs
are eliminatedIll].
It aiseexplains
tuerepertoire
of dilferentcollective responses, of a group of
interacting
neurons,depending
on the afferentsignais.
It is difficult to compare the load
capacity
of this system with that of theHopfield
model oc mo.138 for N
- cc, since many dilferent criteria cari be used to make such an
analysis.
Notice that as subnetwork 1 is an N-elementHopfield Network,
we carnet store moresubpatterns (and
therefore morepatterns)
than p m acN;therefore,
tue relation between tue number p ofcomplete
patterns and tue total number 2N of unitsgives
us as a storage limit of tuecomplete
network half that of tue commonHopfield
Network.However,
in order to recoverinformation,
we aise need half the percentage of bits
required
to recover it in aHopfield
NN. For it isonly
necessary for the initial state of subnetwork 1 to be within the basin of attraction of a
given
subpattern, independently
of the initial state of subnetwork 2(it
could even coincide with any580 JOURNAL DE PHYSIQUE I N°5
of the other
subpatterns).
On the otherhard,
each of the2p subpatterns
could be consideredas a pattern
itself,
soby taking
thisposition,
forlarge
N this system wouldequalthe capacity
limit of the
Hopfield network,
but with half the total number ofconnections,
as these would be 2N~ for aHopfield
NN with 2N elements and N~ in this case; butwe would still need
only
ualf the bits which are necessary to recover eacu of the p patterns in the usual
Hopfield
model.Tue results tuat we found for this system
composed by
two finitecoupled
subnetworksare very
promising. Tuey
could be useful toimplement
cuains ofcoupled
networksstoring
informationcomposed
ofmeaningful subsets,
wuicu are to be recalledsequent1ally
in time and space. It isimportant
topoint
out tuat in tuismodel,
tue fraction of the total information needed to make the recoverypossible
is much lower than in theusua1Hopfield model,
since tuepersistent signal coming
from tue previous subnetwork when it has reached a stablestate acts as a turesuold on tue neurons of tue second subnetwork
drawing collectively
itsdynamic
state into tue basin of attraction related to tue stored memory inconsideration,
tuusplaying
a rote similar to tuat of an "init1al state".However,
one of tue mainproblems
foundis tuat information recalled
by
tue second subnetwork bas many errors witu respect to tue stored information(low mi values).
Tuis suggests to us, as a next step, tostudy
adynamical
process
composed by
as many stages as subnetworks arecoupled, wuere,
alter subnetwork n reacues asteady
state, tuesignal coming
from subnetwork(n
1) issuppressed.
In tuis way,subnetwork n would be allowed to searcu for a new minimum
taking
as an initial state tueone obtained wuen still connected to its
feeding
partner. Tuispossibility
iscurrently
underinvestigation
[12].Acknowledgments
Tue autuors wisu to tuank Dr.
Miguel
Virasoro foruaving suggested
us tostudy
a cuain ofnetworks,
and M.C. ArmandoReyes
forcomputing
support. Tuis work waspartially supported
by
DGAPAProject
No. 1N100294 of tue NationalUniversity
of Mexico(UNAM),
and aCONACYÎ scuolarsuip
for C.M. Numerical simulations wereperformed
on UNAM'SCray
Y-MP
4/432
supercomputer.References
[1] Hopfield J-J-, Froc. Nat. Acad. Soi. USA 79
(1949)
2554.[2] Hebb D.O., The Organization of Behaviour
(Wiley,
New York, 1949).[3] Amit D.J., Gutfreund H. and Sompolinsky H., Phys. Reu. A32
(1985)
1007.[4] This value for oc was calculated within the rephca symmetric approximation, m the limit N - ce.
[Si One word of caution about this analogy must be said: the analogy has its pitfalls, when used for the visual representation of letters, since here we are considering non-correlated patterns with
zero mean and without any spatial structure.
[fil Forrest B-M-, J. Phys. A21
(1988)
245.[7] Coolen A.C.C. and
Ruijgrok'Th.W.,
Phys. Reu. A38(1988)
4253.[8] Viana L. and Coolen A.C.C, J. Phys. I France 3
(1993)
777.(9] Virasoro M.A., Reich S. and Laurogrotto R., to be published.
[10] This o-dependent probability has its origm in the cross-talk noise.
[Il]
Johnson S-W-, Seutin V. and North R-A-, Science 258(1992)
665.[12] Viana L., in preparation.