• Aucun résultat trouvé

Supplemental Material for

N/A
N/A
Protected

Academic year: 2022

Partager "Supplemental Material for"

Copied!
2
0
0

Texte intégral

(1)

Supplemental Material for

Adaptive Cluster Expansion for Boltzmann Machines with Noisy Data

S. Cocco, R. Monasson

This document details the pseudo-codes useful for the practical implementation of the inference algorithm. Given a cluster Γ the partition function of theK–spin system restricted to Γ,

Z(J,Γ) = X

i=0,1;i∈Γ}

exp X

i∈Γ

hiσi+ X

i<j; i,j∈Γ

Jijσiσj

, (1)

can be computed in time∝2K. The entropy S(Γ) is then obtained from definition (2) (main text) and the use of a convex minimization algorithm. The computation of the entropyS0 requires the eigenvalues Λa of the matrixˆc(Γ), theK×K restriction of the matrixˆcto the spins in Γ:

S0(Γ) = 1 2

K

X

a=1

log Λa . (2)

In the case of undersampling, some eigenvalues Λa may be equal to zero. To regularize the expressions ofS andS0

we introduce aL2penaltyµX

i<j

Jij2 pi(1−pi)pj(1−pj) over the couplings, where µ >0. Formula (2) above becomes

S0(Γ) =1 2

K

X

a=1

log ˆΛa+ 1−Λˆa

, (3)

where

Λˆa= 1 2

Λa−µ+p

a−µ)2+ 4µ

(4) is strictly positive. The optimal value forµcan be determined through Bayesian inference, see D.J.C. MacKay,Neural Computation4, 415 (1991).

The first pseudo-code obtains ∆S(Γ) (through M¨obius inversion formula):

Algorithm 1Computation of cluster-entropy ∆S(Γ) Require: Γ (of sizeK),p, routines to calculateS0 andS

∆S(Γ)←S(Γ)−S0(Γ) SIGN←1

forSIZE =K−1to1do SIGN← −SIGN

forevery Γwith SIZE spins in Γdo

∆S(Γ)←∆S(Γ)+ SIGN×`

S(Γ)−S0)´ end for

end for Output: ∆S(Γ)

When the above routine is called several times (to compute the entropies of various clusters) a substantial speed-up can be achieved by memorizing the entropies ∆S(Γ) of every cluster. The above routine is simply turned into a recursive procedure by changing the line ∆S(Γ)←∆S(Γ)+ SIGN× S(Γ)−S0)

into ∆S(Γ)←∆S(Γ)−∆S(Γ) – the variable SIGN becomes useless. Hence, the computation ofS(Γ), which is the slowest step for large cluster sizes, is done only once for every Γ.

(2)

2 The core of our inference algorithm is the recursive building-up and the selection of new clusters:

Algorithm 2Adaptive Cluster Expansion Require: N, Θ,S0, routine to calculate ∆S(Γ) fromp

LIST← ∅ {All selected clusters}

SIZE←1

LIST(1)←(1)∪(2)∪. . .∪(N){Clusters of SIZE=1}

repeat{Building-up of clusters with one more spin}

LIST←LIST∪LIST(SIZE){Store current clusters}

LIST(SIZE+1)← ∅

forevery pairs Γ12∈LIST(SIZE)do

ΓI←Γ1∩Γ2 {Spins belonging to Γ1 and to Γ2} ΓU←Γ1∪Γ2 {Spins belonging to Γ1 or to Γ2} if ΓIcontains (SIZE-1) spinsand|∆S(ΓU)|>Θthen

LIST(SIZE+1)←LIST(SIZE+1)∪ΓU {add ΓU to the list of selected clusters}

end if end for

SIZE←SIZE+1 untilLIST(SIZE) =∅

S←S0,J← −ddpS0 {Calculation of S,J}

forΓ∈LISTdo

S←S+ ∆S(Γ),J←J−ddp∆S(Γ) end for

Output: S,Jand LIST of clusters

Références

Documents relatifs

2: The loading curve is the shear stress measured in the cell as a function of time during the growth of the crystal in the gap, which takes about 3 minutes in this case.. Near the

In this paper, the dual feasible functions will be applied to the knapsack poly- topes associated to different constraints of (8)–(11), namely the demand con- straints (9),

We study the performance of empirical risk minimization (ERM), with respect to the quadratic risk, in the context of convex aggregation, in which one wants to construct a

only supposed to be bounded. Moreover, no Haar condition is introduced. At each itération, we consider a séquence of nested minimization problems. The algorithm is based on an

We study the performance of empirical risk minimization (ERM), with re- spect to the quadratic risk, in the context of convex aggregation, in which one wants to construct a

The problem of convex aggregation is to construct a procedure having a risk as close as possible to the minimal risk over conv(F).. Consider the bounded regression model with respect

Since a gapped boundary is a gapped domain wall between a bulk topological order and the vacuum, regarded as the trivial topological order, our theory of gapped domain walls

We have shown by an example that the direct extension of the alternating direction method of multiplier (ADMM) is not necessarily convergent for solving a convex minimization model