OATAO is an open access repository that collects the work of Toulouse
researchers and makes it freely available over the web where possible
Any correspondence concerning this service should be sent
to the repository administrator:
tech-oatao@listes-diff.inp-toulouse.fr
This is a publisher’s version published in:
https://oatao.univ-toulouse.fr/2
7580
To cite this version:
Mouradi, Rem-Sophia and Goeury, Cédric and Thual, Olivier
and Zaoui,
Fabrice and Tassi, Pablo Physically interpretable machine learning algorithm on
multidimensional non-linear fields. (2021) Journal of Computational Physics, 428.
110074. ISSN 0021-9991.
Official URL:
Contents lists available atScienceDirect
Journal
of
Computational
Physics
www.elsevier.com/locate/jcp
Physically
interpretable
machine
learning
algorithm
on
multidimensional
non-linear
fields
Rem-Sophia Mouradi
a,b,∗
,
Cédric Goeury
a,
Olivier Thual
b,c,
Fabrice Zaoui
a,
Pablo Tassi
a,daEDFR&D,NationalLaboratoryforHydraulicsandEnvironment(LNHE),6QuaiWatier,78400Chatou,France
bClimate,Environment,CouplingandUncertaintiesresearchunit(CECI)attheEuropeanCenterforResearchandAdvancedTrainingin ScientificComputation(CERFACS),FrenchNationalResearchCenter(CNRS),42AvenueGaspardCoriolis,31820Toulouse,France cInstitutdeMécaniquedesFluidesdeToulouse(IMFT),UniversitédeToulouse,CNRS,Toulouse,France
dSaint-VenantLaboratoryforHydraulics(LHSV),Chatou,France
a
r
t
i
c
l
e
i
n
f
o
a
b
s
t
r
a
c
t
Articlehistory:
Received25May2020
Receivedinrevisedform6November2020 Accepted10December2020
Availableonline7January2021
Keywords:
Data-DrivenModel(DDM)
ProperOrthogonalDecomposition(POD) DimensionalityReduction(DM) PolynomialChaosExpansion(PCE) MachineLearning(ML)
Geosciences
In an ever-increasinginterestfor MachineLearning (ML) and afavorable data develop-ment context, we here propose an original methodology for data-based prediction of two-dimensional physical fields. Polynomial Chaos Expansion (PCE), widely used inthe UncertaintyQuantificationcommunity(UQ),haslongbeenemployedasarobust represen-tationforprobabilisticinput-to-outputmapping.IthasbeenrecentlytestedinapureML context,andshowntobeaspowerfulasclassicalMLtechniquesforpoint-wiseprediction. Some advantagesareinherenttothemethod,suchasitsexplicitnessand adaptabilityto smalltrainingsets,inadditiontotheassociatedprobabilisticframework.Simultaneously, DimensionalityReduction(DR)techniquesareincreasinglyusedforpatternrecognitionand datacompressionandhavegainedinterestduetoimproveddataquality.Inthisstudy,the interestofProperOrthogonalDecomposition(POD)fortheconstructionofastatistical pre-dictivemodelisdemonstrated.BothPODandPCEhaveamplyprovedtheirworthintheir respective frameworks.The goalof thepresent paperwas tocombine themfora field-measurement-basedforecasting. Thedescribedstepsare alsousefultoanalyzethe data. Somechallengingissuesencounteredwhenusingmultidimensionalfieldmeasurementsare addressed,forexamplewhendealingwithfewdata.ThePOD-PCEcouplingmethodology ispresented,withparticularfocusoninputdatacharacteristicsandtraining-setchoice.A simplemethodologyforevaluatingtheimportanceofeachphysicalparameterisproposed forthePCEmodelandextendedtothePOD-PCEcoupling.
©2020TheAuthor(s).PublishedbyElsevierInc.Thisisanopenaccessarticleunderthe CCBY-NC-NDlicense(http://creativecommons.org/licenses/by-nc-nd/4.0/).
1. Introduction
Deep Learning techniques(DL [1])and more generally Machine Learning (ML [2]), andtheir applications to physical problems(fluidmechanics[3];plasmaphysics[4];quantummechanics[5],etc.)havemadeapromisingtake-offinthelast few years.Thishasbeenparticularlythecaseforfieldswherethemeasurement potentialhasdramaticallyincreased(e.g. GeoscienceData[6]).Inthiscontext,learningtechniquesareofinteresttoestablishnon-linearphysicalrelationships from
*
Correspondingauthorat:EDFR&D,NationalLaboratoryforHydraulicsandEnvironment(LNHE),6QuaiWatier,78400Chatou,France.E-mailaddress:remsophia.mouradi@gmail.com(R.-S. Mouradi).
https://doi.org/10.1016/j.jcp.2020.110074
0021-9991/©2020TheAuthor(s).PublishedbyElsevierInc.ThisisanopenaccessarticleundertheCCBY-NC-NDlicense (http://creativecommons.org/licenses/by-nc-nd/4.0/).
Fig. 1. Representation of the POD-PCE ML approach.
the databya combinationofsteps,inparticularusingtransformation functions,to capturethe complexityofthe system [2].
In particular, multi-layer NeuralNetworks (NN) [7] are widely used for physical applications.Their popularitycomes fromthiscomplexstructure,whichmakesthemadaptableforvariousapplications[8,9].However,somelimitationsprevent theuseofNNforphysicalapplications:(i)itisdifficulttoprovideanexplicitinput-to-outputformulation,duetothe com-binationsofstepsinvolvedinthelearning(ActivationFunctions,HiddenLayers [1]).Physicalinterpretationoftheconstructed model is therefore tedious[10]; (ii) too manyhyper-parameters and choicesare involved, depending on the number of neuronsandlayers(curseofdimensionality)[11];(iii)nogeneralproofforthetheoreticalabilityofapproximatingarbitrary functionsisavailable,excepttheUniversalApproximationTheorem anditsextensions[12,13] forparticularcases.
Toovercometheselimitations,wepropose analternativeML method,basedonacouplingbetweenProperOrthogonal Decomposition (POD) [14] and Polynomial Chaos Expansion (PCE) [15,16]. This approach is proposed for the prediction ofspatially-distributed physicalfieldsthat varyin time.The ideais tousePOD toseparate thespatial patternsfromthe temporalvariations,thatarerelatedtotheconditioningparametersusingPCE.TocorrespondtocommonNNparadigms,an adequate representationofthisideaisgiveninFig.1.Inparticular, PODisusedforbothEncoding and Decoding whereas
PCEisusedasanActivationFunction intheLatentRepresentation [1]. TheproposedPOD-PCEaddressesthesedrawbacksofML.
(
i)
Itisexplicitandsimpletoimplement, asitconsistsoftheassociationoftwo lineardecompositions.PODisalinear separationof the spatiotemporal patterns [17], shown to be accurate for both linearand non-linearproblems [18], combiningsimplicityandrelevance.PCEisawell-establishedmethodinUncertaintyQuantification(UQ)[19,20],widely used for the study of stochastic behavior in physics [21,22]. It is a linear polynomial expansion that allows non-linearitiestobegradually addedto themodelby increasingthepolynomialdegree. Thelinearityandorthonormality ofthePODandPCEcomponentsandtheprobabilisticframeworkofPCEmaketheoutput’sstatisticalmomentseasier tostudy[23],enablingstraightforwardphysicalinterpretationofthemodel[24].(
ii)
Itonlyhastwo hyper-parameters:anumberofPODcomponents,andaPCEpolynomial degree.Both canbe chosen according to quantitative criteria[14,25]. All other forms of parameterization (choice of the polynomial basis) can be achieved with robust physical and/or statistical arguments [26], as assessed in the present paper. Furthermore, theorthonormality ofthe PODand PCEbases minimizesthe numberof componentsnecessary to captureessential variationsindata.Additionally,thePODmodescapturemoreenergythananyotherdecomposition[27],PCEisknown toexponentiallyconvergewithpolynomialdegree[16],andthecardinalityofthelattercanbereducedbysparsebasis selection[28].(
iii)
Itcan beconsidered asa universalexpansion forphysicalfield approximation: aphysical fieldhasa finitevariance, whichimpliesthatitbelongstotheHilbertspaceofrandomvariableswithfinitesecondordermoments.There there-foreexists a numerable setoforthogonal random variables, thatform thebasis of thisHilbertspace, on whichthe fieldofinterestcan beexpanded(strictequality,notapproximation)[20].Amathematicalsettingforbasis construc-tion based on input was established by Soize and Ghanem [26] for the general case of dependent variables with arbitrarydensity,providedthatthesetofinputsisfinite.Intheliterature,associatingregressiontechniquestoReducedOrderModels(ROM), thatincludePOD,isnot novel[29,
30].Thecitedstudies,however,focusedondimensionalityreduction,whereas theexplicitformulationandapplicabilityto complex physicalprocessesare emphasized inthe presentstudy.Secondly,coupling PCEtoROM was recentlyaddressed
[31,32] and theuseofPCEasML isconsistentwiththe workofTorreet al. [33],wherethe authorsshowedthat PCEis aspowerfulasclassicalMLtechniques.However,neitherspatiotemporalfieldsnorphysicalinterpretabilitywereaddressed. Thedatainthesestudieswereeitherobtainedfromnumericalexperiments,emulatedfromanalyticalbenchmarkfunctions such asSobolorIshigami,orbasedonone-dimensionaldatasets[33].Incontrast,theproposedPOD-PCE methodologyis herein assessedon two-dimensionalphysicalfields.Inparticular, atoy examplewheresyntheticdataare emulatedusing an analytical function (groundwater perturbations due to tidal loadings [34]), and a real data set (high-resolution field measurementsofunderwatertopography)areused.Althoughsimilarfromalearningpointofview,thesetwoapplications are characterized withdifferences. In particular, thetoy problemispurely parametric and controllable,whereas the real data concern temporaldynamics andare of limitedsize. The casesare therefore complementary, in the sense that they allow demonstrating differentpropertiesof the proposed methodology. Hence, usingthe particularities ofeach case, the studyconsistsin:i)theevaluationofthecombineduseofPODandPCEasMLforpoint-wiseprediction;ii)therobustness ofthemethodologytonoise;iii)theapplicationtofielddatawiththeinherentchallengesnotencounteredwithnumerical data (e.g.paucity); iv) a focuson model explicitnessas a key condition forphysical understanding andv) theinfluence of forcing variablesstudy, basedon a classicalmeasure ofimportance (Garson weights [35]) directlycomputedwith the POD-PCEexpansioncoefficients.
Thepaperisorganizedasfollows.Section2givesadetailedexplanationofthemethodology,withaproposalforphysical importance measures inSubsection2.2.2.Section 3deals withthe assessmentofthe methodologyonsynthetic data,for bothpredictionandphysicalinterpretation.Inparticular,therobustnessoftheapproachtonoiseisevaluatedinSubsection
3.3.The modelis then deployed onfield measurements in Section 4.The study caseanddataare described in 4.1. POD andPCEperformancesare thendemonstratedindependentlyin4.2witha deepphysicalanalysis. Theperformance ofthe POD-PCEpredictorisdiscussedin4.3.Asummaryofthestudyandperspectivesoftheproposedmethodologyarepresented inSection5.
2. Theoreticalframework
Inthissection,theobjectiveistodefinetheframeworkoftheproposedPOD-PCEMachineLearningmethodology,along withphysicalinfluenceindicatorsforthe inputs.Thisis theobjectofSubsection2.3,butfirst,areminder oftheexisting PODandPCEtheoreticalbasesispresentedin2.1and2.2respectively.
2.1. ProperOrthogonalDecomposition
PODisa dimensionalityreductiontechnique [17] thatiswell documentedinliterature [14,18].Theoretical detailsand demonstrationscanbefoundin[27,36].Forclarity’ssake,theessentialelementsofPODaresummarizedbelow.
The goal of POD is to extract the main patterns of continuous bi-variate functions. These patterns, when added and multipliedbyappropriatecoefficients,explainthedynamicsofthevariableofinterest:areal-valuedphysicalfield.
Letu
:
× T → D
beacontinuous functionoftwovariables(
x,
t)
∈
× T
.Thefollowingrelationshipsandproperties holdforany× T
andHilbertspaceD
characterizedbyitsscalarproduct(. ,
.)
D andinducednorm||.||
D.However,as isthecaseforamajorityofphysicalfields,weshallconsiderasasetofspatialcoordinates(e.g.
R
2 orR
3),T
aneventspace(e.g.parametersspace
R
V with V∈ N
∗,ora temporalsubset[
0,
T]
⊆ R
+),andD
asasetofscalarrealvaluesor vector realvalues(e.g.R
orR
2). PODconsistsinan approximationof u(
x,
t)
atagivenorderd∈ N
(Lumley[17])asinEquation(1), u
(
x,t)
≈
d k=1 vk(t)φ
k(
x) ,
(1)where
{
vk(.)
}
dk=1⊂
C(T,
R)
and{φ
k(.)
}
kd=1⊂
C(,
D)
,withC(A,
B)
denotingthe spaceof continuous functionsdefinedover
A
andarrivingatB
.TheobjectiveofPODistoidentify{φ
k(.)
}
dk=1 thatminimizesthedistanceoftheapproximationfromthetruevalue u
(.,
.)
,overthewhole× T
domain,withanorthogonalityconstraintfor{φ
k(.)
}
dk=1 usingthescalar product(. ,
.)
D.Thiscanbedefined,intheleast-squaressense,asaminimizationproblem.The minimization problem is defined for all orders d
∈ N
, so that the membersφ
k are ordered according to theirimportance.Inparticular,fororder1,
φ
1isthelineargeneratorofthesub-vectorspacemostrepresentativeofu(
x,
t)
inD
.For
D
=
Im(
u)
,thefamily{φ
k(.)
}
dk=1iscalledthePODbasisofD
ofrankd.Thesolutiontothisproblemhasalreadybeen established in literature [17,37]. The theoretical aspects ofPOD and demonstrations of mathematical properties can, for example,befoundin[27]:thePODbasisofD
oforderd istheorthonormalsetofeigenvectorsofanoperatorR
: D → D
definedasRφ = (
u, φ)
D×
uT,iftheeigenvectorsaretakenindecreasingorderofthecorrespondingeigenvalues
{λ
k}
dk=1.Forthisexpansion,anaccuracyrate,alsocalledtheExplainedVarianceRate(EVR),denoteded atrankd,canbe calcu-latedasinEquation(2).EVRtendsto1(perfectapproximation)whend
→ +∞
.ed
=
k≤dλ
k +∞ k=1λ
k.
(2)Inpractice,for
D
= R
,whenu(.,
.)
isadiscretesample onasetofm∈ N
spacecoordinatesX = {
x1,
. . . ,
xm}
andfor n∈ N
measurement eventsT = {
t1,
. . . ,
tn}
(e.g.realizationsof theparameters, timecoordinates, etc.),the available datasetisarrangedinamatrixU
(
X ,
T )
= [
u(
xi,
tj)
]
i,j∈ R
m×n,calledthesnapshotmatrix,soastobeabletoworkinadiscretespace.ThePODproblemformulatedinEquation(1) canbewritteninitsdiscreteformasU
(
X ,
T )
=
(d)(
X )
V(d)(T )
,where(d)
(
X )
:= [φ
j
(
xi)
]
i,j∈ R
m×d andV(d)(
T )
:= [
vi(
tj)
]
i,j∈ R
d×n.TheproblemcanthereforebeviewedasifworkingwithanewfunctionU
(
X ,
.)
= [
u(
xi,
.)
]
i∈{1,...,m}:
T → D = R
M.Then,theaverageoverT
canbe definedasthestatisticalmeanoverthesubset
T
,andthescalarproduct(. ,
.)
DasthecanonicalproductoverR
m.ThePODoperatorR
canbewrittenas inEquation(3),R
φ(X ) =
1 n n j=1 U(
X ,
tj)
T(
X )
U(
X ,
tj)
=
1 nU(
X ,
T
)
U(
X ,
T
)
T(
X ) ,
(3)where U
(X ,
tj)
= [
u(
xi,
.)
]
i∈{1,...,m} isthecolumnnumber j of thematrixU(
X ,
T )
(i.e realizationtj ofthemeasurementover
X
),and(
X )
= [φ(
xi)
]
i∈{1,...,m}.AsfindingthePODbasisisequivalenttoidentifyingtheorthonormalsetof eigenvec-torsoftheoperatorR
,thenforthisdiscreterepresentationtheproblembecomesequivalent tosolvingtheeigenproblem ofthematrixR:=
n1U(
X ,
T )
U(
X ,
T )
T,calledthecovariancematrix.Anumberd∈ N
ofeigenvectors(
X )
areidentifiedandstoredinthecolumnsofthematrix
(d)
(
X )
.FortheeigenvaluesofthecovariancematrixR denoted{λ
k}
dk=1,theex-pansioninEquation(1) canalsobewrittenasinEquation(4),where
{φ
k(.)
}
dk=1togetherwith{
ak(.)
}
d k=1arebi-orthonormal, andvk
(.)
=
ak(.)
√
n× λ
k. u(
x,t)
≈
d k=1 ak(t
)
n× λ
kφ
k(
x) .
(4)BydefiningthematrixA(d)
(T )
:= [
ai
(
tj)
]
i,j∈ R
d×n andtheoperatorD(d)(λ
1,
...,
λ
d)
correspondingtothediagonalmatrixofelements
λ
i,wehaveU(
X ,
T )
=
(d)(
X )
D(d)(
√
n× λ
1,
...,
√
n× λ
d)
A(d)(
T )
.ThereforethetransposedformisU(
X ,
T )
T=
A(d)
(T )
TD(d)(
√
n× λ
1,
...,
√
n× λ
d)
(d)(
X )
T. Thanksto theorthonormalityof{
ak(.)
}
dk=1,the covariancematrixreadsR=
1
n
(
d)
(
X )
D(d)(
n× λ
1
,
...,
n× λ
d)
(d)(
X )
T=
(d)(
X )
D(d)(λ
1,
...,
λ
d)
(d)(
X )
T.Whenn
<<
m,itismorecomputationallyefficienttosolvetheeigenproblemofRT insteadoftheeigenproblemofR ashighlightedbySirovich [37].Thisisoftenthecasewhenalimitednumberofoccurrencesismeasuredforatwo-dimensional physicalfield,asisthecaseencounteredfortheapplicationdescribedinSection4.
When an order d
<<
min(
m,
n)
corresponds to a high EVR as defined in Equation (2), we speak of dimensionality reduction, because the data are projected in a sub-space that is of much smaller dimension thanR
m×n. When diverseenoughrecordsareavailableforthevariableunderstudy,wemayconsiderthat
{φ
k(
X )}
dk=1= {[φ
k(
xi)
]
i∈{1,...,m}}
dk=1,i.e.theresulting POD basis,is a generator ofall possiblestates. Predictingthe associated expansion coefficients
{
ak(
t)
}
dk=1 foragiveneventt wouldthereforebeenoughtopredictthewholestate.Hence,weproposetousethePODasabasisextractor. Thiswouldfirstenable studyofthedynamicsofthevariableofinterestandeventually extractionofphysicalinformation, as shownin theapplications Sections 4and 3. Then, the basis can be used asa generator forthe predictionof diverse states.Thisimpliespredicting
{
ak(
t)
}
dk=1,forwhichwepropose tousePolynomialChaosExpansion(PCE),asdescribed inthefollowingSection2.2.
2.2. PolynomialChaosExpansion
AreminderofthetheoreticalbaseofPCEispresentedinSubsection2.2.1.Theoreticaldetails,demonstrationsand inter-esting referencescanbe foundin [23,19]. Then,a simpleindicator isproposed inSubsection2.2.2forthe analysisofthe variablesinfluenceontheoutputvalue.ThelatterislatergeneralizedforPOD-PCEinSection2.3.
2.2.1. Learning
The idea behind Polynomial ChaosExpansion (PCE)is to formulate an explicitmodel that linksa variable ofinterest (output)toconditioningparameters(inputs),bothinaprobabilityspace.Thisenablesthepropagationpathofprobabilistic information(uncertainties,occurrence frequencies)tobe mappedfromthe inputto theoutput space. Thevariable of in-terest,Y,andtheinputparameters
= (θ
1,
θ
2,
...,
θ
V)
arethereforeconsideredrandomvariables,characterizedbyagivenProbabilityDensityFunction(PDF)denoted f.ItshouldbekeptinmindthattheoutputsofourproblemarethePOD ex-pansioncoefficientsY
= [
ak(
t)
]
k∈{1,...,d},andthattheinputscorrespondtophysicalforcings,asdescribedlaterinSection2.3.The objective isto derive the variations of the PODcoefficients asthe outcome ofthe forcings.Let usnow recall some fundamentalsofthemathematical probabilisticframework, takingtheexampleofaone dimensionalreal-valued variable. Thedefinitionscanbeeasilyextendedto
R
M.Let
(,
F,
P )
be a probability space, whereis the eventspace (space ofall the possible events
ω
) equipped withσ
-algebra F (someeventsof)anditsprobabilitymeasure
P
(likelihoodofagiveneventoccurrence).Arandomvariable definesanapplicationY(
ω
)
:
→
DY⊆ R
,withrealizationsdenotedby y∈
DY.ThePDF ofY isa function fY:
DY→ R
thatverifies
P (
Y∈
E⊆
DY)
=
E fY(
y)
dy.Thekth moments ofY aredefinedasE
[
Yk]
:=
DY y
kfY
(
y)
dy,thefirstbeingtheexpectationdenoted
E
[
Y]
.Inthesamemanner,wedefinethekth centralmoments ofY asE
[(
Y− E[
Y])
k]
,thefirstbeing0 andthe secondthe varianceof Y denotedby
V
[
Y]
.The covarianceoftworandomvariables isdefinedascov(
X,
Y)
=
E
[(
X− E[
X])(
Y− E[
Y])]
andaresultingpropertyisV
[
Y]
=
cov(
Y,
Y)
.ReturningtothePCEconstruction,inputs
= (θ
1,
θ
2,
...,
θ
V)
areconsideredtoliveinthespaceofrealrandomvariableswithfinitesecondmoments(andfinitevariances).Thisspaceisdenotedby
L
2R(,
F,
P
;
R)
andisaHilbertspaceequipped with the inner product(θ
1,
θ
2)
L2R
:= E[θ
1θ
2]
=
θ
1(
ω
)θ
2(
ω
)
dP (
ω
)
and its induced norm||θ
1||
L2 R:=
E
[θ
21
]
. The PCEobjectiveistomaptheoutputspacefromtheinputspacewithamodel
M
asinEquation(5):Y
=
M
()
=
I⊆{1,...,V}M
I(θ
I)
=
M
0+
V i=1M
i(θ
i)
+
1≤i<j≤VM
i,j(θ
i, θ
j)
+ ... +
M
1,...,V(θ
1, θ
2, ..., θ
V) ,
(5)where
M
0 isthe expectation of Y andM
I⊆{1,...,V} represents thecommon contribution ofthe variablesI ⊆ {
1,
...,
V}
to the variation in Y . For the PCE model, these contributions have a polynomial form. We shall define, for each input variable
θ
i,an orthonormalunivariate polynomial basisξ
β(i)(.), β
∈ [|
0,
p|]
where p
∈ N
is a chosen polynomialdegree andξ
β(i)(.)
isof degreeβ
.The orthonormality isdefined withrespect to the inner product(.,
.)
L2R. If we introduce the
multi-indexnotation
α
= (
α
1,
...,
α
V)
∈ N
V sothat|
α
|
=
Vi=1α
i,wecandefine amultivariatebasisζ
α(.),
|
α
| ∈ [|
0,
p|]
asζ
α(θ
1,
θ
2,
...,
θ
V)
:=
iV=1ξ
(i)αi
(θ
i)
.Therefore,themodelinEquation(5) canbewrittenas:Y
=
M
()
=
|α|≤P
cα
ζ
α(θ
1, θ
2, ..., θ
V) ,
(6)wherecα
∈ R
are deterministiccoefficientsthat canbe estimatedthanksto differentmethods.Itcan be formulatedasa minimization problem, andregularization methodscan be usedwhen dealingwith smalldatasets. Inthepresentstudy, weusedtheLeastAngleRegressionStagewisemethod(LARS)inordertoconstructanadaptivesparsePCE.Itisaniterative procedure,consistingonanimprovedversionofforwardselection.Thealgorithmbeginsbyfindingthepolynomialpattern, heredenotedζ
i forsimplicity,that isthemostcorrelatedtotheoutput.Thelatterislinearlyapproximatedbyi
ζ
i,wherei
∈ R
. Coefficienti is not setto its maximal value, butincreasedstarting from0,until another pattern
ζ
j is found tobe ascorrelatedto Y
−
i
ζ
i,andsoon.Inthisapproach,acollectionofpossiblePCE,orderedbysparsity,isprovidedandan optimumcanbechosenwithanaccuracyestimate.Itwasperformedinthisstudyusingcorrectedleave-one-outerror. The reader can refer to the work ofBlatman and Sudret [25] forfurther details on LARS andmore generallyon sparse constructions.
Thechoiceofthebasisiscrucialandisdirectlyrelatedtothechoiceofinputvariablemarginals,viatheinnerproduct
(.,
.)
L2R.Chaospolynomialswerefirstintroducedin[38] forinputvariablescharacterizedbyGaussiandistributions.The
or-thonormalbasiswithrespecttothismarginalistheHermitepolynomialsfamily.Later,otherAskeyschemehypergeometric polynomial families were associatedto some well-knownparametric distributions [39]. Forexample,theLegendre family is orthonormalwithrespecttothe Uniformmarginals.This iscalled g P C (generalizedPolynomialChaos) whenvariables ofdifferentPDFsareusedasinputs.Inpracticehowever,theinputdistributionsofphysicalvariablescanbedifferentfrom usualparametricmarginals.Insuchcases,themarginalscanbeinferredbyempiricalmethodssuchastheKernelSmoother (see[40] fortheoreticalelements).Inthiscase,anorthonormalpolynomialbasiswithrespecttoarbitrarymarginalscanbe builtwithaGram-Schmidtorthonormalizationprocessasin[41] orviatheStieltjesthree-termrecurrenceprocedureasin [42].
Tohighlighttheimportanceofthemarginalsandchoiceofpolynomialbasisforthelearningprocess,several configura-tions areattempted inSection4.Different inputsetsanddistributions (Gaussian,Uniform,inferred byKernelSmoothing) weretested.TheinfluenceofthepolynomialbasisontheMLisinvestigatedinSection4.2.2.
2.2.2. Physicalimportancemeasures
OncethePCEconstructionisachieved,aphysicalinterpretationcanbeperformed.ItisnotablethatclassicalNN indica-torscanbeused[35].ThePCEcanberepresentedintheFeedforwardNNparadigmasinFig.2.Suchnetworksareclassically composed,inadditiontotheinputandoutputlayers,ofsuccessivehiddenlayers.Eachhiddenlayeriscomposedofneurons
that transformthe variables of the previous layer (outputs of the previous neurons)into a new set of variables. Thisis donebycombiningalineartransformation,givingdifferentweights tothepreviousneurons,andatransformationfunction, calledActivationFunction (AF). Thissuccession of layers iscalled thelatentrepresentation. Fora numberof hiddenlayers
L
≥
1,theNNcan be formallywritten asY≈
fout(
AL fL(. . .
A2 f2(
A1 f1(
Ain))))
where{
Ak}
k∈[|1,L|] and{
fk}
k∈[|1,L|] are thehiddenlayerweightmatricesandAFs,Ain istheinput-to-hiddenconnectionmatrixand fout isthefinalhidden-outputtransformation[2].
The PCE-basedNNrepresentedinFig.2isasingle layerfeedforward,composed ofl
∈ N
neurons,that canbe written asY≈
fout(
A1 f1(
Ain))
.ThefirstmatrixAin istheinput-to-hiddenconnectionmatrixofdimensionV×
V ,thatlinkstheinputlayertothePCEhiddenlayercontainingthemultivariatepolynomials
ζ
α
,
α
∈ {
α
1, ...,
α
l}
Fig. 2. Representation of the PCE learning in the NN paradigm.
inputsandthemultivariateindexes
{
α
1,
...,
α
l}
areconditionedbythechosenpolynomialdegree p suchas∀
i∈ [|
1,
l|]
0≤
|
α
i|
≤
p,andby thenumberofselectedfeatures ifa sparsepolynomial isconstructed,asinthepresentcaseusingLARS[28]. Matrix Ain represents thecontributions of the V variables to the multivariate polynomials
ζ
α
,
α
∈ {
α
1, ...,
α
l}
.It is a diagonal matrix such that
[
Ain]
j,j∈[|1,V|]2 is 0 if∀
i∈ [|
1,
l|] (
α
i)
j=
0 and 1 if not. The first multi-dimensional AF f1 is a vector ofmultivariate functionsthat transforms theset of selectedinputs corresponding to[
Ain]
i,i∈[|1,V|]2=
1 tothe multivariate polynomials of the chosen basis (Hermite, Legendre, etc.) by tensor product over the univariate basis. The hiddenlayerweightmatrixA1 givesdifferentweightsto theconstructedpolynomial features.It isadiagonal matrix composedofthePCEexpansioncoefficientssuchas
[
A1]
i,j∈[|1,l|]2= [
cαi]
i∈[|1,l|].The final hidden-output transformation fout is a summation.Fig. 2 can alsobe presented differently:another hidden layer can beadded to thePCElatent representationas Y
≈
fout(
A2 f2(
A1 f1(
Ain))
.The first AF f1 wouldrepresentatransformation ofeach input variableto a list ofmonomialsof degrees1 to p (here, Ain isidentity). The second AF f2
thereforerepresentsthetensorproductthattransformsthedifferentmonomialstomultivariatefeatures,withA1
appropri-atelyfilledwithzerosandones,and
[
A2]
i,j∈[|1,l|]2= [
cαi]
i∈[|1,l|].To capture the importance of each feature, the Garson relative Weights (GW) defined in Equation (7) are a classical measure to quantify the relative importance of each neuron of the last hiddenlayer, andtherefore of each polynomial pattern,fortheoutputvalue[35,43].
wζ α
=
|
cα|
0≤β≤1|
cβ|
.
(7)This measure can be used to understand the importance given by the NN algorithm to the variables andtheir possible interactions, especially when using feature selection algorithms asLARS: “feature interactions [...] are created at hidden unitswithnonlinear activationfunctions, andthe influencesofthe interactionsare propagated layer-by-layerto thefinal output”[43].Intheparticularcaseofapolynomialexpansion,theinterpretationisstraightforward,theimportanceofeach variablealonecorrespondstoitsmonomials,andtheimportanceofitsinteractionswithothervariablescorrespondstothe multivariatepolynomialsinwhichitisinvolved.
FortheparticularcaseoftheorthonormalbasisprovidedbyPCE,theGWdefinedin(7) canbeinterpretedintermsof Pearson’scorrelationsbetweenoutputY andthepolynomialbasiselements
ζ
α denotedρ
Y, ζα,with
α
= (
0,
...,
0)
.Indeed, Pearson’scorrelationsρ
Y, ζα aredefinedasinEquation(8),
ρ
Y, ζ α=
E
(Y
− E(
Y))(ζ
α− E(ζ
α))
V (
Y)
V (ζ
α)
=
cα 1≤|β|≤pc2β
,
(8)thankstotheorthonormalityofthebasiswithrespecttothescalarproduct
(. ,
.)
L2R thatguarantees:
• E
ζ
α=
ζ
α, ζ
β=(0,...,0)=
1 L2 R=
0;• E
[Y ]=
0≤|β|≤pcβζ
β, ζ
β=(0,...,0) L2 R=
cβ=(0,...,0);• E
Y, ζ
α=
0≤|β|≤pcβζ
β, ζ
α L2 R=
cα;• V
ζ
α= E
ζ
α− E
ζ
α2
= E
ζ
α2= ||ζ
α||
2L2 R=
1;• V
[Y ]= E
Y− E
[Y ]2=
1≤|β|≤pcβζ
β,
1≤|β|≤pcβζ
β L2 R=
1≤|β|≤pc2β.Therefore, the weights wζ
α can alsobe computedas
|
ρ
Y, ζα|/
1≤|β|≤p
|
ρ
Y, ζβ
|
. Thismeans that they measure theofthe latter.These“relativePearson’s correlations”can beseen asaphysicalcontributionsince thePCEmodelisstrictly linear.ThesumoftheGW wζ
α forallthepolynomialfeaturesequals1.Thismeansthattheyallow
{ζ
α}
|α|≤p toberanked in terms ofrelative contribution to the output Y . The contributions can be analyzed eitherfor each polynomial pattern separately,orforasinglevariableθ
i byaddingallthepolynomialsharesrelatedtothisvariablealone,orbyaddingallthepolynomialsharesrelatedtothisvariableanditsinteractions(Sobol’indicesanalogy[23]).
2.3. POD-PCEbasedpredictor
PODandPCEwere introduced separatelyinSubsections 2.1and2.2respectively.We arenow fullyequippedwiththe adequate theoretical basis andmathematical notations,topresentthe POD-PCE ML methodologyfor adata-based model learning of a multidimensional physical field. In this Subsection, we will first summarize the proposed approach, then theformal details ofthecouplingwillbe givenwiththe definitionofadequate accuracymeasures. Finallythepreviously discussedimportancemeasureswillbegeneralizedforthePOD-PCEphysicalstudy.
TheproposedPOD-PCEMLconsistsofsteps,inalearningandapredictionphase,summedupasfollows:
•
Learningphase:∗
PODbasis construction:givenasetofmeasurements U(
X ,
T )
= [
u(
xi,
tj)
]
i,j∈ R
m×n (snapshotmatrix), constructaspatial POD basis accordingly.Variable tj can represent time in caseof temporal dynamics,or more generallyan
occurrenceofU
(
X ,
.)
.Then,ingeneral,T
wouldbeaneventspace;∗
PCE learning: constructPCE models that map each POD coefficient, obtained inthe previous step along withthe spatial basis, to a set ofinputs. In theparticular caseof temporaldynamics, previous valuesof the physicalfield, representedby previous PODcoefficients, can bepartof thelearning inputs. Forexample,onecould usean initial fieldvalue U(
X ,
tj)
tolearnafuturefieldU(
X ,
tj+1)
fromasetofphysicalparametersthatcondition theevolutionover
[
tj,
tj+1]
.Thelattercanconsistintimeseriesofphysicalvariables,representativestatisticsofthelatter,physicalconstants,etc.andcanbedenoted
(
tj→
tj+1)
;•
Predictionphase:∗
Givena newrealizationofthe inputs,predict thenewPODcoefficientsusingthelearnedPCEmodels,then recon-struct the newestimate U(
X ,
tk)
on thePOD basis. As previously explained, an initial value to the physicalfield,denotedU
(
X ,
tk−1)
,maybe partoftheinputsfortemporal dynamics.Inparticular, itsreducedform, consistingintemporal PODcoefficients, isused. Inthis case, an additionalstep is needed:U
(
X ,
tk−1)
is projected onthecon-structedPODbasisinordertoretrievethevaluesofassociatedPODcoefficientswhicharethenusedasPCEinputs. The learning andpredictionset-upsaremore complextoestablishfortemporal evolutionproblems,because thefield informationatprevioustimesarerequired.Therefore,forthesakeofclarity,thestepsareexplicitlydevelopedinthe follow-ingSubsection2.3.1.TheaccuracyofthemethodologyislaterdemonstratedonbothaparametrictoyprobleminSection3, and a field measurements-based temporal problem in Section 4. Thesetwo can be considered as complementary appli-cations, anddemonstratethat thePOD-PCE MLcanbe appliedindifferentlearning set-upsofmulti-dimensionalphysical fields.Similaritiesinthetreatmentofbothproblemscanbenoticed,buttheirparticularitiesarealsousedtodemonstrate differentpropertiesofthePOD-PCElearning,thatareshortlydescribedatthebeginningofeachsection.
2.3.1. Machinelearningmethodology
Here, the formal hypothesis behind the POD-PCE ML reasoning and its mathematical formulation are discussed. Let
U
(
X ,
.)
= [
u(
xi,
.)
]
i∈{1,...,m}beafieldofinterestdefinedonasetofm∈ N
spacecoordinatesX = {
x1,
. . . ,
xm}
.Let(.)
=
(θ
1(.),
θ
2(.),
...,
θ
V(.))
be a vector of the inputs supposed to condition the evolution of U(
X ,
.)
over time. The dynamicmodel,denoted
H
,that givesan estimationofafuture state U(
X ,
tj+1)
fromapast state U(
X ,
tj)
andan estimationof(
tj→
tj+1)
overthetimeinterval[
tj,
tj+1]
,wheretj<
tj+1∈ R
+,isformulatedasinEquation(9).U
(
X ,
tj+1)
≈
H
U
(
X ,
tj),
tj+1−
tj, (t
j→
tj+1)
.
(9)If thefield of interest hasbeen recorded overa set of pasttimes
T = {
t1,
. . . ,
tn}
⊂ R
+, wheretj<
tj+1,a PODba-sis can be constructed as in Section 2.1, consisting of d
∈ N
vectors of dimension m stored in a matrix as(d)
(
X )
=
(
1(d)(
X ),
. . . ,
d(d)
(
X ))
∈ R
m×n,andcanbe seenasageneratorofall possiblestatesifenoughrecordsareavailable.Ifso,anyfuturestateU
(
X ,
tj)
canbeexpandedonthisPODbasisandtheassociatedtemporalcoefficientsaresimplytheweightsofU
(
X ,
tj)
onthePODbasis.TheyarethereforeobtainedusingthecanonicalscalarproductoverR
m,asinEquation(10). U(
X ,
tj)
≈
dk=1ak(t
j)
√
n× λ
k(kd)
(
X )
≈
d k=1(
U(
X ,
tj),
k(d)(
X ))
Rm(d) k
(
X )
≈
d k=1U(
X ,
tj)
Tk(d)
(
X )
(kd)(
X ) .
(10)Hence,thevariablepartofU
(
X ,
tj)
isfullyexpressedinthetemporalcoefficientsak(
tj)
.ThefieldofinterestU(
X ,
tj)
canbyarandomprocess“inthesensethatnaturehappenswithoutconsiderationofwhatcouldbethebestrealizationsforthe learning algorithm” [2].Therefore,the coefficientsak
(
tj)
canalsobe seenasthe jth realizationofarandom variable Ak.WecanthereforeconstructaPCEapproximation
H
k thatmaps Ak fromitsinput space.Thelatteristakenasacollectionof randomvariables, composed fromthe set
(
A1,
...,
Ad)
at aprevious time, theduration of thedynamic, andthe inputvariables
(
tj→
tj+1)
.ThisisformulatedasaclassicaldynamicmodelinEquation(11).ak
(t
j+1)
≈
H
ka1
(t
j), . . . ,
ad(t
j),t
j+1−
tj, (t
j→
tj+1)
.
(11)Themodel
H
inEquation(9) isapproximatedasinEquation(12).H
U(
X ,
tj),
tj+1−
tj, (t
j→
tj+1)
≈
d k=1H
k a1(t
j), . . . ,
ad(t
j),t
j+1−
tj, (t
j→
tj+1)
n
× λ
k(kd)
(
X ) .
(12)SomelimitationstotheintroducedformulationsinEquations(11) and(12) canbehighlighted.Afirstlimitationconcerns discontinuities thatcan bemetinphysicalfields.Thiscanoccureitherinthecomplete spatialfield U
(.,
.)
,inits reduced version representedbythe PODcoefficientsak(.)
,orintheinputs.Inthefirstcase, theclassicallinearapproximations
as POD may be inefficient [44]. One solution developed by Taddei [44], called RePOD (Registration POD), consists in a parametrictransformationoftheinterestdiscontinuousfieldintoasmootheroneforlineartransformations.Inthesecond case, wherediscontinuity happensinthePOD temporalcoefficients, thiswouldimpact thelearning withPCE.Innovative solutions were identified toapply PCEwhen the output’s space ischaracterized with rapidvariations ordiscontinuities, for instance near a critical point in the inputs space. As an example, a method calledadaptive Multi-Element PCE was developed for Legendrepolynomialsin [45] andextended toarbitrarymeasures in[42]. The inputsspaceisdecomposed toaunionofsubsets, andtheoutputvariableislocallyexpandedoneachsubset.Thefinalsolutionisthenacombination ofPCEsub-problems.Inthelast discontinuitycasethatconcernsthe inputs
,theprevious splittingtechniquescanalso beused.Forexample,thesub-intervalsintheinputsspacecanbeconstructedinsuch waytoavoidthediscontinuity.PCE sub-problemswouldthereforebetreatedasusual.
Asecondlimitationconcernsthechoiceofinputvariablesforregressionmodels,andisanongoingresearchquestionin statistics[46].Asapracticalillustration,thedynamicalproblemwritteninEquation(11) canincorporateadditionalinputs, forexamplethe informationatprevious timestj−1,tj−2,etc.However, whena large setofinputscanbe used andonly
asmallsetofrealizationsisavailable forlearning,awell-posednessproblemoccurs.Onesolutionconsistsintransforming thelargesetofinputstoa reducedversion,forexamplewiththehelpofPCA[47] forDR.Thisapproachwasnot studied hereandwill be thetopicofa futurestudy.However, differentinput configurations willbe evaluated, toinvestigatethe influenceofvariableselection ontheproposedlearning. Forexample,thehypothesis ofdependencebetweentherandom variables
(
A1,
. . . ,
Ad)
could be relaxed. Thiswould implywritingthe approximation in Equation(11) in a relaxedformas
H
kak
(
tj),
tj+1−
tj, (
tj→
tj+1)
.Inthatcaseasimplermodel
H
,underthestrongindependenceassumption,canbe formulatedasinEquation(13).H
U(
X ,
tj),
tj+1−
tj, (t
j→
tj+1)
≈
d k=1H
k ak(t
j),t
j+1−
tj, (t
j→
tj+1)
n
× λ
k(kd)
(
X ) .
(13)Both alternativesare testedinSection4.Toinvestigatetheinfluenceofinputselection onlearning accuracy,a quanti-tative evaluationofthe hypothesisis needed.Moregenerally, whetherforthe above-mentionedsimplificationsorforthe approximatedformofthemodelingeneral,accuracyestimatorsareneeded.Thesearepresentedbelow.
2.3.2. Accuracytestsfortheapproximation
There are two determining parts in the POD-PCE learning process. Firstly, the PCE learning
H
k(.)
of each mode Akshould beasaccurateaspossible.Secondly,thereconstructedfield
dk=1H
k(.)
√
n
× λ
k(kd)
(
X )
foragivenrankd shouldbeasclosetotherealfield U
(
X )
aspossible.The distance between each mode and its PCE approximate can be evaluated using the generalizationerror, denoted
δ(
Ak,
H
k)
anddefinedasinEquation(14).δ(A
k,
H
k)
= E
(
Ak−
H
k(.))
2.
(14)Forthe modeldefinedinEquation(13),thiserrorcan be estimated,on aset ofpairedrealizations
(
ak(
t1),
. . . ,
aj(
tn))
and
((
t0→
t1),
. . . ,
(
tn−1→
tn))
,asin Equation(15) asexplained by Blatman [28].This approximatedversion of thegeneralizationerror iscalledtheempiricalerror.
δ(A
k,
H
k)
≈ δ
emp(
Ak,
H
k)
:=
1 n n j=1 ak(t
j)
−
H
k ak(t
j−1),
tj+1−
tj, (t
j−1→
tj)
2
.
(15)Itsrelativeestimatedenoted
emp
(
Ak,
H
k)
canbedefinedasinEquation(16).emp
(
Ak,
H
k)
:=
δ
emp(
Ak,
H
k)
V
[
Ak]
.
(16)OncethePCElearningscanbetrusted,thedistanceattimetj betweenthetruestateU
(
X ,
tj)
andthePOD-PCEapprox-imation
H
U(
X ,
tj),
tj+1−
tj, (
tj→
tj+1)
canbe defined.It mightbe estimatedusing therelativeRoot MeanSquared Error (relativeRMSE), denoted r
[
U,
H](
tj)
and calculated asin Equation (17), where h(
xi,
tj)
refers to the value of thePOD-PCEapproximationatcoordinatexi andtimetj.
r
[
U,
H
](
tj)
:=
mi=1 u(
xi,t
j)
−
h(
xi,t
j)
2 m i=1 u
(
xi,
tj)
2
.
(17)AmeanvalueoftherelativeRMSEiscalculatedoverasetofrealizationscorrespondingtoasetoftimes
T = {
t1,
. . . ,
tn}
.Itisdenotedr
[
U,
H]
(T )andestimatedasinEquation(18).r
[
U,
H
]
(T ):=
1 n n j=1 r[
U,
H
](
tj) .
(18)OncetheaccuraciesofthePCElearningsandthe POD-PCEcouplinghavebeenevaluated,afinal model,whichwillbe themostaccurateone,canbechosen.Thismodelwould, forourMLset-up,bethebestrepresentationofthedependence structurebetweeninputsandoutputs.Itisusedtoshedlightontheunderlyingphysicalrelationships.Thereforetheinputs arerankedintermsofphysicalinfluence,usinganappropriaterankingindicator,presentedinthefollowingSubsection.
2.3.3. PhysicalinfluenceofinputsbasedonthePOD-PCEmodel
TheGWinfluencemeasurespresentedforthePCEmodelsinSubsection2.2arehereextendedforthePOD-PCEcoupling. These indicatorsare adequate for theanalysis ofeach PCE model
H
k,i.e.for interpreting thecontribution ofthe inputstoeachrandomvariable Ak separately.However,calculatingthecontributionstoeach Ak independentlyprecludesputting
them in perspective according to the importance of Ak in the final reconstructed model
H
that approximates U(
X ,
.)
. Hence,adaptedindicatorsshouldbecalculated.LetU
(
X ,
.)
betherandomspatiotemporal fieldapproximatedby thePOD-PCE ML,forpredictionfromtimetj totime tj+1 andletH
kbethePCEapproximationatdegree p(k)thatmapstherandomPODtemporalcoefficient Akfromasetofinput variables,usingtheexpansiononthemultivariate polynomialbasis
ζ
α(k)(.)
|α|≤p(k).ThePOD-PCEmodelformulated
inEquation(12) iswrittenasinEquation(19):
U
(
X , .) ≈
d k=1 Ak n× λ
kk(d)
(
X ) ≈
d k=1⎛
⎝
|α|≤p(k) c(αk)ζ
α(k)(.)
⎞
⎠
n× λ
k(kd)
(
X ) .
(19)Thankstoitslinearity,thePOD-PCEMLcanberepresentedasasingle-layeredNN,asshowninFig.3. Therefore,anewindicator,GeneralizedGarsonWeights (GGW),denotedWζ(k)
α ,iscomputedandsimplyre-evaluatedfrom thePCEGarsonweights(GW),heredenotedw
ζα(k),asinEquation(20). W ζα(k)
:=
|
c(αk)|
√
n× λ
k d e=1 |β|≤p(e)|
c(βe)|
√
n× λ
e=
|β|≤p(k)|
c(βk)|
w ζα(k)√
λ
k d e=1 |β|≤p(e)|
cβ(e)|
√
λ
e=
⎛
⎝
|β|≤p(k)|
c(βk)|
√
λ
k d e=1 |β|≤p(e)|
c(βe)|
√
λ
e⎞
⎠
w ζα(k).
(20)TheseGGWindicatorsshowthatthecontributionofthepolynomials
{ζ
α(k)}
|α|≤p(k)ofAkareenhancedwiththeeigenvalueλ
k,whichisdirectlylinkedto theimportanceofthePODmodek(d)
(
X )
(EVR inEquation(2)).Ananalogycanbedrawnwiththegeneralizedsensitivityindicesforareducedordermodel[48].The
dk=1|α|≤p(k)Wζα(k)
=
1 propertyholds.This means that the indices allow{{ζ
α(k)}
|α|≤p(k)}
k∈{1,...,d} to be ranked altogether in terms of contribution to output U. TheFig. 3. Representation of the POD-PCE ML approach in the NN paradigm.
3. Applicationonaparametrictoyproblem
The theoretical framework ofthe proposed POD-PCE learning was presented in the previous Section 2, including the detailedcouplingformulation, accuracyestimatorsandphysicalinfluencemeasures inSubsection2.3.Inthe latter,itwas highlightedthatthereisaslightdifferenceinthelearningandpredictionstepsbetweentemporalproblemsandparametric problems. In this section, the POD-PCE ML is applied to a parametric toy problem, for which the analytical solution is introduced inSubsection3.1.Theproblemissimpleandcontrollable,andallowsdemonstratingthelearning performance, theconsistencyofphysicalinterpretationsincomparisonwiththeanalyticalinformation,andtherobustnessofthelearning tonoiseinthedata.Subsection3.2thereforedealswiththeapplicationofthePOD-PCEmethodologyforphysicalanalysis andprediction,whileinSubsection3.3,robustnesstodifferentnoiselevelsisinvestigated.
3.1. Problemdescription
The chosen toy problem deals with the representation of groundwater flow in a confined aquifer. Such a flow can be complex todescribe andisgenerally representedusingthe depth-averaged groundwaterflow equations[34]. Analyti-cal solutions for theseequationscan be found forparticular configurations. Forexample,a solution was identified by Li et al. [34] incaseofa semi-infinitecoastalaquifersubjectto oscillatingboundary conditions,resultingfromoceanicand estuarine tidal loadings. The solution is givenforthe particular casewhere the estuary andcoastline are perpendicular. The oceanicBC(alongcoastline) istakenasasingle andspatiallyuniformtidalharmonicconstituent Acos
(
ω
t)
,where Aand
ω
arethe tidalamplitudeandpulsation respectively.The corresponding BC alongthe estuaryis anon-uniformtidal loading Aexp(
−
κ
erx)
cos(
ω
t−
κ
eix)
, whereκ
er andκ
ei are theestuary’s tidaldamping coefficientandtidal wave numberrespectively,thatrepresentchangesintheamplitudeandphasealongtheestuary.
Thisforcing resultswithfluctuationsinthewatertable,that isdefinedasthelevelseparatingthewaterandsaturated groundfromtheremainingupperunsaturated ground.The fluctuations,denoted f ,can becalculatedusingtheanalytical solutiondefinedin[34] asinEquation(21).
⎧
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎩
f(x,
y,t)
=
f0(x,
t)+
f1(x,
y,t)
f0(x,t)
=
Aexp−
ω 2Dx cosω
t−
ω 2Dx f1(x,
y,t)=
A×
Re t 0[g(k
1,
x)−
g(k1,
−
x)−
g(k2,
x)+
g(k2,
−
x)] dt0 (21)whereconstant D isthediffusivityoftheaquifer[34],t isthetime variable,and