• Aucun résultat trouvé

Density estimation by Monte Carlo and randomized quasi-Monte Carlo

N/A
N/A
Protected

Academic year: 2022

Partager "Density estimation by Monte Carlo and randomized quasi-Monte Carlo"

Copied!
54
0
0

Texte intégral

(1)1. aft. Density estimation by Monte Carlo and randomized quasi-Monte Carlo Pierre L’Ecuyer. Dr. Joint work with Amal Ben Abdellah, Art B. Owen, and Florian Puchhammer. Séminaire au GERAD, Montréal, 27 mars 2019.

(2) 2. What this talk is about. aft. Quasi-Monte Carlo (QMC) and randomized QMC (RQMC) methods have been studied extensively for estimating an integral, say E[X ], and for approximating a function from its evaluation at a finite number of points.. Dr. How can we use them to estimate the entire distribution of X ? Here we will focus on estimating the density of X over [a, b] ⊂ R..

(3) 2. What this talk is about. aft. Quasi-Monte Carlo (QMC) and randomized QMC (RQMC) methods have been studied extensively for estimating an integral, say E[X ], and for approximating a function from its evaluation at a finite number of points. How can we use them to estimate the entire distribution of X ? Here we will focus on estimating the density of X over [a, b] ⊂ R.. Dr. People often look at empirical distributions via histograms, for example. More refined methods: kernel density estimators (KDEs). Can RQMC improve such density estimators, and by how much? Are there other types of density estimators than KDEs, that work better with RQMC?.

(4) 2. What this talk is about. aft. Quasi-Monte Carlo (QMC) and randomized QMC (RQMC) methods have been studied extensively for estimating an integral, say E[X ], and for approximating a function from its evaluation at a finite number of points. How can we use them to estimate the entire distribution of X ? Here we will focus on estimating the density of X over [a, b] ⊂ R.. Dr. People often look at empirical distributions via histograms, for example. More refined methods: kernel density estimators (KDEs). Can RQMC improve such density estimators, and by how much? Are there other types of density estimators than KDEs, that work better with RQMC? We will discuss an alternative that takes the sample derivative of a smoothed estimator of the cumulative distribution function (cdf). The smoothing can be achieved via conditional Monte Carlo, for example..

(5) 3. Small example: A stochastic activity network. aft. Gives precedence relations between activities. Activity k has random duration Yk (also length of arc k) with known cdf Fk (y ) := P[Yk ≤ y ]. Project duration X = (random) length of longest path from source to sink. Want to estimate the density of X , d P[X ≤ x]. f (x) = F 0 (x) = dx. 5. Dr. Y5. Y1. 0 Y0 source. 2. sink Y10. 8. Y9. Y4. 4. Y2. Y12 Y8. 7. Y7 1. Y3. 3. Y11 Y6. 6.

(6) 3. Small example: A stochastic activity network. aft. Gives precedence relations between activities. Activity k has random duration Yk (also length of arc k) with known cdf Fk (y ) := P[Yk ≤ y ]. Project duration X = (random) length of longest path from source to sink. Want to estimate the density of X , d P[X ≤ x]. f (x) = F 0 (x) = dx. Y5. Dr. The sample cdf P F̂n (x) = n1 ni=1 I[Xi ≤ x] is an unbiased estimator of the cdf F (x) = P[X ≤ x].. 5. Y1. 0 Y0 source. 2. sink Y10. 8. Y9. Y4. 4. Y2. Y12 Y8. 7. Y7 1. Y3. 3. Y11 Y6. 6.

(7) 3. Small example: A stochastic activity network. aft. Gives precedence relations between activities. Activity k has random duration Yk (also length of arc k) with known cdf Fk (y ) := P[Yk ≤ y ]. Project duration X = (random) length of longest path from source to sink. Want to estimate the density of X , d P[X ≤ x]. f (x) = F 0 (x) = dx. F̂n0 (x). Y5. Dr. The sample cdf P F̂n (x) = n1 ni=1 I[Xi ≤ x] is an unbiased estimator of the cdf F (x) = P[X ≤ x].. 5. Y1. But its derivative is not a meaningful estimator of f (x), because it is 0 0 Y0 almost everywhere. source. 2. sink Y10. 8. Y9. Y4. 4. Y2. Y12 Y8. 7. Y7 1. Y3. 3. Y11 Y6. 6.

(8) 4. Numerical illustration from Elmaghraby (1977): N(µk , σk2 ). aft. Yk ∼ for k = 0, 1, 3, 10, 11, and Yk ∼ Expon(1/µk ) otherwise. µ0 , . . . , µ12 : 13.0, 5.5, 7.0, 5.2, 16.5, 14.7, 10.3, 6.0, 4.0, 20.0, 3.2, 3.2, 16.5. Results of an experiment with n = 100 000. Frequency. mean = 64.2 Xdet = 48.2. 5000. Dr. 10000. ξˆ0.99 = 131.8. 0 0. 25. 50. 75. 100. 125. T 150. 175. 200.

(9) 5. aft. Setting Classical density estimation was developed in the context where independent observations X1 , . . . , Xn of X are given and one wishes to estimate the density f of X from that. Here we assume that X1 , . . . , Xn are generated by simulation from a stochastic model. We can choose n and we have some freedom on how the simulation is performed.. Dr. The Xi ’s are realizations of a random variable X = g (U) ∈ R with density f , where U = (U1 , . . . , Us ) ∼ U(0, 1)s and g (u) can be computed easily for any u ∈ (0, 1)s . Can we obtain a better estimate of f with RQMC instead of MC? How much better? Whats about taking a stratified sample over [0, 1)s ?.

(10) 6. aft. Density Estimation Suppose we estimate the density f over a finite interval [a, b]. Let fˆn (x) denote the density estimator at x, with sample size n. We use the following measures of error:. Z. MISE = mean integrated squared error =. b. E[(fˆn (x) − f (x))2 ]dx. a. = IV + ISB. b. Dr. Z. Var[fˆn (x)]dx a Z b (E[fˆn (x)] − f (x))2 dx ISB = integrated squared bias = IV = integrated variance =. a.

(11) 7. aft. Density Estimation Simple histogram: Partition [a, b] in m intervals of size h = (b − a)/m and define nj for x ∈ Ij = [a + (j − 1)h, a + jh), j = 1, ..., m fˆn (x) = nh where nj is the number of observations Xi that fall in interval j.. Dr. Kernel Density Estimator (KDE) : Select kernel k (unimodal symmetric density centered at 0) and bandwidth h > 0 (horizontal stretching factor for the kernel). The KDE is n. 1 X fˆn (x) = k nh i=1. . x − Xi h.  ..

(12) 8. KDE bandwidth selection: an illustration in s = 1 dimension lattice + shift (left),. aft. KDE (blue) vs true density (red) with RQMC point sets with n = 219 : Stratified sample of U = F (X ) (right). 80 70 60 50 40 30 20 10 0 -5.1. (10−5 ). -4.6. 80 70 60 50 40 30 20 10 0 -5.1. Dr. (10−5 ). -4.1. -3.6. -4.6. -4.1. -3.6.

(13) 9. For any g : R → R, define Z R(g ) =. aft. Asymptotic convergence with Monte Carlo for smooth f b. (g (x))2 dx,. Za ∞ µr (g ) =. x r g (x)dx,. for r = 0, 1, 2, . . .. −∞. Dr. For histograms and KDEs, when n → ∞ and h → 0:. AMISE = AIV + AISB ∼. Histogram KDE. C + Bhα . nh. C 1. B R(f 0 ) /12. α 2. µ0 (k 2 ). (µ2 (k))2 R(f 00 ) /4. 4.

(14) 10. The asymptotically optimal h is h = and it gives AMISE = Kn−α/(1+α) . C. KDE. B R(f 0 ). 1 µ0. C Bαn. 1/(α+1). 12 (µ2 (k))2 R(f 00 ). α. h∗. AMISE. 2. (nR(f 0 )/6)−1/3. O(n−2/3 ). Dr. Histogram. . aft ∗. (k 2 ). 4. . 4. µ0 (k 2 ) (µ2 (k))2 R(f 00 )n. 1/5. O(n−4/5 ). To estimate h∗ , one can estimate R(f 0 ) and R(f 00 ) via KDE (plugin). This is under the simplifying assumption that h must be the same all over [a, b]..

(15) 11. Elementary quasi-Monte Carlo (QMC) Bounds (Recall). aft. Integration error for g : [0, 1)s → R with point set Pn = {u0 , . . . , un−1 } ⊂ [0, 1)s : n−1. 1X En = g (ui ) − n i=0. Z. g (u)du.. [0,1)s. Koksma-Hlawka inequality: |En | ≤ VHK (g )D ∗ (Pn ) where VHK (g ) =. X Z. D ∗ (Pn ) =. (Hardy-Krause (HK) variation). Dr. ∅6=v⊆S. [0,1)s. ∂ |v| g (u) du, ∂v. sup. u∈[0,1)s. vol[0, u) −. |Pn ∩ [0, u)| n. (star-discrepancy).. There are explicit point sets for which D ∗ (Pn ) = O((log n)s−1 /n) = O(n−1+ ), ∀ > 0. Explicit RQMC constructions for which E[En ] = 0 and Var[En ] = O(n−2+ ), ∀ > 0. With ordinary Monte Carlo (MC), one has Var[En ] = O(n−1 )..

(16) 12. Asymptotic convergence of KDE with RQMC RQMC does not change the bias.. aft. Idea: Replace U1 , . . . , Un by RQMC points. For a KDE with smooth k, one could hope (perhaps) to get AIV = C 0 n−β h−1. for β > 1, instead of C n−1 h−1 .. Dr. If the IV is reduced, the optimal h can be taken smaller to reduce the ISB as well (re-balance) and then reduce the MISE..

(17) 12. Asymptotic convergence of KDE with RQMC RQMC does not change the bias.. aft. Idea: Replace U1 , . . . , Un by RQMC points. For a KDE with smooth k, one could hope (perhaps) to get AIV = C 0 n−β h−1. for β > 1, instead of C n−1 h−1 .. If the IV is reduced, the optimal h can be taken smaller to reduce the ISB as well (re-balance) and then reduce the MISE.. Dr. Unfortunately, things are not so simple.. Roughly, decreasing h increases the variation of the function in the KDE estimator. So we rather have something like AIV = C 0 n−β h−δ. or IV ≈ C 0 n−β h−δ in some bounded region, for β > 1 and δ > 1..

(18) 13. Bounding the AIV under RQMC for a KDE. aft. KDE density estimator at a single point x:   n n x − g (Ui ) 1X 1X 1 ˆ k = g̃ (Ui ). fn (x) = n h h n i=1 i=1 R With RQMC points Ui , this is an RQMC estimator of E[g̃ (U)] = [0,1)s g̃ (u)du = E[fˆn (x)]. RQMC does not change the bias, but may reduce Var[fˆn (x)], and then the IV.. Dr. To get RQMC variance bounds, we need bounds on the variation of g̃ ..

(19) 13. Bounding the AIV under RQMC for a KDE. aft. KDE density estimator at a single point x:   n n x − g (Ui ) 1X 1X 1 ˆ k = g̃ (Ui ). fn (x) = n h h n i=1 i=1 R With RQMC points Ui , this is an RQMC estimator of E[g̃ (U)] = [0,1)s g̃ (u)du = E[fˆn (x)]. RQMC does not change the bias, but may reduce Var[fˆn (x)], and then the IV. To get RQMC variance bounds, we need bounds on the variation of g̃ .. Dr. The partial derivatives are:. ∂ |v| 1 ∂ |v| g̃ (u) = k ∂uv h ∂uv. . x − g (u) h.  .. We assume they exist and are uniformly bounded. E.g., Gaussian kernel k. By expanding via the chain rule, we obtain terms in h−j for j = Q2, . . . , |v| + 1. One of the term for v = S grows as h−s−1 k (s) ((g (u) − x)/h) sj=1 gj (u) = O(h−s−1 ) when h → 0, so this AIV bound grows in h as h−2s−2 . Not so good!.

(20) 14. aft. Improvement by a Change of Variable, in One Dimension. Suppose g : [0, 1] → R is monotone. Change of variable w = (x − g (u))/h. In one dimension (s = 1), we have dw /du = −g 0 (u)/h, so VHK (g̃ ) =. 1 h. 1. Z. k0. 0. . x − g (u) h. . −g 0 (u) h. . du =. 1 h. Z. ∞. k 0 (w )dw = O(h−1 ).. −∞. Dr. Then, if k and g are continuously differentiable, with RQMC points having D ∗ (Pn ) = O(n−1+ ), we obtain AIV = O(n−2+ h−2 ). With h = Θ(n−1/3 ), this gives AMISE = O(n−4/3+ )..

(21) 15. Higher Dimensions. For each j ∈ S, let Gj =. Q. `∈S\{j}. g{`}. aft. Assumptions. Let g : [0, 1]s → R be piecewise monotone in each coordinate uj when the other coordinates are fixed, with at most Mj pieces. Assume that all first-order partial derivatives of g are continuous and that kgw1 gw2 · · · gw` k1 < ∞ for all selections of non-empty, mutually disjoint index sets w1 , . . . , w` ⊆ S = {1, . . . , s}. 1.  and cj = Mj kk (s) k∞ Gj + I(s = 2) kg{1,2} k1 .. Proposition Then the Hardy-Krause variation of g̃ satisfies. Dr. VHK (g̃ ) ≤ cj h−s + O(h−s+1 ). for each j.. Corollary. With RQMC point sets having D ∗ (Pn ) = O(n−1+ ) for all  > 0 when n → ∞, using KH and squaring gives the bound AIV = O(n−2+ h−2s ). for all  > 0.. By picking h to minimize the AMISE bound, we obtain AMISE = O(n−4/(2+s)+ ) . Worst than MC when s > 3. The factor h−2s hurts! But this is only an upper bound..

(22) 16. Stratification of the unit cube. aft. Q Partition [0, 1)s into n = b s congruent cubic cells Si := sj=1 [ij /b, (ij + 1)/b), i ∈ I = {i = (i1 , i2 , . . . , is ) : 0 ≤ ij < b for each j}, for some b ≥ 2. Construct Pn = {U1 , . . . , Un } by sampling one point uniformly in each subcube Si , independently, and put Xi = g (Ui ) for i = 1, . . . , n. Proposition. Suppose g is monotone. Then the KDE obtained from those points has. Dr. IV ≤ (b − a)s · k 2 (0) · h−2 n−(s+1)/s .. Corollary. By taking h = κn−(s+1)/(6s) , one has AMISE = O(n−(2/3)(s+1)/s ) . This gives a better rate than MC for s < 4. The factor h−2 hurts, but much less than h−2s ..

(23) 17. Empirical Evaluation with Linear Model in a limited region. aft. Regardless of the asymptotic bounds, the true IV and MISE may behave better than for MC for pairs (n, h) of interest. In a region of interest, we consider the model MISE = IV + ISB ≈ C n−β h−δ + Bhα . The optimal h for this model satisfies. C δ −β n . Bα. Dr. hα+δ =. and it gives MISE ≈ Kn−αβ/(α+δ) .. We can take the asymptotic α (known) and B (estimated as for MC). To estimate C , β, and δ, estimate the IV over a grid of values of (n, h), and fit a linear regression model: log IV ≈ log C − β log n − δ log h..

(24) Model estimation. aft. 18. For each (n, h), we estimate the IV by making nr = 100 indep. replications of the RQMC density estimator, compute the variance at ne = 1024 evaluation points (stratified) over [a, b], and multiply by (b − a)/n. We use logs in base 2, since n is a power of 2. Validation. Dr. After estimating model parameters, we can test out-of-sample with independent simulation experiments at pairs (n, h) with h = ĥ∗ (n). For test cases in which density is known, to assess what RQMC can achieve, we can compute a MISE estimate at those pairs (n, h), and obtain new parameter estimates K̃ and ν̃ of model MISE ≈ Kn−ν ..

(25) 19. RQMC Point sets:. aft. Numerical illustrations. I MC: Independent points (Crude Monte Carlo), I Stratification: stratified unit cube,. I LMS: Sobol’ points with left matrix scrambling (LMS) + digital random shift,. Dr. I NUS: Sobol’ points with nested uniform scrambling..

(26) 20. aft. Simple test example with standard normal density Let Z1 , . . . , Zs i.i.d. standard normal generated by inversion, and X = Then X ∼ N (0, 1).. Z1 + · · · + Zs √ . s. Dr. Here we can estimate R b 00IV, ISB, and MISE accurately. We can compute a f (x)dx exactly. We take [a, b] = [−2, 2]..

(27) 21. Estimates of model parameters for KDE. We have α = 4 and B = 0.04754. MC 1 0.999 1.017 1.144 3.758 0.770 16.96. MISE ≈ κn−ν. 1 0.999 2.787 2.997 3.798 1.600 34.05. Sobol + NUS 2 3 5 1.000 0.995 0.979 2.110 1.788 1.288 3.195 3.356 2.293 3.846 3.860 3.782 1.172 0.981 0.827 24.37 20.80 17.91. Dr. method s R2 β δ α ν̃ e19. IV ≈ Cn−β h−δ ,. aft. ISB = Bhα ,. For n = 219 , we have MISE ≈ 2−e19 .. 20 0.993 1.026 1.450 3.870 0.730 17.07.

(28) aft. 22. Convergence of the MISE, for s = 2, for histograms (left) and KDE (right). Independent points Stratification Sobol + LMS Sobol + NUS. log2 (MISE). −10. −15. −15. Dr. log2 (MISE). −10. Independent points Stratification Sobol + LMS Sobol + NUS. −20. −25. 10. 15 log2 (n). 20. 10. 15 log2 (n). 20.

(29) 22. Displacement of a cantilever beam (Bingham 2017). aft. Displacement D of a cantilever beam with horizontal load X and vertical load Y : r 4L3 Y 2 X 2 D= + 4 Ewt t 4 w where L = 100, w = 4, t = 2 (in inches), X , Y , and E are independent and normally distributed with means and standard deviations: Symbol E X Y. Mean 2.9 × 107 500 1000. Dr. Description Young’s modulus Horizontal load Vertical load. We want to estimate the density of X̃ = D/2.2535 − 1 over [a, b] = [0.336, 1.561] (about 99.5% of density).. St. dev. 1.45 × 106 100 100.

(30) 23. aft. Parameter estimates of the linear regression model for IV and MISE: IV ≈ Cn−β h−δ ,. MISE ≈ κn−ν. Dr. Ĉ β̂ δ̂ Point set KDE with Gaussian kernel, α = 4 Independent 0.210 0.993 1.037 Sobol+LMS 5.28E-4 1.619 2.949 Sobol+NUS 5.24E-4 1.621 2.955. ν̂. 0.789 0.932 0.932.

(31) aft. 24. log2 (IV) vs log2 n for cantilever with KDE, for fixed h. 0. Independent, Sobol+NUS, Independent, Sobol+NUS, Independent, Sobol+NUS,. −20. −30 10. Dr. log2 (IV). −10. h = 2−9.5 h = 2−9.5 h = 2−7.5 h = 2−7.5 h = 2−5.5 h = 2−5.5. 15 log2 (n). 20.

(32) 24. A weighted sum of lognormals, in s = 12 dimensions s X. aft. X =. wj exp(Yj ). j=1. where Y = (Y1 , . . . , Ys. )t. ∼ N (µ, C).. Let C = AAt . To generate Y, generate Z ∼ N (0, I) and put Y = µ + AZ. We will use principal component decomposition (PCA).. Dr. This has several applications. In one of them, with wj = s0 (s − j + 1)/s, e −ρ max(X − K , 0) is the payoff of a financial option based on an average price at s observation times, under a GBM process. Want to estimate density of positive payoffs..

(33) 24. A weighted sum of lognormals, in s = 12 dimensions s X. aft. X =. wj exp(Yj ). j=1. where Y = (Y1 , . . . , Ys. )t. ∼ N (µ, C).. Let C = AAt . To generate Y, generate Z ∼ N (0, I) and put Y = µ + AZ. We will use principal component decomposition (PCA).. Dr. This has several applications. In one of them, with wj = s0 (s − j + 1)/s, e −ρ max(X − K , 0) is the payoff of a financial option based on an average price at s observation times, under a GBM process. Want to estimate density of positive payoffs. Numerical experiment: Take s = 12, ρ = 0.037, s0 = 100, K = 101, and C defined indirectly via: Yj = Yj−1 (µ − σ 2 )j/s + σB(j/s) where Y0 = 0, σ = 0.12136, µ = 0.1, and B(·) a standard Brownian motion. We will estimate the density of e −ρ (X − K ) over [a, b] = [0, 50]..

(34) 25. Frequency (×103 ). 30. 10 0 0. Dr. 20. aft. Histogram of positive values from n = 106 independent simulation runs:. 13.1. 50. Payoff 100. 150.

(35) IV as a function of n for fixed h, for KDE.. aft. Independent, h = 2−3,5 Sobol+NUS, h = 2−3,5 Independent, h = 2−1,5 Sobol+NUS, h = 2−1,5 Independent, h = 20,5 Sobol+NUS, h = 20,5. −20. −30. −40 14. 15. Dr. log2 (IV). 26. 16. 17. 18. 19. 20. log2 (n). The RQMC bound in s = 12 dimension gives a worst rate than MC, but we observe a better actual IV and MISE..

(36) 26. aft. Alternative approach: Can we take the stochastic derivative of an estimator of F ? Can we estimate the density f (x) = F 0 (x) by the derivative w.r.t. x of an estimator of F (x). A simple candidate cdf estimator is the empirical cdf n. 1X F̂n (x) = I[Xi ≤ x]. n. Dr. i=1. However dF̂n (x)/dx = 0 almost everywhere, so this cannot be a useful density estimator! We need a smoother estimator of F ..

(37) 27. Conditional Monte Carlo (CMC) for Derivative Estimation. aft. Idea: Replace the indicator I[Xi ≤ x] by its conditional cdf given filtered (reduced) information G: F (x | G) := P[X ≤ x | G]. Dr. where the sigma-field G contains not enough information to reveal X but enough to compute F (x | G), and is chosen so that the following holds:.

(38) 27. Conditional Monte Carlo (CMC) for Derivative Estimation. aft. Idea: Replace the indicator I[Xi ≤ x] by its conditional cdf given filtered (reduced) information G: F (x | G) := P[X ≤ x | G] where the sigma-field G contains not enough information to reveal X but enough to compute F (x | G), and is chosen so that the following holds:. Dr. Assumption 1. For all realizations of G, F (x | G) is a continuous function of x which is differentiable except perhaps over a denumerable set of points, and for which the derivative F 0 (x | G) = dF (x | G)/dx (when it exists), which is itself a random variable for each x, is bounded uniformly in x by some random variable Γ having finite variance. Theorem: F 0 (x | G) is an unbiased estimator of f (x), with variance bounded uniformly in x. P Overall estimator with n iid replicates: fˆcmc,n (x) = n1 ni=1 F 0 (x | Gi ). With this estimator, ISB = 0, so MISE = IV, which is rather easy to estimate. The idea was introduced by Asmussen (2018) for a sum of random variables..

(39) 28. aft. Application to the normalized sum of standard normals We had. X =. Z1 + Z2 + · · · + Zs √ s. Dr. where each Zj ∼ N (0, 1). For CMC, we can leave out Zs , so G = (Z1 , . . . , Zs−1 ) and √ F (x | G) = P[X ≤ x | Z1 , . . . , Zs−1 ] = P[Zs ≤ x s − (Z1 + Z2 + · · · + Zs−1 )] √ = Φ(x s − (Z1 + Z2 + · · · + Zs−1 )). The resulting density estimator is. √ √ F 0 (x | G) = φ(x s − (Z1 + Z2 + · · · + Zs−1 )) s. Assumption 1 is easily verified..

(40) CMC KDE. ν e19 ν e19. MC 1.019 21.36 0.798 17.01. aft. 29. s=3 Sobol+LMS 2.116 40.81 0.976 20.79. Sobol+NUS 2.094 40.65 0.975 20.80. MC 0.988 19.27 0.769 17.00. s = 20 Sobol+LMS Sobol+NUS 0.961 0.982 19.58 19.54 0.771 0.760 17.10 17.07. Dr. For n = 219 , we have MISE ≈ 2−e19 .. We see that CMC is always helping, with both MC and RQMC. CMC+RQMC brings a huge gain in 3 dimensions, much less in 20 dimensions..

(41) MISE convergence. s=3. s = 20. 30. MC+CMC MC+KDE LMS+CMC LMS+KDE. aft. MC+CMC MC+KDE LMS+CMC LMS+KDE. −14. −20. −16. −30. −40. Dr. −18. −20. 14. 16 log2 (n). 18. 14. 16 log2 (n). 18.

(42) 30. Cantilever beam. aft. r 4L3 Y 2 X 2 D= + 4 Ewt t 4 w where E , X , Y are normal r.v.’s. Want density over [a, b] = [0.407, 1.515]. For CMC, we can leave out E , i.e., take G = (X , Y ). Then, ". 3. Y2 t4. +. X2 w4. #. ≤ d | X,Y. Dr. 4L F (d | G) = P[D ≤ d | X , Y ] = P Ewt. r. . = 1 − Φ. 4L3 dwt. q. Taking the derivative w.r.t. d, we get . 0. F (d | G) = φ . 4L3 dwt. q. Y2 t4. +. σE. X2 w4. − µE. . ×. 4L3 d 2 wtσE. r. Y2 X2 + . t4 w4. Y2 t4. + σE. X2 w4. − µE.  ..

(43) Estimated MISE = Kn−ν .. aft. 31. For n = 219 , we have MISE ≈ 2−e19 .. CMC. MC 1.02 18.54 0.81 15.12. Strat 1.80 30.64 0.89 18.08. Dr. KDE. ν e19 ν e19. LMS 2.22 43.18 0.94 21.33. NUS 2.16 43.08 0.96 21.35.

(44) −10. aft. −15. −25 −30 −35 −40 −45 14. Dr. log2 (MISE). −20. 15. 16. 17. log2 (n). 18. 19. MC+CMC MC+KDE Strat+CMC Strat+KDE LMS+CMC LMS+KDE. 32.

(45) 32. Small example: A stochastic activity network. aft. Gives precedence relations between activities. Activity k has random duration Yk (also length of arc k) with known cdf Fk (y ) := P[Yk ≤ y ]. Project duration X = (random) length of longest path from source to sink.. 5. Y5. Dr. Want to estimate the density of X , d P[X ≤ x]. f (x) = F 0 (x) = dx. We saw that F̂n0 (x) is not a meaningful estimator of f (x).. Y1. 0 Y0 source. 2. sink Y10. 8. Y9. Y4. 4. Y2. Y12 Y8. 7. Y7 1. Y3. 3. Y11 Y6. 6.

(46) 33. Numerical illustration from Elmaghraby (1977): N(µk , σk2 ). aft. Yk ∼ for k = 0, 1, 3, 10, 11, and Yk ∼ Expon(1/µk ) otherwise. µ0 , . . . , µ12 : 13.0, 5.5, 7.0, 5.2, 16.5, 14.7, 10.3, 6.0, 4.0, 20.0, 3.2, 3.2, 16.5. Results of an experiment with n = 100 000. Frequency. mean = 64.2 Xdet = 48.2. 5000. Dr. 10000. ξˆ0.99 = 131.8. 0 0. 25. 50. 75. 100. 125. T 150. 175. 200.

(47) aft. 34. The SAN example, Sobol+NUS vs Independent points, KDE, summary for n = 219 = 524288.. h 0.10 0.18 0.32 0.43. Independent points log2 IV IV rate -16.64 -0.999 -17.96 -0.999 -19.33 -0.998 -19.99 -0.998. Dr. Density. Sobol+NUS log2 IV IV rate -16.71 -1.006 -18.18 -1.015 -19.79 -1.035 -20.71 -1.064.

(48) 35. CMC for the SAN Example. aft. Want to estimate the density of the longest path length X . CMC estimator of P[X ≤ x]: F (x | G) = P[X ≤ x | {Yj , j 6∈ L}]. where L = {4, 5, 6, 8, 9} and Yj = Fj−1 (Uj ). This estimator continuous in the Uj ’s and in x. 5. Y5. Y10. Y9. Dr 2. Y1. source. 0. Y0. 8. Y4. 4. Y2. Y12. Y8. 7. Y7. 1. Y3. 3. sink. Y11 Y6. 6.

(49) 36. aft. To compute F (t | G): for each l ∈ L, say from al to bl , compute the length αl of the longest path from 1 to al , and the length βl of the longest path from bl to the destination.. Dr. The longest path that passes through link l does not exceed t iff αl + Yl + βl ≤ x, which occurs with probability P[Yl ≤ x − αl − βl ] = Fl [x − αl − βl ]..

(50) 36. aft. To compute F (t | G): for each l ∈ L, say from al to bl , compute the length αl of the longest path from 1 to al , and the length βl of the longest path from bl to the destination. The longest path that passes through link l does not exceed t iff αl + Yl + βl ≤ x, which occurs with probability P[Yl ≤ x − αl − βl ] = Fl [x − αl − βl ]. Since the Yl are independent, we obtain. F (x | G) =. Y. Fl [x − αl − βl ].. l∈L. Dr. To estimate the density of X , take the derivative w.r.t. x: F 0 (x | G) =. Y d w.p.1 X F (x | G) = fj [x − αj − βj ] Fl [x − αl − βl ]. dx j∈L. l∈L, l6=j. Assumption 1 holds if the Fj are smooth enough, and then E[F 0 (x | G)] = fX (x)..

(51) 37. For n = 219 , MISE ≈ 2−e19 .. CMC. aft. Estimated MISE = Kn−ν , for KDE with CMC.. ν e19. MC 0.99 25.48. LMS 1.34 29.67. NUS 1.32 29.66. Dr. For comparison, the base case without CMC and RQMC gave e19 ≈ 20. With MC, the IV converges as O(1/n) and there is no bias, so MISE = IV. With RQMC, we observe a convergence rate near O(n−4/3 ) for the IV and the MISE..

(52) MC+CMC LMS+CMC NUS+CMC. aft. −20. −24. −26. −28. −30 14. Dr. log2 (MISE). −22. 15. 16. 17. log2 (n). 18. 19. 38.

(53) 38. aft. Conclusion. Dr. I Both CMC and RQMC can improve the convergence rate of the IV and MISE when estimating a density. I With KDEs, the convergence rates observed in small examples are much better than the bounds proved from standard QMC theory. There are opportunities for QMC theoreticians here! I The combination of CMC with RQMC for density estimation is very promising. Lots of potential applications. I Future: Density estimation for a function of the state of a Markov chain, using Array-RQMC. I More applications..

(54) Some references. 38. aft. I S. Asmussen. Conditional Monte Carlo for sums, with applications to insurance and finance, Annals of Actuarial Science, prepublication, 1–24, 2018. I A. Ben Abdellah, P. L’Ecuyer, A. B. Owen, and F. Puchhammer. Density estimation by Randomized Quasi-Monte Carlo. Submitted, 2018. I J. Dick and F. Pillichshammer. Digital Nets and Sequences: Discrepancy Theory and Quasi-Monte Carlo Integration. Cambridge University Press, Cambridge, U.K., 2010. I P. L’Ecuyer. A unified view of the IPA, SF, and LR gradient estimation techniques. Management Science 36: 1364–1383, 1990.. Dr. I P. L’Ecuyer. Quasi-Monte Carlo methods with applications in finance. Finance and Stochastics, 13(3):307–349, 2009. I P. L’Ecuyer. Randomized quasi-Monte Carlo: An introduction for practitioners. In P. W. Glynn and A. B. Owen, editors, Monte Carlo and Quasi-Monte Carlo Methods 2016, 2017. I P. L’Ecuyer and G. Perron. On the Convergence Rates of IPA and FDC Derivative Estimators for Finite-Horizon Stochastic Simulations. Operations Research, 42 (4):643–656, 1994. I A. B. Owen. Scrambled Net Variance for Integrals of Smooth Functions. Annals of Statistics, 25 (4):1541–1562, 1997. I D. W. Scott. Multivariate Density Estimation. Wiley, 2015..

(55)

Références

Documents relatifs

Randomized quasi-Monte Carlo (RQMC) methods replace the independent random numbers by dependent vectors of uniform random numbers that cover the space more evenly.. When estimating

With this scrambling just by itself, each point does not have the uniform distribution (e.g., the point 0 is unchanged), but one can apply a random digital shift in base b after

Keywords: random number generator, pseudorandom numbers, linear generator, multi- ple recursive generator, tests of uniformity, random variate generation, inversion,

The estimated variance, bias, MSE and fraction of the MSE due to the square bias on the likelihood function estimator with constructed lattices and other point sets are plotted

However, a new RQMC method specially designed for Markov chains, called array-RQMC (L’Ecuyer et al., 2007), is often very effective in this situation. The idea of this method is

In summary, for array-RQMC, we have the follow- ing types of d-dimensional RQMC point sets for P n : a d + 1-dimensional Korobov lattice rule with its first coordinate skipped,

When &amp; 2 − &amp; 1 becomes small, the parameters of the beta distribution decrease, and its (bimodal) density concentrates near the extreme values 0 and 1.. 82) say in

Like for RNGs, different discrepancy measures are often used in practice for different types of point set constructions, for computational efficiency considera- tions (e.g.. This