Final examen in Asymptotic Statistics Preparation 1 hour
Manuscript notes and handout of the lectures are allowed
1 Cheap Bin
Let Z be a random variable with binomial distribution of parameters M ∈ N
∗and 0 < π < 1 (Z ∼ B(M, π)). We recall the following formulas
E (Z ) = M π,
E (Z
2) = M π(1 + (M − 1)π),
E (Z
3) = M π(1 − 3π + 3M π + 2π
2− 3M π
2+ M
2π
2),
E (Z
4) = M π(1 − 7π + 7M π + 12π
2− 18M π
2+ 6M
2π
2− 6π
3+ 11M π
3− 6M
2π
3+ M
3π
3).
For 0 < π
∗< 1, let X
1, . . . , X
nbe an i.i.d. sample with common law B(1, √
π
∗). To estimate the parameter π
∗, we consider the maximum likelihood estimator b π.
1. Show that b π = X
n2. Here, as usual, X
ndenotes the empirical mean built on the sample X
1, . . . , X
n. Compute the two first moments of b π. Compute its mean square error :
R
πb(π
∗) := E [( π b − π
∗)
2].
2. Modify b π to build an unbiased estimator b
b π. Compute R
bb π