Convergence in distribution #
We introduce a definition of convergence in distribution of random variables: this is the
weak convergence of the laws of the random variables. In Mathlib terms this is a Tendsto in the
ProbabilityMeasure type.
The definition assumes that the random variables are defined on the same probability space, which
is the most common setting in applications. Convergence in distribution for random variables
on different probability spaces can be talked about using the ProbabilityMeasure type directly.
We also state results relating convergence in probability (TendstoInMeasure)
and convergence in distribution.
Main definitions #
TendstoInDistribution X l Z μ: the sequence of random variablesX nconverges in distribution to the random variableZalong the filterlwith respect to the probability measureμ.
Main statements #
TendstoInDistribution.continuous_comp: Continuous mapping theorem. IfX ntends toZin distribution andgis continuous, theng ∘ X ntends tog ∘ Zin distribution.tendstoInDistribution_of_tendstoInMeasure_sub: the main technical tool for the next results. LetX, Ybe two sequences of measurable functions such thatX nconverges in distribution toZ, andY n - X nconverges in probability to0. ThenY nconverges in distribution toZ.TendstoInMeasure.tendstoInDistribution: convergence in probability implies convergence in distribution.TendstoInDistribution.prodMk_of_tendstoInMeasure_const: Slutsky's theorem. IfX nconverges in distribution toZ, andY nconverges in probability to a constantc, then the pair(X n, Y n)converges in distribution to(Z, c).
Convergence in distribution of random variables.
This is the weak convergence of the laws of the random variables: Tendsto in the
ProbabilityMeasure type.
- forall_aemeasurable (i : ι) : AEMeasurable (X i) μ
- aemeasurable_limit : AEMeasurable Z μ
- tendsto : Filter.Tendsto (fun (n : ι) => ⟨Measure.map (X n) μ, ⋯⟩) l (nhds ⟨Measure.map Z μ, ⋯⟩)
Instances For
Continuous mapping theorem: if X n tends to Z in distribution and g is continuous,
then g ∘ X n tends to g ∘ Z in distribution.
Let X, Y be two sequences of measurable functions such that X n converges in distribution
to Z, and Y n - X n converges in probability to 0.
Then Y n converges in distribution to Z.
Convergence in probability (TendstoInMeasure) implies convergence in distribution
(TendstoInDistribution).
Convergence in probability (TendstoInMeasure) implies convergence in distribution
(TendstoInDistribution).
Slutsky's theorem: if X n converges in distribution to Z, and Y n converges in
probability to a constant c, then the pair (X n, Y n) converges in distribution to (Z, c).
Slutsky's theorem for a continuous function: if X n converges in distribution to Z,
Y n converges in probability to a constant c, and g is a continuous function, then
g (X n, Y n) converges in distribution to g (Z, c).