Convergence in probability

The first stochastic convergence concept that we illustrate is not the strongest one, but it can be easier to grasp.

DEFINITION 9.7 (Convergence in probability) A sequence of random variables, X1X2,…, converges in probability to a random variable X if for every images

images

The definition may look intimidating, but it is actually intuitive: Xn tends to X if, for an arbitrarily small images, the probability of the event images goes to zero. In other words

images

i.e., the difference between the two realizations cannot be larger than a small and arbitrary constant images. Convergence in probability is denoted as follows:

images

or

images

We may also apply the definition to express convergence to a specific number a, rather than a random variable.

Example 9.29 Consider the following sequence of random variables:

images

We see that, for n going to infinity, Xn may take a larger and larger value n, but its probability is vanishing. Indeed, the probability mass gets concentrated on the value 0 and

images

Using this concept, we may clearly state (without proof) one form of the law of large numbers.

THEOREM 9.8 (Weak law of large numbers) Let X1X2,… be a sequence of i.i.d. random variables, with E[Xi] = μ and finite variance Var(Xi) = σ2 < +∞. Then

images

As you may imagine, there must be a different, strong form of this law; as we shall see, it uses a stronger form of stochastic convergence. Despite its relative weakness, convergence in probability is relevant in the theory of estimation, in that it allows us to define an important property of an estimator. Consider a sequence of estimators Yn of an unknown parameter θ; each Yn is a statistic depending on a sample of size n. We say that this is a consistent estimator if

images

The weak law of large numbers implies that the sample mean is a consistent estimator of expected value.


Comments

Leave a Reply

Your email address will not be published. Required fields are marked *