* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Converge in probability and almost surely
Survey
Document related concepts
Transcript
Converge in probability and
almost surely
Definition: A random sample
The random variables X1 , · · · , Xn are called a random sample
of size n from population f (x) if X1 , · · · , Xn are mutually
independent and each Xi has the same distribution f (x).
Usually X1 , · · · , Xn are called independent and identically
distributed (iid) random variables.
The joint pdf or pmf of X1 , · · · , Xn is given by
f (x1 , · · · , xn ) =
n
Y
i=1
f (xi ).
Example
Suppose X1 , · · · , Xn are observed failure time of n bulbs. We
might assume X1 , · · · , Xn be a random sample from an
exponential(λ) population, where λ is unknown. The joint pdf of
the sample is
f (x1 , x2 , · · · , xn ) = λ
−n
exp(−
n
X
i=1
xi /λ).
Definition: Statistic
I
Let X1 , · · · , Xn be a random sample of size n from a
population and let T (x1 , · · · , xn ) be a real-valued or
vector-valued function. The random variable
Y = T (X1 , · · · , Xn ) is called a statistic, which does not
depend on any unknown parameter. The probability
distribution of a statistic Y is called sampling distribution of
Y.
I
Examples: Let X1 , · · · , Xn be a random sample.
P
1 Pn
¯
X̄n = n1 ni=1 Xi and Sn2 = n−1
i=1 (Xi − Xn ) are the
sample mean and sample variance respectively.
Converge in probability
Definition: Let X1 , · · · , Xn be a sequence of random variables
in probability space (S, F , P). The sequence {Xn } is said to
converge in probability to a random variable X if for any > 0,
lim P(|Xn − X | > ) = 0.
n→∞
Example
Let X1 , · · · , Xn be a sequence of random variables from
Unif(0,1) distribution. Let Mn = max{X1 , · · · , Xn }. Show that Mn
converge to 1 in probability.
Weak law of large numbers
Let X1 , · · · , Xn be iid random variables with mean E(Xi ) = µ
P
and variance Var(Xi ) = σ 2 < ∞. Define X̄n = n1 ni=1 Xi . Then,
for every > 0,
lim P(|X̄n − µ| > ) = 0.
n→∞
That is X̄n converges in probability to µ.
Example: Monte Carlo integration
Suppose we want to evaluate the integral
1
Z
I(h) =
h(x)dx
0
for a complicated function h. If the integration exists and but
hard to calculate, we can use the following approximation:
generating a large number of iid random variables
U1 , U2 , · · · , Un from Unif(0,1) and approximate I(h) by
n
Î(h) =
1X
h(Ui ).
n
i=1
Convergence of function of random variables
Suppose that X1 , X2 , · · · , converges in probability to a random
variable X and h is a continuous function. Then
h(X1 ), h(X2 ), · · · converges in probability to h(X ).
Example
Let Xn ∼ Binomial(n, p) and p̂n = Xn /n. (a) Show that p̂n → p in
probability. (b) Does p̂n2 converge to p2 in probability?
Almost surely convergence
Definition: Let X1 , · · · , Xn be a sequence of random variables
in probability space (S, F , P). The sequence {Xn } is said to
converge almost surely to a random variable X if
P({s : lim Xn (s) = X (s)}) = 1.
n→∞
Example
Let the sample space S be the closed interval [0,1] and P be
the uniform probability measure on [0, 1]. Define Xn (s) = s + sn
and X (s) = s. Does Xn (s) converge to X (s) almost surely?
Example: Converge in probability, but not almost
surely
Let Xn be a sequence of random variables on ([0, 1], F , P) and
P be the uniform probability measure on [0, 1]. Define
Xn (s) = IAn (s)
m
m
where An = [ 2km , k+1
2m ], n = k + 2 , k = 0, · · · , 2 − 1 and
m = 0, 1, 2, · · · . Show that Xn (s) converge to 0 in probability
but not almost surely.
Strong law of large numbers
Let X1 , · · · , Xn be iid random variables with mean E(Xi ) = µ
P
and variance Var(Xi ) = σ 2 < ∞. Define X̄n = n1 ni=1 Xi . Then,
for every > 0,
P( lim |X̄n − µ| > ) = 0.
n→∞
That is X̄n converges almost surely to µ.