Download Special random variables

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
Special random variables
Chapter 5
Some discrete or continuous
probability distributions
Some special random variables






Bernoulli
Binomial
Poisson
Hypergeometric
Uniform
Normal and its derivatives





Chi-square
T-distribution
F-distribution
Exponential
Gamma
Bernoulli random variable

A random variable X is said to be a Bernoulli
random variable if its probability mass function
is given as the following:



P{X=0}=1-p, P{X=1}=p; (you may assume X=1
when the experimental outcome is successful and
X=0 when it is failed.)
E[X]=1×P{X=1}+0×P{X=0}=p
Var[X]=E[X2]-E[X]2 =p(1-p)
Binomial random variable


Suppose there are n independent Bernoulli trials,
each of which results in a “success” with the
probability p.
If X represents the number of successes that occur
in the n trials, then X is said to be a binomial
random variable with parameters (n,p).

P{X=i}= n!/[(n-i)! ×i!] × pn(1-p)(n-i) , i=0, 1, 2,…n
Cin
n
n
 p( X  i)   C ( p) (1  p)
i 0
i 0
n
i
i
n i
 [ p  (1  p)]  1
n
The expectation of binomial
random variable




The Binomial random X is composed of n
independent Bernoulli trials
∴X=Σ1~nxi, xi =1 when the ith trial is a
success or xi =0 otherwise
E[X]=Σ1~nE[xi]=n×p
Var[X]=Σ1~nVar[xi]=np(1-p)
Patterns of binomial
distribution




If p=0.5, then X will distribute
symmetrically
If p>0.5, then X will be a left-skewed
distribution
If p<0.5, then X will be a right-skewed
distribution
If n∞, then X will distribute as a
symmetric bell/normal pattern
Poisson random variable

The random variable X is a Poisson
distribution if its prob. mass function is
given by
P[ X  i]  e


i
i!
, i  0,1,2,...


i
i 0
i 0
i!

P
(
X

i
)

e


(By the Taylor series)
 e   e   e 0  1
Expectation and variance of
Poisson distribution

By using the moment generating function


i 0
i 0
 (t )  E[etX ]   e ti (e  i / i!) e   (e t  ) i / i!)

e e
e t
 exp{ (e t  1)}
 ' (t )  e t exp{ (e t  1)};
 " (t )  (e t ) 2 exp{ (e t  1)}  e t exp{ (e t  1)}
E[ X ]   ' (0)   ,
Var[ X ]   " (0)   ' (0) 2  2    2  
Poisson vs. Binomial


A binomial distribution, (n,p),  a Poisson with
meanλ=np, when n is large and p is small.
In other words, if the successful probability of trial is
very small, then the accumulative many trials
distribute as a Poisson.


The probability of one person living over 100 years of age
is 0.01, then the mean number of over-100-year-old
persons 10 may occur within 1000 elders.
λ means the average occurrence among a large number of
experiments, in contrast to, p means the occurring chance
of every trial
Comparisons
A binomial random variable with p near to 0.5
A Poisson random variable
approximates to a binomial
distribution when n becomes
large.
Hypergeometric random
variable


There are N+M objects, of which N are desirable and the
other M are defective. A sample size n is randomly
chosen from N+M without replacements.
Let X be the number of desirable within n chosen objects.
Its probability mass function as the following.
N M
i
n i
N M
n
C C
P[ X  i] 
C
, i  0,1,2,... min( N , n)
We said X is a hypergeometric distribution with
parameters (N,M,n)
Expectation & variance of
hypergeometric distribution
1, if the ith selection is desirable
Xi  
0, otherwise
n
X   X i , X is the number of desired outcome in the sample of size n
i 1
n
n
i 1
i 1
E[X]   E[ X i ]   P( X i  1)  nN
n
Var ( X )   Var ( X i )  2
i 1
(N  M )
n
 Cov( X , X
1i  j n
i
j
)
Expectation & variance of
hypergeometric distribution (cont.)
N 1
N

E[ X i X j ]  P( X i X j  1)  P( X i  1, X j  1) 
N  M N  M 1
N
N ( N  1)
) 2 , for i  j
(
Cov( X i , X j )  E[ X i X j ]  E[ X i ]E[ X j ] 
N M
( N  M )( N  M  1)
 NM

( N  M ) 2 ( N  M  1)
n 1
nMN
 NM
nMN
n
)

1
(

C
2

Var ( X ) 
2
2
2
2
N  M 1
( N  M ) ( N  M  1) ( N  M )
(N  M )
n 1 

if let p  N/(N  M), then E(X)  np, Var(X)  np(1 - p) 1 
 N  M  1
Moreover, if N+M increases to ∞, then Var(X) converges to np(1-p), which
is the variance of a binomial random variable with parameters (n,p).
hypergeometric vs. binomial

Let X, and Y be independent binomial
random variables having respective
parameters (n,p) and (m,p). The
conditional p.m.f. of X given that X+Y=k is
a hypergeometric distribution with
parameters (n,m,k).
The Uniform random variable

A random variable X is said to be uniformly
distributed over the interval [α,β] if its p.d.f.
is given by
 1
, if α  X  

f ( x)     
0,
otherwise

1

(    ) 
dx  1
Pattern of uniform distribution
Expectation and variance of
uniform distribution
E[X]=∫α~β x[1/(β-α)]dx=(α+β)/2

Var(X)=E[X2]-E[X]2
=(β-α)2/12

P.161, Example 5.4b

Normal random variable

A continuous r.v X has a normal distribution
with parameter μ and σ2 if its probability
density function is given by:
We write X~N(μ,σ2)
By using the M.G.F., we obtain E[X]=ψ’(0)=μ, and
Var[X]=ψ”(0)-ψ’(0)2=σ2
Standard normal distribution
The Cumulative Distribution function of Z
(a) 
1
2
P{ X  b}  P{

P{a  X  b}  P{
e
z2 / 2

X 

a

a

b
a
 (
)  (
)


b


dz, -   a  
}  (
X 


b

)

}
b
 (a )  P{Z  a}  P{Z  a}  1   (a ), (by symmetry)
Percentiles of the normal
distribution

Suppose that a test score distributes as a normal
distribution with mean 500 and standard
deviation of 100. What is the 75th percentile
score of this test?
Characteristics of Normal
distribution
P{∣X-μ∣<σ}=ψ(-1<X-μ/σ<1)=ψ(-1<Z<1)= ψ(1)-ψ(-1)=2ψ(1) -1=2× 0.8413-1
P{∣X-μ∣<2σ}=ψ(-2<X-μ/σ<2)=ψ(-2<Z<2)= ψ(2)-ψ(-2)=2ψ(2) -1=2× 0.9772-1
P{∣X-μ∣<3σ}=ψ(-3<X-μ/σ<3)=ψ(-3<Z<3)= ψ(3)-ψ(-3)=2ψ(3) -1=2× 0.9987-1
The pattern of normal
distribution
Exponential random variables
A nonnegative random variable X with a parameter λ obeying
the following pdf and cdf is called an exponential distribution.
x
e x , if x  0
f ( x)  
, F ( x)  P{ X  x}   e y dy  1  e x , x  0
0
0
,
if
x

0

• The exponential distribution is often used to describe the
distribution of the amount of time until some specific
event occurs.
• The amount of time until an earthquake occurs
• The amount of time until a telephone call you receive
turns to be the wrong number
Pattern of exponential
distribution
Expectation and variance of
exponential distribution

E[X]=ψ’(0)=1/λ



E[X] means the average cycle time, λ presents the occurring
frequency per time interval
Var(X)=ψ”(0)-ψ’(0)2=1/λ2
The memoryless property of X

P{X>s+t/X>t}=P{X>s}, if s, t≧0; P{X>s+t}=P{X>s} ×P{X>t}
P{ X  s  t , X  t} P{ X  s  t}
P{ X  s  t / X  t} 

P{ X  t}
P{ X  t}
1  F ( s  t ) 1  [1  e  ( s t ) ]
 s



1

[
1

e
]  1  F ( s)  P{ X  s}
 t
1  F (t )
1  [1  e ]
Poisson vs. exponential

Suppose that independent events are occurring at
random time points, and let N(t) denote the number
of event that occurs in the time interval [0,t]. These
events are said to constitute a Poisson process
having rate λ, λ>0, if




N(0)=0;
The distribution of number of events occurring within an
interval depends on the time length and not on the time
point.
lim h0 P{N(h)=1}/h=λ
lim h0 P{N(h)≧2}/h=0
Poisson vs. exponential (cont.)

P{N(t)=k}=P{k of the n subintervals contain exactly 1
event and the other n-k contain 0 events}
P{ecactly 1 event in a subinterval t/n}≒λ(t/n)
 P{0 events in a subinterval t/n}≒1-λ(t/n)
k
nk
n
P{N(t)=k} ≒   t  1  t  , a binomial
 
n  with p=λ(t/n)
distribution  k  n  



A binomial distribution approximates to a Poisson
distribution with k=n(λt/n) when n is large and p is
k
small.
(

t
)
 t
P[ N (t )  k ]  e
k!
, k  0,1,2,...
Poisson vs. exponential (cont.)

Let X1 is the time of first event.
P{X1>t}=P{N(t)=0} (0 events in the first t time
length)=exp(-λt)
 ∵F(t)=P{X1≦t}=1- exp(-λt) with mean 1/λ


∴X1 is an exponential random variable
Let Xn is the time elapsed between (n-1)st and
nth event. P{Xn>t/Xn-1=s}=P{N(t)=0} =exp(-λt)
 ∵F(t)=P{Xn≦t}=1- exp(-λt) with mean 1/λ

∴ Xn is also an exponential random variable
Gamma distribution

See the gamma definition and proof in p.182-183


If X1 and X2 are independent gamma random variables
having respective parameters (α1,λ) and (α2,λ), then
X1+X2 is a gamma random variable with (α1+α2,λ)
The gamma distribution with (1,λ) reduces to the
exponential with the rate λ.

If X1, X2, …Xn are independent exponential random
variables, each having rate λ, then X1+ X2+ …+Xn is a
gamma random variable with parameters (n,λ)
Expectation and variance of
Gamma distribution

The gamma distribution with (α,1) reduces
to the normal distribution when α becomes
large.


The patters of gamma distribution move from
right-skewed toward symmetric as α increases.
See p.183, using the G.M.F. to compute


E[X]=α/λ
Var(X)=α/(λ2 )
Patterns of gamma distribution
Derivatives from the normal
distribution



The chi-square distribution
The t distribution
The F distribution
The Chi-square distribution



If Z1, Z2,…Zn are independent standard normal
random variables, the X, defined by
X=Z12+Z22+…Zn2, is called chi-square distribution
with n degrees of freedom, and denoted by X~Xn2
If X1 and X2 are independent chi-square random
variables with n1 and n2 degrees of freedom,
respectively, then X1+X2 is chi-square with n1+n2
degrees of freedom.
p.d.f. of Chi-square

The probability density function for the
distribution with r degrees of freedom is
given by
The relation between chi-square
and gamma


A chi-square random variable with n degrees of
freedom is identical to a gamma random variable
with (n/2,1/2), i.e., α=n/2,λ=1/2
The expectation of chi-square distribution


The variance of chi-square distribution


E[X] is the same as the expectation of gamma
withα=n/2,λ=1/2, so E[X]=α/λ=n (degrees of
freedom)
Var(X)=α/(λ2 )=2n
The chi-square distribution moves from rightskewed toward symmetric as the degree of
freedom n increases.
Patterns of chi-square
distribution
The t distribution


If Z and Xn2 are independent random variables, with Z
having a std. normal dist. And Xn2 having a chi-square
dist. with n degrees of freedom,
2
2
2
2
Z

Z

...
Z
Z
X
then Tn defined by
2
n
Tn 
, n  1
n
X n2 / n n
• For large n, the t distribution approximates to the standard
normal distribution
• The t distribution move from flatter and having thicker tails
toward steeper and having thinner tails as n increases.
p.d.f. of t-distribution
B(p,q) is the beta function
Patterns of t distribution
The F distribution


If Xn2 and Xm2 are independent chi-square random
variables with n and m degrees of freedom, respectively,
2
then the random variable Fn,m defined by
Xn / n
Fn,m 
2
m
X /m
• Fn,m is said an F-distribution with n and m degrees of freedom
• F1,m is the same as the square of t-distribution, (Tm)²
p.d.f of F-distribution
where
is the gamma function, B(p,q) is the beta function,
Patterns of F-distribution
Relationship between different
random variables
n∞
p0
λ=np
Binomial (n, p)
Repeated
trials
Bernoulli (p)
The very small time partition
Poisson(λ)
Gamma(α, λ)
n∞,
p0.5
λ=1,
α∞
α=1
α=n/2
λ=1/2
Normal (μ,σ2)
Standardization
Z=(X-μ)/σ
Exponential(λ)
F distribution
Chi-square (n)
Z (0, 1)
t distribution
Homework #4

Problem 5,12,27,31,39,44
Related documents