Download Chapter 7 Point Estimation of Parameters

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
Chapter 7 Point Estimation of Parameters
7-1 INTRODUCTION
z
If X is a random variable with probability distribution f (x) ,
characterized by the unknown parameter θ , and if X 1 , X 2 ,..., X n is a
random sample of size n from X, the statistic Θˆ = h( X 1 , X 2 ,..., X ) is
called a point estimator of θ .
Definition
A point estimate of some population parameter θ is a single numerical
value θ̂ of a statistic Θ̂ . The statistic Θ̂ is called the point estimator.
z
z
z
z
As an example, suppose that the random variable X is normally
distributed with an unknown mean µ .
The sample mean is a point estimator of the unknown population mean
µ . That is, µ̂ = X .
After the sample has been selected, the numerical value x is the point
estimate of µ .
Thus, if x1 = 25 , x2 = 30 , xc = 29 , and x4 = 31 , the point estimate of µ
is
25 + 30 + 29 + 31
= 28.75
4
Similarly, if the population variance σ 2 is also unknown, a point
estimator for σ 2 is the sample variance S 2 .
x=
z
z
z
We often need to estimate
z The mean µ of a single population.
2
z The variance σ
(or standard deviation σ ) of a single population.
z The proportion p of items in a population that belong to a class of
interest.
z The difference in means of two populations, µ1 − µ 2 .
z The difference in two population proportions, p1 − p 2 .
Reasonable point estimates of these parameters are as follows:
z For µ , the estimate is µ̂ = x , the sample mean.
2
ˆ 2 = s 2 , the sample variance.
z For σ , the estimate is σ
1
z
z
z
For p, the estimate is pˆ = x / n , the sample proportion, where x is the
number of items in a random sample of size n that belong to the
class of interest.
For µ1 − µ 2 , the estimate is µˆ 1 − µˆ 2 = x1 − x2 , the difference between
the sample means of two independent random samples.
For p1 − p2 , the estimate is pˆ 1 − pˆ 2 , the difference between two
sample proportions computed from two independent random
samples.
7-2 GENERAL CONCEPTS OF POINT ESTIMATION
7-2.1 Unbiased Estimators
Definition
The point estimator Θ̂ is an unbiased estimator for the parameter θ if
ˆ = θ (7-1)
E (Θ)
If the estimator is not unbiased, then the difference
ˆ − θ (7-2)
E (Θ)
is called the bias of the estimator Θ̂ .
EXAMPLE 7-1
2
z X is a random variable with mean µ and variance σ .
z Let X 1 , X 2 ,..., X n be a random sample of size n from the population
represented by X.
X and sample variance S 2 are
z Show that the sample mean
unbiased estimators of µ and σ 2 , respectively.
z First consider the sample mean. In Equation 5.40a in Chapter 5, we
showed that E (X ) = µ .
z Therefore, the sample mean X is an unbiased estimator of the
population mean µ .
2
z
Now consider the sample variance. We have
⎡ n
2⎤
⎢∑(Xi − X ) ⎥
n
1
⎥=
E ( S 2 ) = E ⎢ i =1
E∑ ( X i − X )2
n −1
⎥ n − 1 i =1
⎢
⎥⎦
⎢⎣
n
n
n
⎤
1
1 ⎡ n 2
2
+
−
E ∑ ( X i2 + X 2 − 2 XX i ) =
E
(
X
)
E
(
X
)
2
E
(
Xi )X ⎥
∑
∑
∑
i
⎢
n − 1 i =1
n − 1 ⎣ i =1
i =1
i =1
⎦
n
n
⎤
⎤
1 ⎡
1 ⎡
=
E (∑ X i2 ) + nX 2 − 2nX ⋅ X ⎥ =
E (∑ X i2 ) − nX 2 ⎥
⎢
⎢
n − 1 ⎣ i =1
⎦ n − 1 ⎣ i =1
⎦
=
=
z
z
n
⎤
1
1 ⎡n
E (∑ X i2 − nX 2 ) =
E ( X i2 ) − nE ( X 2 )⎥
∑
⎢
n − 1 i =1
n − 1 ⎣ i =1
⎦
The last equality follows from Equation 5-37 in Chapter 5.
Since E ( X i2 ) = µ 2 + σ 2 and E ( X 2 ) = µ 2 + σ 2 / n (See 5-40b), we have
⎤
1 ⎡n 2
(µ + σ 2 ) − n(µ 2 + σ 2 / n)⎥
∑
⎢
n − 1 ⎣ i =1
⎦
1
=
( nµ 2 + nσ 2 − nµ 2 − σ 2 )
n −1
= σ2
Therefore, the sample variance S 2 is an unbiased estimator of the
population variance σ 2 .
E (S 2 ) =
z
z
z
z
Sometimes there are several unbiased estimators of the sample
population parameter.
For example, suppose we take a random sample of size n = 10 from a
normal population and obtain the data x1 = 12.8, x2 = 9.4, x3 = 8.7,
x4 = 11.6, x5 = 13.1, x6 = 9.8, x7 = 14.1, x8 = 8.5, x9 = 12.1, x10 =
10.3.
The sample mean
x=
z
12.8 + 9.4 + 8.7 + 11.6 + 13.1 + 9.8 + 14.1 + 8.5 + 12.1 + 10.3
= 11.04
10
Sample median
10.3 + 11.6
~
x=
= 10.95
2
z
10% trimmed mean (obtained by discarding the smallest and largest
10% of the sample before averaging)
xtr (10 ) =
8.7 + 9.4 + 9.8 + 10.3 + 11.6 + 12.1 + 12.8 + 13.1
= 10.98
8
3
z
We can show that all of these are unbiased estimates of µ .
7-2.3 Variance of a Point Estimator
Definition
If we consider all unbiased estimators of θ , the one with the smallest
variance is called the minimum variance unbiased estimator (MVUE).
Theorem 7-1
If X 1 , X 2 ,..., X n is a random sample of size n from a normal distribution with
mean µ and variance σ 2 , the sample mean X is the MVUE for µ .
7-2.4 Standard Error: Reporting a Point Estimate
Definition
The standard error of an estimator Θ̂ is its standard deviation, given by
ˆ ) . If the standard error involves unknown parameters that can be
σ Θˆ = V (Θ
estimated, substitution of those values into σ Θˆ produces an estimated
standard error, denoted by σˆ Θˆ .
z
z
z
z
The estimated standard error is denoted by sΘ̂ or se(Θˆ ) .
Suppose we are sampling from a normal distribution with mean µ and
variance σ 2 .
The distribution of X is normal with mean µ and variance σ 2 / n .
The standard error of X
σX =
z
σ
n
If we did not know σ but substituted the sample standard deviation S
into the above equation, the estimated standard error of X
σ̂ X =
S
n
4
EXAMPLE 7-2
o
z Using a temperature of 100 F and a power input of 550 watts, the
following 10 measurements of thermal conductivity (in Btu/hr-ft- o F)
were obtained:
41.60, 41.48, 42.34, 41.95, 41.86,
42.18, 41.72, 42.26, 41.81, 42.04
o
z A point estimate of the mean thermal conductivity at 110 F and 550
watts is the sample mean or
x = 41.924 Btu/hr-ft- o F
z The standard error of the sample mean is σ X = σ / n
z Since is unknown, we may replace it by the sample standard
deviation s = 0.284 to obtain the estimated standard error of X as
σˆ X =
0.284
s
=
= 0.0898
10
n
7-2.6 Mean Square Error of an Estimator
z
The mean square error of an estimator Θ̂ is the expected squared
difference between Θ̂ and θ .
Definition
The mean square error of an estimator Θ̂ of the parameter θ is defined
ˆ ) = E (Θ
ˆ − θ) 2
as
MSE (Θ
(7-3)
z
The mean square error can be rewritten as follows:
ˆ ) = E[ Θ
ˆ − E (Θ
ˆ )]2 + [θ − E (Θ
ˆ )]2 = V (Θ
ˆ ) + (bias ) 2
MSE (Θ
Exercise 7-2 : 7-11, 7-13, 7-17
5
7-3 METHODS OF POINT ESTIMATION
7-3.1 Method of Moments
Definition
Let X 1 , X 2 ,..., X n be a random sample from the probability distribution f (x),
where f (x) can be a discrete probability mass function or a continuous
probability density function. The kth population moment (or distribution
moment) is E ( X k ) , k = 1, 2, … . The corresponding kth sample moment is
n
(1 / n)∑i =1 X ik , k = 1,2,... .
z
z
To illustrate, the first population moment is E(X) = µ , and the first
n
sample moment is (1 / n)∑i =1 X i = X .
The sample mean is the moment estimator of the population mean.
Definition
Let X 1 , X 2 ,..., X n be a random sample from either a probability mass
function or probability density function with m unknown parameters
ˆ ,Θ
ˆ ,..., Θ
ˆ are found by equating the
θ1 , θ 2 ,..., θ m . The moment estimators Θ
1
2
m
first m population moments to the first m sample moments and solving the
resulting equations for the unknown parameters.
EXAMPLE 7-3
z Suppose that X 1 , X 2 ,..., X n is a random sample from an exponential
distribution with parameter λ .
z Now there is only one parameter to estimate, so we must equate E(X)
to X .
z For the exponential, E (X ) = 1 / λ .
ˆ = 1 / X is the moment
z Therefore E ( X ) = X results in 1 / λ = X , so λ
estimator of λ .
Ex:
z Suppose that the time to failure of an electronic module used in an
automobile engine controller is tested at an elevated temperature to
accelerate the failure mechanism.
z The time to failure is exponentially distributed.
z Eight units are randomly selected and tested, resulting in the
following failure time (in hours): x1 = 11.96, x2 = 5.03, x3 = 67.40,
x4 = 16.07, x5 = 31.50, x6 = 7.73, x7 = 11.10, and x8 = 22.38 .
6
z
Because x = 21.65 , the moment estimate of λ is λ = 1 / x = 1 / 21.65
= 0.0462 .
7-4 SAMPLING DISTRIBUTIONS
Definition
The probability distribution of a statistic is called a sampling distribution.
7-5 SAMPLING DISTRIBUTIONS OF MEANS
z
The sample mean
X =
X 1 + X 2 + ... + X n
n
has a normal distribution with mean
µX =
µ + µ + ... + µ
=µ
n
σ 2X =
σ 2 + σ 2 + ... + σ 2 σ 2
=
n
n2
and variance
z
z
If we are sampling from a population that has an unknown probability
distribution, the sampling distribution of the sample mean will still be
approximately normal with mean µ and variance σ 2 / n , if the sample
size n is large.
This is one of the most useful theorems in statistics, called the central
limit theorem.
Theorem 7-2: The Central Limit Theorem
If X 1 , X 2 ,..., X n is a random sample of size n taken from a population (either
finite or infinite) with mean µ and finite variance σ 2 , and if X is the
sample mean, the limiting form of the distribution of
Z=
X −µ
σ/ n
(7-6)
as n → ∞ , is the standard normal distribution.
7
z
z
z
z
The normal approximation for X depends on the sample size n.
Figure 7-6(a) shows the distribution obtained for throws of a single,
six-sided true die. The probabilities are equal (1/6) for all the values
obtained, 1, 2, 3, 4, 5, or 6.
Figure 7-6(b) shows the distribution of the average score obtained when
tossing two dice.
Fig. 7-6(c), 7-6(d), and 7-6(e) show the distributions of average scores
obtained when tossing three, five, and ten dice, respectively.
Figure 7-6
Distributions of average scores from throwing dice. [Adapted
with permission from Box, Hunter, and Hunter(1978).]
8
z
z
In many cases of practical interest, if n ≥ 30, the normal approximation
will be satisfactory regardless of the shape of the population.
If n < 30, the central limit theorem will work if the distribution of the
population is not severely nonnormal.
EXAMPLE 7-13
z An electronics company manufactures resistors that have a mean
resistance of 100 ohms and a standard deviation of 10 ohms.
z The distribution of resistance is normal.
z Find the probability that a random sample of n = 25 resistors will
have an average resistance less than 95 ohms.
z The sampling distribution of X is normal, with mean µ X = 100
and a standard deviation of
σX =
z
z
σ
10
=
=2
n
25
The desired probability corresponds to the shaded area in Fig. 7-7.
Standardizing the point X = 95 in Fig. 7-7, we find that
95 − 100
= −2 . 5
2
p( X < 95) = P( Z < −2.5) = 0.0062
z=
EXAMPLE 7-14
z Suppose that a random variable X has a continuous uniform
distribution.
⎧1 / 2,4 ≤ x ≤ 6
f ( x) = ⎨
⎩ 0, otherwise
z
z
z
z
Find the distribution of the sample mean of a random sample of size
n = 40.
The mean and variance of X are µ = 5 and σ 2 = (6 − 4) 2 / 12 = 1 / 3 .
The central limit theorem indicates that the distribution of X is
approximately normal with mean µ X = 5 and variance σ 2X = σ 2 / n
= 1 /[3(40)] = 1 / 120 .
The distributions of X and are shown in Fig. 7-8.
9
Figure 7-7 Probability for Example 7-13.
Figure 7-8
z
z
z
z
The distributions of X and X for Example 7-14.
We have two independent populations.
Let the first population have mean µ1 and variance σ12 and the second
population have mean µ 2 and variance σ 22 .
Suppose that both populations are normally distributed.
The sampling distribution of X 1 − X 2 is normal with mean
µ x1 − x2 = µ x1 − µ x2 = µ1 − µ 2
σ 2X1 − X 2 = σ 2X1 − σ 2X 2 =
σ12 σ 22
+
n1 n2
10
Definition
If we have two independent populations with means µ1 and µ 2 and
variances σ 22 and σ 22 and if X 1 and X 2 are the sample means of two
independent random samples of sizes n1 and n2 from these populations,
then the sampling distribution of
Z=
X 1 − X 2 − (µ1 − µ 2 )
(7-9)
σ12 / n1 + σ 22 / n2
is approximately standard normal, if the conditions of the central limit
theorem apply. If the two populations are normal, the sampling distribution
of Z is exactly standard normal.
EXAMPLE 7-15
z
The effective life of a component used in a jet-turbine aircraft engine is
a random variable with mean 5000 hours and standard deviation 40
hours.
z
An improvement into the manufacturing process for this component
that increases the mean life to 5050 hours and decreases the standard
deviation to 30 hours.
z
A random sample of n1 = 16 components is selected from the “old”
process and a random sample of n2 = 25 components is selected from
the “improved” process.
z
What is the probability that the difference in the two sample means
X 2 − X 1 is at least 25 hours?
Old
New
µ1 = 5000
µ 2 = 5050
σ1 = 40
σ 2 = 30
µ X1
n1 = 16
= 5000
σ X1 =
n2 = 25
µ X 2 = 5050
40
= 10
16
σ X2 =
∴ µ X1 − X 2 = µ X1 − µ X 2 = 5000 − 5050 = −50
σ 2X1 − X 2 = σ 2X 1 + σ 2X 2 = 10 2 + 6 2 = 136
11
30
=6
25
z
Corresponding to the value in Fig. 7-9, we find that
25 − 50
= −2.14
136
P( X 2 − X 1 ≥ 25) = P( Z ≥ −2.14) = 0.9838
z=
Figure 7-9
The sampling distribution of in Example 7-15.
Exercise 7-5 : 7-33, 7-35, 7-37, 7-39, 7-41, 7-43, 7-49, 7-55
12
中華大學資訊工程學系
學號:
姓名:
日期:
時間:
輔導老師:
課程名稱:
年級:□一年級
班別:□甲
□乙
□二年級
地點:
□三年級
□丙(二部)
課後輔導記錄表
□四年級
□二技專
□研一
□碩專
問題描述:
問題回覆:
系所主任簽章:
輔導老師簽章:
13
□研二
Related documents