Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Point Estimation Statistic A function of observations, X1 , X 2 ,…, X n Also a random variable Sample mean X Sample variance S 2 Its probability distribution is called a sampling distribution Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Point estimator of ◦ A statistic ˆ h( X1, X 2 ,..., X n ) Point estimate of ◦ A particular numerical value ˆ Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Mean ◦ The estimate ̂ x Variance 2 ◦ The estimate ˆ 2 s 2 Proportion p ◦ The estimate pˆ x / n ◦ x is the number of items that belong to the class of interest Difference in means, 1 2 ◦ The estimate ˆ1 ˆ 2 xˆ1 xˆ2 Difference in two proportions ◦ The estimate pˆ1 pˆ 2 p1 p2 Sampling Distributions and the Central Limit Theorem Random sample ◦ The random variables X1 , X 2 ,…, X n are a random sample of size n if (a) the X i‘s are independent random variables, and (b) every X i has the same probability distribution If X1 , X 2 ,…, X n are normally and independently distributed with mean and variance 2 ◦ X X 1 X 2 X n has a normal distribution n ◦ with mean X n ◦ variance 2 X 2 2 2 n 2 2 n Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Central Limit Theorem ◦ If X1 , X 2 ,…, X n is a random sample of size n taken from a population (either finite or infinite) with mean and finite variance 2 , and if X is the sample mean, the limiting form of the distribution of Z X / n ◦ as n , is the standard normal distribution. Works when ◦ n 30 , regardless of the shape of the population ◦ n 30 , if not severely nonnormal Two independent populations with means 2 2 and 2 , and variances 1 and 2 Z 1 X 1 X 2 ( 1 2 ) 12 / n1 22 / n2 ◦ is approximately standard normal, or ◦ is exactly standard normal if the two populations are normal Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Example 7-1 Resistors ◦ An electronics company manufactures resistors that have a mean resistance of 100 ohms and a standard deviation of 10 ohms. The distribution of resistance is normal. Find the probability that a random sample of n 25 resistors will have an average resistance less than 95 ohms Example 7-2 Central Limit Theorem ◦ Suppose that X has a continuous uniform distribution 1 / 2 4 x 6 f ( x) 0 otherwise ◦ Find the distribution of the sample mean of a random sample of size n 40 Example 7-3 Aircraft Engine Life ◦ The effective life of a component used in a jet-turbine aircraft engine is a random variable with mean 5000 hours and standard deviation 40 hours. The distribution of effective life is fairly close to a normal distribution. The engine manufacturer introduces an improvement into the manufacturing process for this component that increases the mean life to 5050 hours and decreases the standard deviation to 30 hours. Suppose that a random sample of n1 16 components is selected from the “old” process and a random sample of n2 25 components is selected from the “improved” process. What is the probability that the difference in the two sample means X 2 X 1 is at least 25 hours? Assume that the old and improved processes can be regarded as independent populations. Exercise 7-10 ◦ Suppose that the random variable X has the continuous uniform distribution 1 0 x 1 f ( x) 0 otherwise ◦ Suppose that a random sample of n 12 observations is selected from this distribution. What is the approximate probability distribution of X 6 ? Find the mean and variance of this quantity. Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. General Concepts of Point Estimation Bias of the estimator ̂ ˆ E () ̂ is an unbiased estimator if ˆ E () Minimum variance unbiased estimator (MVUE) ◦ For all unbiased estimator of , the one with the smallest variance X is the MVUE for ◦ If X1 , X 2 ,…, X n are from a normal distribution with mean and variance 2 Standard error of an estimator ̂ ˆ) ˆ V ( Estimated standard error ◦ ˆ) ˆ ̂ or s ̂ or se( is normal with mean variance 2 /n IfX X n ̂ X and and S n Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Mean squared error of an estimate ̂ ˆ ) E[( ˆ )2 ] MSE ( ˆ E ( ˆ ) E ( ˆ ) )2 ] E[( ˆ E ( ˆ )) 2 ] E[( E ( ˆ ) ) 2 ] 2 E[ ˆ E ( ˆ )] ( E ( ˆ ) ) E[( ˆ E ( ˆ )) 2 ] ( E ( ˆ ) )2 E[( ˆ ) (bias ) 2 V ( Relative efficiency of ̂ 2 to ̂1 ˆ ) MSE ( 1 ˆ ) MSE ( 2 Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Example 7-4 Sample Mean and Variance Are Unbiased ◦ Suppose that X is a random variable with mean and variance 2 . Let X1 , X 2 ,…., X n be a random sample of size n from the population represented by X . Show that the sample mean X and sample variance S 2 are unbiased estimators of and 2 , respectively. Example 7-5 Thermal Conductivity ◦ Ten measurements of thermal conductivity were obtained: ◦ 41.60, 41.48, 42.34, 41.95, 41.86 ◦ 42.18, 41.72, 42.26, 41.81, 42.04 s 0.284 ◦ Show that x 41.924 and ˆ x 0.0898 n 10 Exercise 7-31 ◦ X 1 and S12 are the sample mean and sample variance from a population with mean 1 and variance 12 . Similarly, X 2 and S 22 are the sample mean and sample variance from a second independent population with mean 2 and variance 22 . The sample sizes are n1 and n2 , respectively. ◦ (a) Show that X 1 X 2 is an unbiased estimator of 1 2 ◦ ◦ (b) Find the standard error of X 1 X 2 . How could you estimate the standard error? ◦ (c) Suppose that both populations have the same 2 2 2 variance; that is, 1 . Show that 2 (n1 1) S12 (n2 1) S 22 S n1 n2 2 2 ◦ Is an unbiased estimator of . 2 p Methods of Point Estimation Moments ◦ Let X1 , X 2 ,…, X n be a random sample from the probability distribution f (x) , where f (x) can be a discrete probability mass function or a continuous probability density function. The kth population moment (or distribution k E ( X ) , k = 1, 2,…. The moment) is 1 n corresponding kth sample moment is i 1 X ik n k = 1, 2, …. Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Moment estimators ◦ Let X1 , X 2 ,…, X n be a random sample from either a probability mass function or a probability density function with m unknown parameters 1 , 2 ,…, m . The moment estimators ̂1 , ̂ 2 ,…, ̂m are found by equating the first m population moments to the first m sample moments and solving the resulting equations for the unknown parameters. Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Maximum likelihood estimator ◦ Suppose that X is a random variable with probability distribution f ( x; ) , where is a single unknown parameter. Let x1 , x2 ,…, xn be the observed values in a random sample of size n . Then the likelihood function of the sample is L( ) f ( x1; ) f ( x2 ; ) f ( xn ; ) ◦ Note that the likelihood function is now a function of only the unknown parameter . The maximum likelihood estimator (MLE) of is the value of that maximizes the likelihood function L ( ) . Properties of a Maximum Likelihood Estimator ◦ Under very general and not restrictive conditions, when the sample size n is large and if ̂ is the maximum likelihood estimator of the parameter , ◦ (1) ̂ is an approximately unbiased estimator for [ E (ˆ ) ] ◦ (2) the variance of ̂ is neatly as small as the variance that could be obtained with any other estimator, and ◦ (3) ̂ has an approximate normal distribution. Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Invariance property ◦ Let ̂1 , ̂ 2 ,…., ̂ k be the maximum likelihood estimators of the parameters 1 , 2 , …, k . Then the maximum likelihood estimator of any function h(1 , 2 ,..., k ) of these parameters is the same function ˆ , ˆ ,..., ˆ ) h( 1 2 k ◦ of the estimators ̂1 , ̂ 2 ,…, ̂ k . Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Bayesian estimation of parameters ◦ Sample X1 , X 2 ,…, X n ◦ Joint probability distribution f ( x1 , x2 ,..., xn , ) f ( x1 , x2 ,..., xn | ) f ( ) ◦ Prior distribution for f ( ) ◦ Posterior distribution for f ( | x1 , x2 ,..., xn ) f ( x1 , x2 ,..., xn , ) f ( x1 , x2 ,..., xn ) ◦ Marginal distribution f ( x1 , x2 ,..., xn , ) f ( x1 , x2 ,..., xn ) f ( x1 , x2 ,..., xn , )d θ discrete θ continuous Example 7-6 Exponential Distribution Moment Estimator ◦ Suppose that X1 , X 2 ,…, X n is a random sample from an exponential distribution with parameter . For the exponential, E ( X ) . ◦ Then E ( X ) X results in 1 / X . ˆ 1 / X Example 7-7 Normal Distribution Moment Estimators ◦ Suppose that X1 , X 2 ,…, X n is a random sample from a normal distribution with parameters and 2 . For the normal distribution, E ( X ) and ◦ E ( X 2 ) 2 2 . Equating E ( X ) to X and E ( X 2 ) n to 1n i 1 X i2 gives n 2 2 2 1 X ◦ i n i 1 X and ◦ Solve these equations. Example 7-8 Gamma Distribution Moment Estimators ◦ Suppose that X1 , X 2 ,…, X n is a random sample from a gamma distribution with parameters r and , For the gamma distribution, E ( X ) r / and ◦ E ( X 2 ) r (r 1) / 2 . Solve n 2 2 1 ◦ r / X and r (r 1) / n i 1 X i Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Example 7-9 Bernoulli Distribution MLE ◦ Let X be a Bernoulli random variable. The probability mass function is p x (1 p)1 x x 0,1 f ( x; p) 0 otherwise ◦ where p is the parameter to be estimated. The likelihood function of a random sample of size n is n n L( p ) p (1 p ) xi 1 xi xi p i1 (1 p ) n n xi i 1 i 1 ◦ Find p̂ that maximizes L( p ) . Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Example 7-10 Normal Distribution MLE ◦ Let X be normally distributed with unknown and known variance 2 . The likelihood function of a random sample of size n , say X1, X 2 ,…, X n , is 1 n ( x )2 2 i 2 2 1 1 2 i 1 L( ) e ( xi ) /(2 ) e (2 2 ) n / 2 2 i 1 ◦ Find ̂ . n Example 7-11 Exponential Distribution MLE ◦ Let X be exponentially distributed with parameter . The likelihood function of a random sample of size n , say X1 , X 2 ,…, X n , is n L( ) e xi n e i 1 ◦ Find ̂ . n xi i 1 Example 7-12 Normal Distribution MLEs for and 2 ◦ Let X be normally distributed with mean and variance 2 , where both and 2 are unknown. The likelihood function of a random sample of size n is 1 n ( x )2 2 i 2 2 1 1 2 i 1 L( , 2 ) e ( xi ) /(2 ) e (2 2 ) n / 2 2 i 1 n ◦ Find ̂ and ˆ 2 . Example 7-13 ◦ From Example 7-12, to obtain the maximum likelihood estimator of the function h( , 2 ) 2 ◦ Substitute the estimators ̂ and ˆ 2 into the function 1/ 2 n , which yields ˆ ˆ 2 1n ( X i X ) 2 i 1 Example 7-14 Uniform Distribution MLE ◦ Let X be uniformly distributed on the interval 0 to a . Since the density function is f ( x) 1 / a for 0 x a and zeros otherwise, the likelihood function of a random sample of size n is n 1 1 L(a ) n a i 1 a ◦ for 0 x1 a, 0 x2 a ,…, 0 xn a ◦ Find â . Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Example 7-15 Gamma Distribution MLE ◦ Let X1 ,X 2 ,…, X n be a random sample from the gamma distribution. The log of likelihood function is n r xir 1e xi ln L(r , ) ln ( r ) i 1 n n i 1 i 1 nr ln( ) (r 1) ln( xi ) n ln[ (r )] xi ◦ Find that r̂ ˆ x n ln( ˆ ) ' (rˆ) ln( xi ) n (rˆ) i 1 n Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Example 7-16 Bayes Estimator for the Mean of a Normal Distribution ◦ Let X1, X 2,…, X n be a random sample from the normal distribution with mean and variance 2 , where is unknown and 2 is known. Assume that the prior distribution for is normal with mean 0 2 and variance 0 ; that is, 1 f ( ) e ( ) /(2 ) 2 0 2 2 0 0 ◦ The joint probability distribution of the sample is f ( x1 , x2 ,..., xn | ) 1 (2 ) 2 n/2 (1 / 2 ) 2 e n ( xi ) 2 i 1 Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. ◦ Show that 02 x ( 2 / n ) 0 1 (1/ 2 )( 2 2 ) 0 / n 02 2 / n 1 f ( | x1 , x2 ,..., xn ) e 2 ◦ Then the Bayes estimate of is 02 x ( 2 / n) 0 02 2 / n Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Exercise 7-42 ◦ Let X1, X 2 ,…, X n be uniformly distributed on the interval 0 to a . Recall that the maximum likelihood estimator of a is aˆ max( X i ) . ◦ (a) Argue intuitively why â cannot be an unbiased estimator for a . ◦ (b) Suppose that E (aˆ ) na /( n 1) . Is it reasonable that â consistently underestimates a ? Show that the bias in the estimator approaches zero as n gets large. ◦ (c) Propose an unbiased estimator for a . ◦ (d) Let Y max( X i ) . Use the fact that Y y if and only if each X i y to derive the cumulative distribution function of Y . Then show that the probability density function of Y is ny n 1 f ( y) a n 0 0 ya otherwiae ◦ Use this result to show that the maximum likelihood estimator for a is biased. ◦ (e) We have two unbiased estimators for a : the moment estimator aˆ1 2 X and ◦ aˆ2 [(n 1) / n] max( X i ) , where max( X i ) is the largest observation in a random sample of size n . It can be shown that V (aˆ1 ) a 2 /(3n) and that ◦ V (aˆ2 ) a 2 /[ n(n 2)] . Show that if n 1 , â2 is a better estimator than â1 . In what sense is it a better estimator of a ? Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Exercise 7-50 ◦ The time between failures of a machine has an exponential distribution with parameter . Suppose that the prior distribution for is exponential with mean 100 hours. Two machines are observed, and the average time between failures is x 1125 hours. ◦ (a) Find the Bayes estimate for . ◦ (b) What proportion of the machine do you think will fail before 1000 hours? Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011. Contents, figures, and exercises come from the textbook: Applied Statistics and Probability for Engineers, 5th Edition, by Douglas C. Montgomery, John Wiley & Sons, Inc., 2011.