STATISTICS 330 COURSE NOTES Cyntha A. Struthers Department of Statistics and Actuarial Science, University of Waterloo Spring 2021 Edition ii 1 Contents 1. Preview 1 1.1 Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.2 Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 2. Univariate Random Variables 5 2.1 Probability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 2.2 Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 2.3 Discrete Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 2.4 Continuous Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . . 19 2.5 Location and Scale Parameters . . . . . . . . . . . . . . . . . . . . . . . . . 27 2.6 Functions of a Random Variable . . . . . . . . . . . . . . . . . . . . . . . . 30 2.7 Expectation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 2.8 Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 2.9 Variance Stabilizing Transformation . . . . . . . . . . . . . . . . . . . . . . 44 2.10 Moment Generating Functions . . . . . . . . . . . . . . . . . . . . . . . . . 45 2.11 Calculus Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 2.12 Chapter 2 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56 3. Multivariate Random Variables 61 3.1 Joint and Marginal Cumulative Distribution Functions . . . . . . . . . . . . 62 3.2 Bivariate Discrete Distributions . . . . . . . . . . . . . . . . . . . . . . . . . 63 3.3 Bivariate Continuous Distributions . . . . . . . . . . . . . . . . . . . . . . . 66 3.4 Independent Random Variables . . . . . . . . . . . . . . . . . . . . . . . . . 76 3.5 Conditional Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84 3.6 Joint Expectations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 90 3.7 Conditional Expectation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97 3.8 Joint Moment Generating Functions . . . . . . . . . . . . . . . . . . . . . . 102 3.9 Multinomial Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108 3.10 Bivariate Normal Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . 111 3.11 Calculus Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114 3.12 Chapter 3 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116 iii iv CONTENTS 4. Functions of Two or More Random Variables 121 4.1 Cumulative Distribution Function Technique . . . . . . . . . . . . . . . . . 121 4.2 One-to-One Transformations . . . . . . . . . . . . . . . . . . . . . . . . . . 125 4.3 Moment Generating Function Technique . . . . . . . . . . . . . . . . . . . . 137 4.4 Chapter 4 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147 5. Limiting or Asymptotic Distributions 151 5.1 Convergence in Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . 152 5.2 Convergence in Probability . . . . . . . . . . . . . . . . . . . . . . . . . . . 156 5.3 Weak Law of Large Numbers . . . . . . . . . . . . . . . . . . . . . . . . . . 159 5.4 Moment Generating Function Technique for Limiting Distributions . . . . . 167 5.5 Additional Limit Theorems . . . . . . . . . . . . . . . . . . . . . . . . . . . 172 5.6 Chapter 5 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 179 6. Maximum Likelihood Estimation - One Parameter 183 6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184 6.2 Maximum Likelihood Method . . . . . . . . . . . . . . . . . . . . . . . . . . 185 6.3 Score and Information Functions . . . . . . . . . . . . . . . . . . . . . . . . 191 6.4 Likelihood Intervals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204 6.5 Limiting Distribution of Maximum Likelihood Estimator . . . . . . . . . . . 208 6.6 Con
dence Intervals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 213 6.7 Approximate Con
dence Intervals . . . . . . . . . . . . . . . . . . . . . . . 218 6.8 Chapter 6 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 224 7. Maximum Likelihood Estimation - Multiparameter 227 7.1 Likelihood and Related Functions . . . . . . . . . . . . . . . . . . . . . . . . 228 7.2 Likelihood Regions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237 7.3 Limiting Distribution of Maximum Likelihood Estimator . . . . . . . . . . . 240 7.4 Approximate Con
dence Regions . . . . . . . . . . . . . . . . . . . . . . . . 242 7.5 Chapter 7 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 249 8. Hypothesis Testing 253 8.1 Test of Hypothesis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 253 8.2 Likelihood Ratio Tests for Simple Hypotheses . . . . . . . . . . . . . . . . . 257 8.3 Likelihood Ratio Tests for Composite Hypotheses . . . . . . . . . . . . . . . 265 8.4 Chapter 8 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 273 9. Solutions to Chapter Exercises 275 9.1 Chapter 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 276 9.2 Chapter 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 288 9.3 Chapter 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 305 9.4 Chapter 5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 316 CONTENTS v 9.5 Chapter 6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 318 9.6 Chapter 7 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325 9.7 Chapter 8 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 334 10. Solutions to Selected End of Chapter Problems 337 10.1 Chapter 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 338 10.2 Chapter 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 383 10.3 Chapter 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 412 10.4 Chapter 5 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 435 10.5 Chapter 6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 448 10.6 Chapter 7 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 460 10.7 Chapter 8 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 465 11. Summary of Named Distributions 469 12. Distribution Tables 471 0 CONTENTS Preface In order to provide improved versions of these Course Notes for students in subsequent terms, please email corrections, sections that are confusing, or comments/suggestions to
[email protected]. 1. Preview The following examples will illustrate the ideas and concepts discussed in these Course Notes. They also indicate how these ideas and concepts are connected to each other. 1.1 Example The number of service interruptions in a communications system over 200 separate days is summarized in the following frequency table: Number of interruptions 0 1 2 3 4 5 > 5 Total Observed frequency 64 71 42 18 4 1 0 200 It is believed that a Poisson model will
t these data well. Why might this be a reasonable assumption? (PROBABILITY MODELS ) If we let the random variable X = number of interruptions in a day and assume that the Poisson model is reasonable then the probability function of X is given by P (X = x) = xe x! for x = 0; 1; : : : where is a parameter of the model which represents the mean number of service inter- ruptions in a day. (RANDOM VARIABLES, PROBABILITY FUNCTIONS, EXPEC- TATION, MODEL PARAMETERS ) Since is unknown we might estimate it using the sample mean x = 64(0) + 71(1) + + 1(5) 200 = 230 200 = 1:15 (POINT ESTIMATION ) The estimate ^ = x is the maximum likelihood estimate of . It is the value of which maximizes the likelihood function. (MAXIMUM LIKELIHOOD ESTIMATION ) The likelihood function is the probability of the observed data as a function of the unknown parameter(s) in the model. The maximum likelihood estimate is thus the value of which maximizes the probability of observing the given data. 1 2 1. PREVIEW In this example the likelihood function is given by L() = P (observing 0 interruptions 64 times,. . . , > 5 interruptions 0 times;) = 200! 64!71! 1!0! 0e 0! 64 1e 1! 71 5e 5! 1 1P x=6 xe x! 0 = c64(0)+71(1)++1(5)e (64+71++1) = c 230e 200 for > 0 where c = 200! 64!71! 1!0! 1 0! 64 1 1! 71 : : : 1 5! 1 The maximum likelihood estimate of can be found by solving dLd = 0 or equivalently d logL d = 0 and verifying that it corresponds to a maximum. If we want an interval of values for which are reasonable given the data then we could construct a con
dence interval for . (INTERVAL ESTIMATION ) To construct con
dence intervals we need to
nd the sampling distribution of the estimator. In this example we would need to
nd the distribution of the estimator X = X1 +X2 + +Xn n where Xi = number of interruptions in a day i, i = 1; 2; : : : ; 200. (FUNCTIONS OF RANDOM VARIABLES: cumulative distribution function technique, one-to-one transfor- mations, moment generating function technique) Since Xi Poisson() with E(Xi) = and V ar(Xi) = the distribution of X for large n is approximately N(; =n) by the Central Limit Theorem. (LIMITING DISTRIBUTIONS ) Suppose the manufacturer of the communications system claimed that the mean number of interruptions was 1. Then we would like to test the hypothesis H : = 1. (TESTS OF HYPOTHESIS ) A test of hypothesis uses a test statistic to measure the evidence based on the observed data against the hypothesis. A test statistic with good properties for testing H : = 0 is the likelihood ratio statistic, 2 log [L (0) =L (^)]. (LIKELIHOOD RATIO STATISTIC ) For large n the distribution of the likelihood ratio statistic is approximately 2 (1) if the hypothesis H : = 0 is true. 1.2 Example The following are relief times in hours for 20 patients receiving a pain killer: 1:1 1:4 1:3 1:7 1:9 1:8 1:6 2:2 1:7 2:7 4:1 1:8 1:5 1:2 1:4 3:0 1:7 2:3 1:6 2:0 It is believed that the Weibull distribution with probability density function f(x) = x 1e (x=) for x > 0; > 0; > 0 1.2. EXAMPLE 3 will provide a good
t to the data. (CONTINUOUS MODELS, PROBABILITY DENSITY FUNCTIONS ) Assuming independent observations the (approximate) likelihood function is L(; ) = 20Q i=1 x 1i e