Chapter 11: Testing a Claim Confidence intervals are one of the two most common types of statistical inference. We use confidence intervals when we wish to estimate a population parameter. The second type of inference is called significance tests. The goal here is to assess the evidence provided by data about some claim concerning the population. Example: I claim I can make 80% of my basketball free throws. We could test this by collecting data on free throws and seeing if the % is close to the claim. If it is not, we can determine if the % if “far enough” away from the claim to reject that original claim. Summary of Significance Testing 1) Hypothesis about a population parameter (m or p) 2) Check conditions for testing 3) Collect data to find a test statistic 4) Find a probability that extreme values are possible (p-value) 5) Look at the significance level and interpret the results of the claim 6) The reasoning is that we are looking at what will happen in the long run with repeated sampling or experiments. Significant = not likely to happen by chance Stating the Hypothesis Null Hypothesis (H0): the statement being tested in a significance test (“no effect”, “no difference”, “no change”) We are trying to find evidence AGAINST this claim Alternative Hypothesis (Ha): the claim about the population that we are trying to find evidence FOR. These are both stated BEFORE the data (evidence) is provided. Ha > H0 value Ha < H0 value Ha can also be ≠ (this makes the significance testing two-sided) Conditions for Significance Testing 1) SRS 2) Normality • For means: population distribution is Normal or n ≥ 30 (CLT) • For proportions: npˆ 10 and n(1 pˆ ) 10 3) Independence As with any type of inference, you MUST check the conditions and state the reasoning for each condition EVERY TIME!! Test Statistic •Collect data and calculate x or p̂ and standard error •We want to compare the sample’s calculations with H0, so we compute a test statistic (z-score) test statistic (z) = estimate ( x or p̂ ) – H0 value standard error But how large is large enough to prove H0 untrue? P-value: the probability of getting another sample statistic as or more extreme than the current estimate when H0 is assumed true …the probability that the current estimate is a “fluke” The smaller the P-value, the stronger evidence AGAINST H0 To determine if the P-value is “good enough”, we compare that probability to the a significance level (a). a= the maximum P-value to still provide evidence AGAINST H0 Significance levels, like confidence levels, are chosen at the discretion of the statistician depending on the situation. a typically is set at 0.01, 0.05, or 0.10 We want P ≤ a 1) Hypothesis: Let’s assume the mean times did NOT decrease H0: m = 6.7 minutes Ha: m < 6.7 minutes 2) Conditions: - randomness: SRS of 400 was collected - Normality: by CLT this distribution will also be normal and standard error will be 2 0.10 n 400 3) Test Statistic: 6.48 6.7 z 2.20 2 / 400 If z is large and in the direction of Ha, then it is unlikely that H0 is true. To find the P-value, look up the test statistic (z) in the table z=-2.20 0.0139 Compare to a significance level… For a = 0.01, our P-value of 0.0139 would NOT be significant (ie: the 6.48 could be a fluke) For a = 0.05, our P-value of 0.0139 WOULD be significant (ie: the 6.48 is mostly likely NOT a fluke and there is a significant decrease in response time) The significance level should be determined BEFORE you collect the data and calculate a test statistic. (At the same time you create the hypotheses.) Interpreting the Results/Draw a Conclusion P≤a P>a we reject H0 we fail to reject H0 If H0 is true, the sample results are too unlikely to occur by chance The sample results could possibly occur by chance This does not mean we accept Ha This does not mean we accept H0
© Copyright 2026 Paperzz