1 / 27

Power and Sample Size

1  .  /2.  /2. H 0 :  =  0. Acceptance Region for H 0. Rejection Region. Rejection Region. Power and Sample Size. “Null Distribution”.

karik
Download Presentation

Power and Sample Size

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. 1  /2 /2 H0: = 0 Acceptance Region for H0 Rejection Region Rejection Region Power and Sample Size “Null Distribution” • IF the null hypothesis H0: μ = μ0is true, then we should expect a random sample mean to lie in its “acceptance region” with probability 1 – α, the “confidence level.” • That is, P(Accept H0 | H0 is true) = 1 – α. • Therefore, we should expect a random sample mean to lie in its “rejection region” with probability α, the “significance level.” • That is, P(Reject H0 | H0 is true) = α. “Type 1 Error” μ0+ zα/2(σ /

  2. 1  /2 /2 H0: = 0 Acceptance Region for H0 Rejection Region Rejection Region Power and Sample Size “Alternative Distribution” “Null Distribution” • IF the null hypothesis H0: μ = μ0is false, then the “power” to correctly reject it in favor of a particular alternative HA: μ = μ1is P(Reject H0 | H0 is false) = 1 – . Thus, P(AcceptH0 | H0 is false) = . 1 –   HA: μ = μ1 Set them equal to each other, and solve for n… “Type 2 Error” μ1 –z(σ / μ0+ zα/2(σ /

  3. Given: • X ~ N(μ, σ) Normally-distributed population random variable, with unknown mean, but known standard deviation • H0: μ = μ0Null Hypothesis value • HA: μ = μ1Alternative Hypothesis specific value • significance level (or equivalently, confidence level 1 –) • 1 – power (or equivalently, Type 2 error rate ) Then the minimum required sample size is: N(0, 1) 1   z  z.025 = 1.96 Example:σ= 1.5 yrs, μ0= 25.4 yrs,  = .05 Suppose it is suspected that currently, μ1= 26 yrs. Want more power! Want 90% powerof correctly rejecting H0 in favor of HA, if it is false  = |26 – 25.4| / 1.5 = 0.4  z.10 = 1.28  1 –  = .90   = .10 n  66 So… minimum sample size required is

  4. Given: • X ~ N(μ, σ) Normally-distributed population random variable, with unknown mean, but known standard deviation • H0: μ = μ0Null Hypothesis value • HA: μ = μ1Alternative Hypothesis specific value • significance level (or equivalently, confidence level 1 –) • 1 – power (or equivalently, Type 2 error rate ) Then the minimum required sample size is: N(0, 1) 1   z  z.025 = 1.96 Example:σ= 1.5 yrs, μ0= 25.4 yrs,  = .05 Change μ1 Suppose it is suspected that currently, μ1= 26 yrs. Want 95% powerof correctly rejecting H0 in favor of HA, if it is false Want 90% powerof correctly rejecting H0 in favor of HA, if it is false  = |26 – 25.4| / 1.5 = 0.4  z.10 = 1.28  1 –  = .90   = .10   = .05  z.05 = 1.645  1 –  = .95 n  82 n  66 So… minimum sample size required is

  5. Given: • X ~ N(μ, σ) Normally-distributed population random variable, with unknown mean, but known standard deviation • H0: μ = μ0Null Hypothesis value • HA: μ = μ1Alternative Hypothesis specific value • significance level (or equivalently, confidence level 1 –) • 1 – power (or equivalently, Type 2 error rate ) Then the minimum required sample size is: N(0, 1) 1   z  z.025 = 1.96 Example:σ= 1.5 yrs, μ0= 25.4 yrs,  = .05 Suppose it is suspected that currently, μ1= 25.7 yrs. Suppose it is suspected that currently, μ1= 26 yrs. Want 95% powerof correctly rejecting H0 in favor of HA, if it is false  = |25.7 – 25.4| / 1.5 = 0.2  = |31 – 30| / 2.4 = 0.41667   = .05  1 –  = .95  z.05 = 1.645 n  325 n  82 So… minimum sample size required is

  6. Given: • X ~ N(μ, σ) Normally-distributed population random variable, with unknown mean, but known standard deviation • H0: μ = μ0Null Hypothesis value • HA: μ = μ1Alternative Hypothesis specific value • significance level (or equivalently, confidence level 1 –) • 1 – power (or equivalently, Type 2 error rate ) Then the minimum required sample size is: N(0, 1) 1   z  z.025 = 1.96 Example:σ= 1.5 yrs, μ0= 25.4 yrs,  = .05 Suppose it is suspected that currently, μ1= 25.7 yrs. With n = 400, how much power exists to correctly reject H0 in favor of HA, if it is false? Power = 1 – = = 0.9793, i.e., 98%

  7. Comments

  8. Given: • X ~ N(μ, σ) Normally-distributed population random variable, with unknown mean, but known standard deviation • H0: μ = μ0Null Hypothesis • HA: μ ≠ μ0 Alternative Hypothesis (2-sided) • significance level (or equivalently, confidence level 1 –) • n sample size From this, we obtain… “standard error” s.e. sample mean sample standard deviation …with which to test the null hypothesis (via CI, AR, p-value). In practice however, it is far more common that the true population standard deviation σ is unknown. So we must estimate it from the sample! But this introduces additional variability from one sample to another… PROBLEM! (estimate) x1, x2,…, xn Recall that

  9. Given: • X ~ N(μ, σ) Normally-distributed population random variable, with unknown mean, but known standard deviation • H0: μ = μ0Null Hypothesis • HA: μ ≠ μ0 Alternative Hypothesis (2-sided) • significance level (or equivalently, confidence level 1 –) • n sample size From this, we obtain… “standard error” s.e. sample mean sample standard deviation …with which to test the null hypothesis (via CI, AR, p-value). SOLUTION: follows a different sampling distribution from before. But this introduces additional variability from one sample to another… PROBLEM! (estimate) x1, x2,…, xn

  10. Student’s T-Distribution … is actually a family of distributions, indexed by the degrees of freedom, labeled tdf. Z ~ N(0, 1) t10 t3 t2 William S. Gossett (1876 - 1937) t1 As the sample size n gets large, tdf converges to the standard normal distribution Z ~ N(0, 1). So the T-test is especially useful when n < 30.

  11. Student’s T-Distribution … is actually a family of distributions, indexed by the degrees of freedom, labeled tdf. Z ~ N(0, 1) t4 William S. Gossett (1876 - 1937) .025 1.96 As the sample size n gets large, tdf converges to the standard normal distribution Z ~ N(0, 1). So the T-test is especially useful when n < 30.

  12. Lecture Notes Appendix… or… qt(.025, 4, lower.tail = F) [1] 2.776445

  13. Student’s T-Distribution … is actually a family of distributions, indexed by the degrees of freedom, labeled tdf. Z ~ N(0, 1) t4 William S. Gossett (1876 - 1937) .025 .025 1.96 2.776 Because any t-distribution has heavier tails than the Z-distribution, it follows that for the same right-tailed area value, t-score > z-score.

  14. Given: X = Age at first birth~ N(μ, σ) • H0: μ = 25.4yrs Null Hypothesis • HA: μ ≠25.4 yrs Alternative Hypothesis Now suppose that σis unknown, and n < 30. Previously… σ= 1.5 yrs, n= 400, statistically significant at  = .05 Example:n = 16, s = 1.22 yrs • standard error (estimate) = • .025 critical value = t15, .025

  15. Lecture Notes Appendix…

  16. Given: X = Age at first birth~ N(μ, σ) • H0: μ = 25.4 yrs Null Hypothesis • HA: μ ≠25.4 yrs Alternative Hypothesis Now suppose that σis unknown, and n < 30. Previously… σ= 1.5 yrs, n= 400, statistically significant at  = .05 Example:n = 16, s = 1.22 yrs 95% margin of error = (2.131)(0.305 yrs) = 0.65 yrs • standard error (estimate) = = 2.131 • .025 critical value = t15, .025 95% Confidence Interval = (25.9 – 0.65, 25.9 + 0.65) = (25.25, 26.55) yrs p-value = • Test Statistic:

  17. Lecture Notes Appendix…

  18. Given: X = Age at first birth~ N(μ, σ) • H0: μ = 25.4 yrs Null Hypothesis • HA: μ ≠25.4 yrs Alternative Hypothesis Now suppose that σis unknown, and n < 30. Example:n = 16, s = 1.22 yrs Previously… σ= 1.5 yrs, n= 400, statistically significant at  = .05 95% margin of error = (2.131)(0.305 yrs) = 0.65 yrs • standard error (estimate) = • .025 critical value = t15, .025 = 2.131 95% Confidence Interval = CONCLUSIONS: (25.9 – 0.65, 25.9 + 0.65) = The 95% CI does contain the null value μ = 25.4. (25.25, 26.55) yrs The p-value is between .10 and .20, i.e., > .05. (Note: The R command 2 * pt(1.639, 15, lower.tail = F) gives the exact p-value as .122.) p-value = = 2 (between .05 and .10) Not statistically significant; small n gives low power! = between .10 and .20.

  19. Lecture Notes Appendix A3.3…(click for details on this section) AssumingX ~ N(, σ), test H0:  = 0vs. HA:  ≠0, at level α… AssumingX ~ N(, σ) To summarize… If the population variance  2 is known, then use it with the Z-distribution, for any n. • If the population variance  2 is unknown, then estimate it by the samplevariances 2, and use: • either T-distribution (more accurate), or the Z-distribution (easier), if n  30, • T-distribution only, if n < 30.

  20. Lecture Notes Appendix A3.3…(click for details on this section) AssumingX ~ N(, σ), test H0:  = 0vs. HA:  ≠0, at level α… AssumingX ~ N(, σ) To summarize… If the population variance  2 is known, then use it with the Z-distribution, for any n. • If the population variance  2 is unknown, then estimate it by the samplevariances 2, and use: • either T-distribution (more accurate), or the Z-distribution (easier), if n  30, • T-distribution only, if n < 30.

  21. How do we check that this assumption is reasonable, when all we have is a sample? And what do we do if it’s not, or we can’t tell? IF our data approximates a bell curve, then its quantiles should “line up” with those of N(0, 1). Z ~ N(0, 1) AssumingX ~ N(, σ)

  22. How do we check that this assumption is reasonable, when all we have is a sample? And what do we do if it’s not, or we can’t tell? Sample quantiles IF our data approximates a bell curve, then its quantiles should “line up” with those of N(0, 1). Z ~ N(0, 1) • Q-Q plot • Normal scores plot • Normal probability plot AssumingX ~ N(, σ)

  23. How do we check that this assumption is reasonable, when all we have is a sample? And what do we do if it’s not, or we can’t tell? IF our data approximates a bell curve, then its quantiles should “line up” with those of N(0, 1). • Q-Q plot • Normal scores plot • Normal probability plot AssumingX ~ N(, σ) qqnorm(mysample) (R uses a slight variation to generate quantiles…)

  24. How do we check that this assumption is reasonable, when all we have is a sample? And what do we do if it’s not, or we can’t tell? IF our data approximates a bell curve, then its quantiles should “line up” with those of N(0, 1). • Q-Q plot • Normal scores plot • Normal probability plot AssumingX ~ N(, σ) qqnorm(mysample) (R uses a slight variation to generate quantiles…) Formal statistical tests exist; see notes.

  25. How do we check that this assumption is reasonable, when all we have is a sample? And what do we do if it’s not, or we can’t tell? • Use a mathematical “transformation” of the data (e.g., log, square root,…). x = rchisq(1000, 15) hist(x) y = log(x) hist(y) X is said to be “log-normal.” AssumingX ~ N(, σ)

  26. How do we check that this assumption is reasonable, when all we have is a sample? And what do we do if it’s not, or we can’t tell? • Use a mathematical “transformation” of the data (e.g., log, square root,…). qqnorm(x, pch = 19, cex = .5) qqnorm(y, pch = 19, cex = .5) AssumingX ~ N(, σ)

  27. How do we check that this assumption is reasonable, when all we have is a sample? And what do we do if it’s not, or we can’t tell? • Use a mathematical “transformation” of the data (e.g., log, square root,…). • Use a “nonparametric test” (e.g., Sign Test, Wilcoxon Signed Rank Test). = Mann-Whitney Test • Makes no assumptions on the underlying population distribution! • Based on “ranks” of the ordered data; tedious by hand… • Has less power than Z-test or T-test (when appropriate)… but not bad. AssumingX ~ N(, σ) • In R, see ?wilcox.testfor details…. SEE LECTURE NOTES, PAGE 6.1-27 FOR SUMMARY OF METHODS

More Related