Sample size and power with the completely randomized designRead the sections on power and sample size (pages 37-39 and page 85) in Carl Schwarz's notes, Chapter 6. Some basic ideas: The significance level of a test protects against incorrect rejection of the null hypothesis of no treatment effect. If the null hypothes is true, we want the probability of rejecting it to be small, specifically as small as the specified significance level. The significance level of the test is the probability of rejecting the null hypothesis when it is true. But if the null hypothesis is false and there really is some effect of the treatments, then we want to reject the null hypothesis in favor of the alternative. The probability of rejecting the null hypothesis when it is false is called the power of the test. To increase the power, one may use larger sample sizes or find a more effective design. If the null hypothesis is true, the test statistic has one distrubution, such as the t or F distribution or randomization distribution. If the null hypothesis is false, however, then one or more of the treatments has an effect and the same test statistic has a different distribution. In the case of the t and F statistics under the standard model assumptions, the distribution of the test statistic when the alternative hypothesis is true is the non-central t or F distribution respectively. The distribution of the test statistic when there is some treatment effect, depends on how big the effect is, or how much the effects of different treatments differ one from another. With the completely randomized design, the power is maximized by assigning the same number of units to each treatment. To calculate the power of a test for the completely randomized design with given sample sizes, one must specify (1) the required significance level, (2) the effect size that needs to be detected, (3) the common variance of the response variables, and (4) the sample size for each group. To calculate the sample sizes needed to achieve a given power, one must specify (1) the required significance level, (2) the effect size that needs to be detected, (3) the common variance of the response variables, and (4) the required power. Learning goals: Understand the concept of power and how it is related to sample size and design. Understand the difference between the distribution that a test statistic would have if the null hypothesis is true and the distribution it would have if the alternative is true, and how this depends on the size or pattern of the treatment effect. Know what must be specified and what must be estimated to compute power or required sample size for an experiment. Know how to obtain these values with standard statistical software. Example: Insect Spray Experiment
With the insect spray experiment, the analysis of variance reject the
null hypothesis of no treatment effect very strongly. Now we'll
ask, with differences so large among the treatment means, would it be
possible to use smaller sample sizes and still have a power
(probability of rejecting) of .90, while holding the significance level
at .05? Then we'll ask, with the sample sizes we had the the
treatment effects so large, what was the power? |