- Why Is Estimating Experimental Power Important?
- What Is the Difference Between Type I and Type II Errors?
- What Does the POWER for Poultry Microsoft Excel Workbook Do?
- What Are the Functions of the Individual POWER for Poultry Worksheets?
- How Are the Individual POWER for Poultry Worksheets Used?
- How Often Will Experimental Results Reflect Reality?
- What Can Be Concluded from POWER for Poultry?
- Computational Requirement
- This workbook (click here to download) runs on all modern versions of Microsoft Excel with the Visual Basic function. Macros need to be enabled. Download instructions for running the workbook here (PowerPoint).
Why Is Estimating Experimental Power Important?
A fundamental question for poultry nutrition researchers is how many replicates to use in their experiments. The question is complicated by the choice of how many birds to place in each replicate. Each replicate (the experimental unit) may be individual birds, but more often is made of pens from five to more than 1,000 birds.
Grouping birds into pens generally saves on labor. For instance, having two pens of 50 birds each may require less labor and record keeping than 10 pens of 10 birds each. And the variation in the averages of pens should increase as the number of birds in the pens decreases. That is, a pen of 50 birds should provide a more reliable sample of the genetic diversity in a pen than, say, a pen of 10 birds. For a constant number of birds in an experiment (and total feed costs), having more birds per pen decreases the total number of pens. Fewer pens mean fewer error degrees of freedom and reduced statistical power just because of the nature of F- and t- distributions used to evaluate the results.
Knowing experimental power is particularly important when the goal of the research -- the null hypothesis being tested -- is to show that products (e.g., dietary supplements) are equivalent. It is particularly important to understand the probability that treated groups are different (Type I error) and the probability of finding a difference of a certain size or magnitude if it really exists (Type II error).
What Is the Difference Between Type I and Type II Errors?
When evaluating experiments, most researchers are primarily concerned with Type I error (α): What is the probability that they will declare a difference significant when none really exists? Most researchers are satisfied if the chance of declaring differences to be real when they are not is one in 20, or P < 0.05. Poultry nutritionists should more often be concerned with another type of error, Type II error (β). This error occurs when we say something is not different when it really is. Poultry nutritionists want to know how much of a nutrient they can add before there is no longer any significant increase in response, or how much of an alternative ingredient they can add before there is no significant decrease in response. Poultry nutritionists rarely determine or formally consider Type II error. The normal standard is to be correct only four times out of five, or P > 0.80. Paradoxically, if the chance of committing a Type I error (declaring something different when it is not) is decreased by increasing the critical probability value, the chance of committing a Type II error (not declaring a real difference) is increased (Aaron and Hays, 2004).
What Does the POWER for Poultry Microsoft Excel Workbook Do?
POWER for Poultry was developed to demonstrate the statistical and economic impacts of choosing different numbers of birds per pen and different numbers of replicate pens per treatment on experimental outcomes and interpretations. The workbook uses inputted descriptive statistics of two populations (a control and a treated group) with potential numbers of birds per pen and pens per treatment to determine the probability that the treatment differences would lead to the declaration of different results (a measure of Type I error) and to show how rigorous comparisons among treatments would be (Type II error).
What Are the Functions of the Individual POWER for Poultry Worksheets?
The POWER for Poultry workbook contains several individual worksheets:
- Click on the ?Dashboard? worksheet to see a summary of inputs and outputs.
- Click on the ?Simulation (Single run)? worksheet to see results of a simulated experiment based on the specified inputs.
- Click on the ?Simulation (Multiple runs)? worksheet to see the output summaries from many simulated experiments with a normal distribution assumption of the population mean.
- Click on the ?Cost_Estimation? worksheet to calculate expected costs (Experiment Cost.xls) from experiments with different numbers of birds per pen and different numbers of pens per treatment (replicates per experiment).
- Click on the ?TC_vs_DD? worksheet to see tables and charts displaying relationships between: 1) total costs and the detectable difference, and 2) birds per pen and the detectable difference.
How Are the Individual POWER for Poultry Worksheets Used?
The ?Dashboard? worksheet contains input, run and output sections. All inputs should be entered on the ?Dashboard? worksheet (Figure 2).
- The population mean, standard deviation, Type I errors and Type II errors are input here. The example values of ?Population Mean? and ?Population Standard Deviation? are for a flock of female broilers when they were 48 days old (Shim et al., unpublished).
- The number of birds per pen and replications per treatment have to be entered. These two numbers will define the experiment you want to simulate.
- The current costs to conduct experiments have to be entered.
- Appropriate historical data should be entered under “Feed Consumption Data.” A graph is automatically adjusted and the coefficients are used to generate costs for different weights of birds.
- Lastly, there is a feed planning table to modify. This table is used to allocate feed charges for different times during the birds? life. Cost calculations will automatically be adjusted without any manual corrections to the other worksheets. For example, if you plan to use ?Starter,? ?Grower,? ?Finisher? and ?Withdrawal? feeds, just fill in the cells with the days that each feed will be fed and its cost. If you do not want to use grower feed, just empty the ?Grower? column and only enter values into the ?Starter,? ?Finisher? and ?Withdrawal? feed columns.
- The number of simulations to be run and the mean for a treated group should be entered just above 001E the ?Run? button (Figure 3). 5FEC
- Each time you want to simulate a new experiment, click the ?Run” button.
- The results of one simulated experiment are displayed on the ?Dashboard? under ?Outputs? (Figure 4).
The ?Simulation (Single run)? worksheet shows a simulation to compare two treatments with each replication (Figure 5). In the example in Figure 5, each replicate contains 15 birds and each treatment contains five replicates. For example, cell B3 contains a randomly generated individual bird weight from a normal distribution with the mean and standard deviation input.
The ?Simulation (Multiple runs)? worksheet shows multiple simulations (Figure 6). Each row shows one experiment. The cell B9 presents the average weight of one pen of randomly generated “birds.” The G9 and O9 cells present average weights of five pens each for the Control and Treated groups, respectively. The H9 and P9 cells present the corresponding standard deviations. These cells will change based on the number of replications. This worksheet also contains the average mean difference, tstatistic and P-values based on the number of replications. The last column is used to tabulate the number of individual simulations meeting the rejection rate probability. It is summed at the bottom of the page. It was an assumption for the simulation to see how accurate the result will be based on the input settings. If the mean values for the control and treated groups are identical, then the Hypothesis Rejection Rate (%) should be 5.0. In the example in Figure 4, it is 5.0. By clicking on the ?Simulation? you can see how the P-values vary based on the assumption of a normal distribution for the population mean, like in a Monte-Carlo Simulation.
The ?Cost_Estimation? worksheet contains the calculation details for experimental costs (Figure 7). Every input value was from the ?Dashboard.? Values should not be overwritten on this sheet.
The ?TC_vs_DD? worksheet displays total costs (from $366.40 to $3,505.33) and detectable differences against various numbers of birds per pen and replications, which make from 120 to 1,920 total birds (Figure 8). This worksheet contains the same graph as in the ?Dashboard? and a graph of birds per pen (from five to 80) and detectable differences.
The ?Chart 1? and ?Chart 2? worksheets show the detectable difference versus total costs (Figure 9) and the detectable difference versus birds per pen (Figure 10) that are also displayed on the ?TC_vs_DD? worksheet. These were included for printing.
How Often Will Experimental Results Reflect Reality?
There is an easy test to demonstrate that the simulation model is working properly: When the Control and Treatments are input with identical means and standard errors, the null hypothesis that there were no differences in means should have a 5 percent rejection rate (P < 0.05) (Pesti and Shim, 2012). For example, the ?Simulation (Multiple runs)? worksheet shows 5 percent rejection (Figure 6). Similar tests can be demonstrated for Type II error by running the problem with real differences between the control and treated groups and different levels of Type II error.
What Can Be Concluded from POWER for Poultry?
POWER analysis can be used to calculate the minimum sample size required so that one can be reasonably likely to detect an effect of a given size. In general, the larger the sample size, the smaller the sampling error tends to be. If the sample size is too small, there is not much point in gathering the data, because the results will tend to be too imprecise to be of much use. However, there is also a point of diminishing returns beyond which increasing sample size provides little benefit. Once the sample size is large enough to produce a reasonable level of accuracy, making it larger simply wastes time and money. Of course, each researcher has to decide what “reasonable” means and then accept that a certain percentage of the time their experiment will not yield the correct conclusions.
POWER calculations can be made whenever one performs a statistical test of a hypothesis and obtains a statistically non-significant result. Because of the very large amount of money at risk from declaring there were no significant differences in an experiment, there should be more information presented than the chance of declaring a significant difference when none exists. The detectable difference displayed on the ?Dashboard? sheet is from Zar (1981).
The number of replications required depends on the experimental design employed, the desired difference from the mean that one would like to be able to detect, and a specific probability level and type II error. The number of replications may be estimated using procedures described by Cochran and Cox (1957), Zar (1981) or Berndtson (1991).
When we assume that we use the same number of birds in an experiment (e.g., 600 total birds), the detectable differences decreases from 127.15 g to 27.78 g by going from two (300 birds per pen) to 12 (50 birds per pen) replications. This means that more replication numbers generally gives better results than more birds per pen. To estimate about 50 g detectable differences in 12 replications, 15 birds are needed. However, when replications were decreased from six to three, the number of birds per pen was increased from 40 to 185 to estimate about 50 g detectable difference.
With two replications per treatment, increasing the number of birds decreased the detectable differences very much. However, even with large numbers of birds per pen, the detectable differences were quite large (Figure 8). With 12 replications, the increase in the number of birds did not affect detectable differences nearly as much as with two replications, but did cause a steep increase in total costs.
In order to have a relatively small detectable difference with a reasonable cost, a proper trade-off study is necessary. POWER for poultry should be helpful for making the appropriate calculations and decisions.
Aaron, D.K. and V.W. Hays. 2004. How many pigs? Statistical power considerations in swine nutrition experiments. J. Anim. Sci. 82 (E. Suppl.): E245-E254.
Berndtson, W.E. 1991. A simple and reliable method for selecting or assessing the number of replicates for animal experiments. J. Anim. Sci. 69:67-76.
Cochran, W. G. and G. M. Cox. 1957. Experimental Designs, 2nd ed. John Wiley & Sons, New York. National Research Council, 1994. Nutrient Requirements of Poultry. 9th rev. ed. National Academy Press, Washington, DC.
Pesti, G. M. and M. Y. Shim. 2012. A spreadsheet to construct power curves and clarify the meaning of the word equivalent in evaluating experiments with poultry. Accepted to Poult. Sci.
SAS Institute, 2006. SAS User?s Guide: Statistics. Version 9.1.3 ed. SAS Inst. Inc., Cary, NC.
Shim, M. Y. and G. M. Pesti. 2012. Poultry Nutritionist Tool Kit. http://www.poultry.uga.edu/extension/Poultry-Nutrition.htm Accessed Aug. 2012.
Shim, M. Y., G. M. Pesti and M. Tahir. Unpublished. How many chickens? Statistical power considerations in experiments with poultry. The University of Georgia, Athens, GA 30602.
Zar, J. H. 1981. Power of statistical testing: Hypothesis about means. American Laboratory 13:102-107.
Status and Revision History
Published on Mar 31, 2013