Use of a new crop variety or production technique may dramatically increase food production in a given area. Alternatively, an innovation successful in North America may utterly fail in the tropics. The goal of “adaptive research” is to evaluate a particular innovation for its usefulness under local conditions. This technical note is written for those who want to improve the quality of their experiments but who have little or no background in statistics. It supplements the 81^{st} issue of ECHO Development Notes with stepbystep instructions on how to manually calculate statistics for the most commonly used experimental designs. A little persistence, very basic math skills, and perhaps a calculator are all you need to do the calculations. If you have a computer equipped with statistical software, doing a set of calculations by hand or with a calculator will help you to understand how to use the software and interpret the output. Examples are given for experiments in which only one factor (e.g. crop variety) is tested. A limited amount of statistical terminology is woven into the text for the benefit of those interested in further study.
Example 1: Comparison of two averages using a Ttest
A Ttest could be used to determine, for example, if the average tomato fruit weight of variety A is similar to that of variety B. Ttests can also be used for analyzing survey results. In doing a Ttest, standard deviation from an average is also calculated.
Imagine a farmer’s field in which we had randomly selected and harvested three rows each of tomato variety A and B. Table 1 contains the yield data collected from each of six rows (10 plants per row). A Ttest would tell us if there is a statistical difference between the average yield of variety A vs. that of variety B. Steps 1 to 3 below illustrate how to calculate the numbers in the bottom 3 rows of Table 1 that are then used in steps 4 to 8 for a Ttest. The “null hypothesis” (see EDN 81) tested is that the average yields are similar.
Table 1. Total fruit production with two tomato varieties. 

Total fruit weight (Kg/10 plants) 

Variety A 
Variety B 

30 
40 

25 
45 

28 
38 

Number of observed values 
3 
3 
Average of observed values 
27.67 
41.00 
Standard deviation 
2.52 
3.61 
 Count the number of observed values (weights) for each variety. In this case, three values were measured for each variety.
 Calculate the average of the three values for each variety.
Average for variety A = (30 + 25 + 28)/3 = 27.67
Average for variety B = (40 + 45 + 38)/3 = 41.00
 Calculate the “standard deviation” of the average weight for each variety. A small value for standard deviation is desirable because it tells you that the individual values (e.g. 30, 25, and 28 for variety A shown in Table 1) are all quite close to the average (27.67 for variety
A). Steps to calculate the standard deviation for variety A are as follows:
A. Subtract the average (27.67) from each of the three values (30, 25, and 28).
30 – 27.67 = 2.33
25 – 27.67 = 2.67
28 – 27.67 = 0.33
B. Square (multiply a value by itself) each difference calculated in step 3A.
2.33 X 2.33 = 5.43
2.67 X –2.67 = 7.13
0.33 X 0.33 = 0.11
C. Add up the values resulting from step 3B.
5.43 + 7.13 + 0.11 = 12.67
D. Subtract 1 from the number of observed values (3).
3 – 1 = 2.
E. Divide the sum from step 3C (12.67) by the number from step 3D (2).
12.67/2 = 6.34 (this number is called the “sample variance”)
F. The standard deviation for variety A is equal to the square root of 6.34.
√6.34 = 2.52
G. Repeat steps 3A to 3F to calculate the standard deviation (3.61) for variety B.
 Obtain what is called the “common variance”.
A. Subtract 1 from the number of values for variety A. Recall from step 1 that 3 values were obtained for each of the varieties.
3 – 1 = 2
B. Multiply the result of step 4A (2) by the square of the standard deviation (2.52; see step 3) for variety A.
2 X 2.52^{2} = 12.70
C. Repeat steps 4A and 4B for variety B. Recall that the standard deviation for variety B was 3.61 (see step 3).
Repeat of step 4A for variety B: 3 – 1 = 2
Repeat of step 4B for variety B: 2 X 3.61^{2} = 26.06
D. Add up the values calculated in steps 4B and 4C.
12.70 + 26.06 = 38.76
E. Subtract 2 from the total number of values (3 for variety A + 3 for variety B = 6).
6 – 2 = 4
F. Divide the result of step 4D by that of step 4E.
38.76/4 = 9.69 (this is the common variance)
 Calculate the “standard deviation of the difference between means” (only need to know that this is not the same thing as described for step 3) using the common variance (9.69; step 4).
A. For each variety, divide 1 by the number of values obtained for that variety. Recall that 3 values were obtained for each variety. Add the resulting numbers together.
Variety A: 1/3 = 0.33
Variety B: 1/3 = 0.33
0.33 + 0.33 = 0.66
B. Multiply the result of step 5A (0.66) by the common variance (9.69).
0.66 X 9.69 = 6.40
C. Calculate the square root of the result of step 5B (6.40).
√6.40 = 2.53 (the standard deviation of the difference between means)
 Obtain a value for “calculated T”.
A. Subtract the average weight of variety A from the average weight of variety B.
27.67 – 41.00 = 13.33
B. Divide the result of step 6A (13.33) by that of 5C (2.53).
13.33/2.53 = 5.27
Use the absolute value of –5.27, so your calculated T statistic is 5.27.
 Obtain a “critical T” value from Table 5 shown near the end of this article. The farleft column of the table contains numbers from 1 to infinity (∞) known as “degrees of freedom”. Degrees of freedom for a Ttest are calculated by subtracting 2 from the total number of values (see step 4E). In this case, therefore, we have 4 (3 + 3 – 2) degrees of freedom.
At the top of each column of Table 5, you will see significance levels ranging from 40% (0.4) down to 0.05% (0.0005). A significance level of 5% is usually considered acceptable. In choosing a significance level of 5%, we are willing to accept a 5% chance of being incorrect if we reject the null hypothesis and, thereby conclude that the average weight for variety A is different from that for variety B. Conversely, we will be 95% (100% 5%) certain of being correct in our conclusion. A significance level of 5% also means that the P (probability; see EDN 81) value is less than or equal to 0.05. Exact P values are difficult to calculate manually.
Critical T, in this case, is 2.132 intersected by the row of numbers across from 4 degrees of freedom and the column of numbers under the significance level of 0.05.
 Compare “calculated T” (5.27) vs. “critical T” (2.132) to determine if the average tomato weight for variety A differs from that for variety B. If “calculated T” is greater than “critical T”, then there is a significant difference between the two averages being tested. If “calculated T” is equal to or less than “critical T”, then the two averages are statistically similar. In Example 1, the value of 5.27 exceeds 2.132, so we reject the null hypothesis and conclude that the total tomato weight for variety A (27.67) is less than that for variety B (41.00).
NOTE: A Ttest could also be used for analyzing results of surveys in which a single factor (e.g. yield) is tested. For instance, suppose we had obtained maize yield data from farmers who A.) used fertilizer and B.) had not used fertilizer. The data would be analyzed similarly as in Example 1 by substituting farmer groups for tomato varieties. I suggest investigating the use of chisquare tests for more elaborate surveys where more than one factor is tested.
Example 2. Analysis of qualitative data with a Completely Randomized Design
In Example 1, we were simply interested in testing marketable yield of two tomato varieties. What would we need to do if we were interested in testing more than two varieties? The statistically correct way to handle this situation is a procedure called analysis of variance (ANOVA), and experiment design becomes important when doing ANOVA.
A CRD is used when the experiment is conducted in a small area such as a shade house bench or a small, uniform plot of ground. Consider the pumpkin variety trial illustrated in EDN 81, a singlefactor (crop variety) experiment in which three pumpkin varieties were tested: 1.) La Primera, 2.) Butternut, and 3.) Acorn. Each of the three varieties were evaluated (replicated) three times, so 9 (3 varieties X 3 reps) planting beds were made with each bed long enough for 6 pumpkin plants. Each of the 9 beds had an equal (random) chance of receiving either of the 3 varieties (see EDN 81 for information on how to randomize). Let us assume that the varieties were planted in the field in a CRD as shown below:
Bed 1 Acorn 
Bed 4 Acorn 
Bed 7 Butternut 
Bed 2 Butternut 
Bed 5 Butternut 
Bed 8 La Primera 
Bed 3 La Primera 
Bed 6 La Primera 
Bed 9 Acorn 
We will use the same data (Table 2A below) as in EDN 81 in testing the null hypothesis that there are no differences between average yields (20.1, 6.72, and 11.4).
Table 2A. Yield (kg fruit per 6 plants) data for tropical pumpkin variety trial. 

Variety 
Replicate 1 
Replicate 2 
Replicate 3 
Totals 
Averages 
La Primera 
21.10 
17.90 
21.20 
60.20 
20.10 
Butternut 
4.42 
2.95 
12.80 
20.17 
6.72 
Acorn 
15.20 
5.78 
13.10 
34.08 
11.40 
Grand total = 
114.45 
In doing ANOVA, an “F” instead of a “T” value will be calculated by doing the following steps:
 For each variety, sum the values obtained in the three replicates (beds).
La Primera: 21.10 + 17.90 + 21.20 = 60.20
Butternut: 4.42 + 2.95 + 12.80 = 20.17
Acorn: 15.20 + 5.78 + 13.10 = 34.08
See these numbers entered in the “Totals” column of Table 2A.
 Sum the numbers obtained in step 1 (60.20, 20.17, and 34.08).
60.20 + 20.17 + 34.08 = 114.45 (shown as “Grand total” in Table 2A)
 Calculate “total sum of squares” abbreviated as SS_{T}.
A. Square each of the 9 values and add the resulting numbers together.
21.1^{2} + 4.42^{2} + 15.2^{2} + 17.9^{2} + 2.95^{2} +5.78^{2} + 21.2^{2} + 12.8^{2} + 13.1^{2} = 1843.20
B. Square the grand total (114.45 in Table 2A) and divide the resulting number by the total number of values (9).
114.45^{2}/9 = 1455.42
C. Subtract the result of step 3B (1455.42) from that of step 2A (1843.20).
1843.20 – 1455.42 = 387.78
Thus, SS_{T} = 387.78
 Calculate treatment sum of squares abbreviated as SS_{Treatments}
A. Square the 3 values (60.2, 20.17, and 34.08; not the “Grand Total”) in the “Totals” column of Table 2A; see step 1 for how these were obtained. Add up the results.
60.2^{2} + 20.17^{2} + 34.08^{2} = 5192.32
B. Divide the previously obtained value (5192.3153) by the number of replicates (3).
5192.3153/3 = 1730.77
C. Subtract the value obtained in step 3B (1455.42) from that in step 4B (1730.77).
Thus, SS_{Treatments} is equal to 1730.77 – 1455.42 = 275.35
 Calculate the “error sum of squares” abbreviated as SS_{E} by subtracting the value for
SS_{Treatments} (275.35) from that for SS_{T} (387.78).
SS_{E} = 387.78 – 275.35 = 112.43
 Calculate “degrees of freedom” for SS_{Treatments}, SS_{E}, and SS_{T}. Two easilyobtained numbers identified as “a” [number of levels of a factor (“variety” in Example 2)] and “N” (total number of values) are required. Three varieties were tested, so “a” is equal to 3 (same as the number of averages). “N” is equal to 9 because there were 9 measured values in Example 2.
A. Degrees of freedom for SS_{Treatments} = a – 1
= 3 –1 = 2
B. Degrees of freedom for SS_{E} = N  a
= 9 – 3 = 6
C. Degrees of freedom for SS_{T} = N  1
= 9 – 1 = 8
 Calculate “mean squares” abbreviated as MS_{Treatments} and MS_{E. }
A. MS_{Treatments }= SS_{Treatments }/Degrees of freedom for SS_{Treatments}
= 275.35/2 =137.68
B. MS_{E }= SS_{E} /Degrees of freedom for SS_{E}
= 112.43/6 = 18.74
 Obtain a calculated F value by dividing MS_{Treatments} (137.68) by MS_{E }(18.74).
F_{calculated} is equal to 137.68/18.74 = 7.35
HINT: Make an ANOVA table (below) to help you understand steps 1 to 8.
Table 2B. Analysis of variance table describing information needed for a CRD. 

Source of variation^{Z} 
Sum of squares 
DF^{Y} 
Mean Square 
F_{Calculated} 
Between treatments 
SS_{Treatments} 
a  1 
MS_{Treatments} 
MS_{Treatments}/ MS_{E} 
Error within treatments 
SS_{E} 
N  a 
MS_{E} 

Total 
SS_{T} 
N  1 


^{Z}A CRD accounts for variation between and within treatments (varieties). 

^{Y}DF = degrees of freedom; see step 5 of Example 2 for definitions of “a” and “N”. 
Table 2C. Analysis of variance table with information for Example 2. 

Source of variation 
Sum of Squares 
DF^{Z} 
Mean Square 
F_{Calculated} 
Between treatments 
275.35 
2 
137.68 
7.35 
Error within treatments 
112.43 
6 
18.74 

Total 
387.78 
8 

^{Z}DF = degrees of freedom 
 Obtain a critical F value. This is done using an F distribution table.
A. Select an “F distribution table” corresponding to your desired significance level. Table 6 at the end of this document contains F values for a significance level of 5%.
B. Calculate degrees of freedom corresponding to the “numerator” (MS_{Treatments}) and “denominator” (MS_{E}) of “F”. The terms “numerator” and “denominator” are used because “F” is a ratio, in this case of MS_{Treatments}/MS_{E}.
1.) Degrees of freedom for “numerator” = a –1 with a equal to 3 (see step 6A)
3 – 1 = 2
2.) Degrees of freedom for “denominator” = N – a with N equal to 9 (see step 6B and a equal to 3 (see step 6A)
9 – 3 = 6.
C. F_{critical} in Table 6 is 5.14, the value intersected by the column of numbers under 2 degrees of freedom for the numerator and the row of numbers across from 6 degrees of freedom for the denominator.
 Decide if the average yields are similar to or different from one another. If F_{Calculated} is greater than F_{Critical}, then the averages are not similar. In example 2, F_{Calculated} (7.35) exceeded F_{Critical} (5.14). Therefore, with 95% certainty, we can say that not all the average marketable yields were similar.
 Make comparisons between averages. Steps 1 to 10 illustrated how to determine that differences between averages exist. This step will illustrate a simple test used to determine how the averages differ. We want to know, for instance, if the yield from La Primera differs significantly from that with Butternut. There are many tests that can be used, but one called “least significant difference” is commonly reported in literature and is calculated as follows:
A. Decide which significance level you will accept and divide that by 2. Let us use a significance level of 5% (expressed as a number, 5% is 0.05).
0.05/2 = 0.025
B. Note the degrees of freedom associated with MS_{E} in this case, 6 (see Table 2C). This number, as well as the result of 11A, is required to obtain a critical T value as described in the following step.
C. Obtain the critical T value (2.447) in Table 5 intersected by the column of numbers under a significance level of 0.025 (step 11A) and the row of numbers across from 6 degrees of freedom.
D. Multiply MS_{E} (18.74; see step 7B) by 2.
18.74 X 2 = 37.48
E. Divide the result of the previous step (37.48) by the number of values used to calculate each average (3).
37.48/3 = 12.49
F. Calculate the square root of the value obtained in step 11E (12.49).
√12.49 = 3.53
G. Multiply the result of step 11C (2.447) by that of step 11F (3.53).
2.447 X 3.53 = 8.64
In Example 2, therefore, the LSD value used to determine if any two averages are statistically different is 8.64. If the difference (absolute value) between any two averages is greater than 8.64, the two averages are not the same. For example, we would want to make the following comparisons using the average yields in Table 2:
La Primera vs. Butternut: 20.10 – 6.72 = 13.28; 13.28 > 8.64 so these averages differ.
La Primera vs. Acorn: 20.10 – 11.40 = 8.7; 8.7 > 8.64 so these averages differ.
Butternut vs. Acorn: 11.40 – 6.72 = 4.68; 4.68 < 8.64 so these averages do not differ.
Example 3. Comparison of averages when treatments are “quantitative”
In Example 2, the treatments tested are “qualitative” because one cannot measure or quantify “tomato variety”. One can, however, measure “quantitative” factors such as rate (e.g. fertilizer, plant density/spacing, temperature, or time after crop establishment). Example 3 illustrates an accepted method of characterizing yield response to a factor such as rate.
Consider an experiment in which we obtained the exact same data as shown in Example 2. Instead of testing pumpkin varieties, suppose we had tested several rates of a herbicide to find out if it can be sprayed over pumpkin plants without injuring the crop. It is important to choose rates that are “equally spaced” to avoid the need for computer software to analyze the data. For instance, the numbers 0, 25, and 75 are not equally spaced because there is a difference of 25 between 0 and 25 and a difference of 50 between 25 and 75. Suppose we had chosen rates of 0, 5, and 10 liters herbicide/plot. Table 2A would then become Table 3A below:
Table 3A. Pumpkin yield (kg fruit per 6 plants) with 3 herbicide rates. 

Herbicide (L/plot) 
Replicate 1 
Replicate 2 
Replicate 3 
Totals 
Averages 
0 
21.10 
17.90 
21.20 
60.20 
20.10 
5 
4.42 
2.95 
12.80 
20.17 
6.72 
10 
15.20 
5.78 
13.10 
34.08 
11.40 
Grand total = 
114.45 
To analyze the data in Table 3A, proceed as follows:
 Do ANOVA because there are more than two averages to compare [Note: if the only rates tested were 0 and 5, then we would only need to conduct an Ftest (Example 2) to find out if yield with the 0liter rate was less than or similar to that with the 5liter rate]. In this case, we have already done the ANOVA in steps 1 to 10 of Example 2. We know there are differences between the averages. We still need to know how pumpkin yield responded to herbicide rate.
The graph (Figure 1) below represents the data in Table 3A. Notice, for example, that there are 3 dots (data points) directly above “5” (herbicide rate). These three points correspond to values 4.42, 2.95, and 12.8 in Table 3A. It appears that there are only 2 points above the “0” rate because two values (21.10 and 21.20) were almost the same. The object is to define a line that can be drawn through the points in Figure 1.
 Determine how many types/shapes of lines are possible by subtracting 1 from the number of rates (3 – 1 = 2). Thus, two line shapes will be tested. The most commonly used line shapes are linear (always the first possibility) and quadratic (always the second possibility). There are more line shapes, but these do not always make biological sense. We want to know if yields decreased linearly (Figure 2) or quadratically (Figure 3) with increased herbicide.
 Obtain “orthogonal coefficients” for each line shape from Table 7 on page 20. The rates (0, 5, and 10 liters/plot) we used are orthogonal because they were equally spaced (difference of 5 liters between each rate). If they had not been equally spaced, we could not use the coefficients in Table 7. Find coefficients for 3 levels of a factor (rate in this case) because we tested 3 rates.
Coefficients for testing significance of a linear line: 1, 0, 1
Coefficients for testing significance of a quadratic line: 1, –2, 1
NOTE: The three numbers for each line shape add up to zero. Making sure your coefficients add up to zero is a quick way to be sure you have the correct numbers.
 Calculate the “contrast estimate” for a linear and quadratic line using the coefficients obtained in step 3 and the average yield (20.1, 6.72, and 11.4) obtained with each herbicide rate. Abbreviations below are “AY” for average yield, “w/” for with, and “L” for liters.
A. For a linear line = (1 X AY w/ 0 L) + (0 X AY w/ 5 L) + (1 X AY w/ 10 L)
NOTE: The first number in each set of parentheses is one of the three coefficients listed in step 3 for a linear line. Notice that they are used in the order that they appear in step 3: the first coefficient (1) is multiplied by the first herbicide rate (0 L); the second coefficient (0) is multiplied by the second rate (5 L); and the third coefficient is multiplied by the third rate (10 L). Thus, the contrast estimate for a linear line is calculated as follows:
= (1 X 20.1) + (0 X 6.72) + (1 X 11.4)
= 20.1 + 0 + 11.4
= 8.7
B. For a quadratic line = (1 X AY w/ 0 L) + (2 X AY w/ 5 L) + (1 X AY w/ 10 L)
NOTE: The first number in each set of parentheses is now the appropriate coefficient for testing the significance of a quadratic line. Plugging in values for average yield (AY):
= (1 X 20.1) + (2 X 6.72) + (1 X 11.4)
= 20.1 + (13.44) + 11.4
= 20.1 – 13.44 + 11.4
= 18.06
 Calculate the sum of squares for each line shape using the formula:
Sum of squares = (number of replications per treatment X (contrast estimate^{2})/sum of squares of the coefficients
There were 3 replications per treatment, and contrast estimates were calculated in the previous step. “Sum of squares of the coefficients” is found by squaring each coefficientand summing the resulting values. The sum of squares of the coefficients is equal to 2 (1^{2} + 0^{2} + 1^{2}) for testing the significance of the linear line and 6 (1^{2} + 2^{2} + 1^{2}) for testing the significance of the quadratic line. Therefore, the sum of squares for the linear and quadratic line are calculated as follows:
A. Sum of squares for linear line = (3 X –8.7^{2})/2
= (3 X 75.69)/2
= 113.54 (we will call this value SS_{1})
B. Sum of squares for quadratic line = (3 X 18.06^{2})/6
= (3 X 326.16)/6
= 163.08 (we will call this value SS_{2})
 For each line shape, calculate a value for “F” by dividing the sum of squares (see step 5A for linear line; step 5B for quadratic line) step 5 by MS_{E} (18.74 used for both line shapes) as already calculated in Example 2 and shown in Tables 2B and 2C):
A. F_{Calculated} for the linear line: 113.54/18.74 = 6.06
B. F_{Calculated} for the quadratic line: 163.08/18.74 = 8.70
Tables 3B and 3C illustrate how to make an ANOVA table with results of steps 1 to 6.:
Table 3B. Analysis of variance table describing information needed for Example 3. 

Source of variation^{Z} 
Sum of squares 
DF^{Y} 
Mean Square 
F_{Calculated} 
Between treatments 
SS_{Treatments} 
a  1 
MS_{Treatments} 
MS_{Treatments}/ MS_{E} 
Linear line 
SS_{1} 
1 
Same as SS_{1} 
SS_{1}/ MS_{E} 
Quadratic line 
SS_{2} 
1 
Same as SS_{2} 
SS_{2}/ MS_{E} 
Error within treatments 
SS_{E} 
N  a 
MS_{E} 

Total 
SS_{T} 
N  1 


^{Z}A CRD accounts for variation between and within treatments (varieties). 

^{Y}DF = degrees of freedom; see step 5 of Example 2 for definitions of “a” and “N”. 
Table 3C. Analysis of variance table with information for Example 2. 

Source of variation 
Sum of Squares 
DF^{Z} 
Mean Square 
F_{Calculated} 
Between treatments 
275.35 
2 
137.68 
7.35 
Linear line 
113.54 
1 
113.54 
6.06 
Quadratic line 
163.08 
1 
163.08 
8.70 
Error within treatments 
112.43 
6 
18.74 

Total 
387.78 
8 

^{Z}DF = degrees of freedom 
 Determine a critical F value for a significance level of 5%. This value will be used for testing the significance of both line shapes.
A. Note degrees of freedom for the numerator (SS_{1 or 2}) and denominator (MS_{E}) of F.
1.) Degrees of freedom for the numerator, regardless of which line shape is being tested, is equal to 1 (see DF for SS_{1} and SS_{2} in Table 3C).
2.) Degrees of freedom for the denominator corresponds to that for MS_{E }in this case, degrees of freedom are equal to 6 (see Table 3C).
B. Find the critical F value of 5.99 in Table 6. It is intersected by the column of numbers under 1 degree of freedom for the numerator (SS_{1 or 2}) and the row of numbers across from 6 degrees of freedom for the denominator (MS_{E}).
8. Determine if the linear and quadratic lines are significant. A line shape is significant if F_{Calculated} exceeds F_{Critical}. In this example, both 6.06 (linear) and 8.70 (quadratic) exceed F_{Critical} of 5.99, so both lines are significant. Which one do we accept? Usually, the highestorder significant line (quadratic in this case) is reported. We now have statistical justification for a quadratic line (Figure 3), indicating that the herbicide was toxic to pumpkin, but yield loss changed little between 5 and 10 liters/plot. The curve in Figure 3 indicates a slight but unlikely increase in yield with an increase in herbicide rate from 5 to 10 liters, but remember that these data were from a variety instead of a herbicide trial.
Example 4: Randomized Complete Block Design
The completely randomized design (CRD) is appropriate for experiments conducted in a laboratory, on a greenhouse or shade house bench, or within a small plot of ground. The reason for this is that, within a small area, environmental factors such as light, temperature, and moisture are usually quite uniform.
A variation of the CRD called the randomized complete block design (RCBD) is often the design of choice for outdoor experiments because soil conditions are seldom uniform throughout a field. If you suspect that conditions may not be uniform, even in a greenhouse, use the RCBD.
In a RCBD, reps are grouped into equalsized portions of space called “blocks”. The field should be “blocked” to minimize variability in soil conditions within each block. The number of blocks is equal to the number of reps.
Consider the treatments (pumpkin varieties) used in Example 2. Suppose this experiment had been carried out on sloped land. Figure 4A illustrates how not to block the reps; the area for each rep contains a large amount of variation in elevation and, thus, soil moisture. Blocks of reps should be arranged perpendicular to the direction of the gradient as illustrated in Figure 4B.
Figure 4A. Ineffective blocking of replicates. 

Primera A 
AcornA 
ButternutA 
AcornB 
PrimeraB 
ButternutB 
ButternutC 
Primera C 
AcornC 
Higher elevation > Lower elevation 

West > East 

Rep A in block indicated by horizontal lines Rep B in block indicated by diagonal lines Rep C in block indicated by hatched lines 
Figure 4B. Correct blocking of replicates. 

Primera A 
AcornB 
ButternutC 
AcornA 
PrimeraB 
PrimeraC 
ButternutA 
ButternutB 
AcornC 
Higher elevation > Lower elevation 

West > East 

Rep A in block indicated by horizontal lines Rep B in block indicated by diagonal lines Rep C in block indicated by hatched lines 
Treatments are randomly assigned to the rows within each block. For instance, in Figure 4B, ‘Primera’ was planted in the topmost row in block A and in the middle row in blocks B and C.
Suppose we had conducted the experiment for Example 2 using a RCBD. The data, as shown in Table 4A below, are the same as in Example 2, but now we can account for error between reps.
Table 4A. Yield data for tropical pumpkin variety trial. 

Replication (Block) 

Variety 
1 
2 
3 
Totals 
Averages 
La Primera 
21.10 
17.90 
21.20 
60.20 
20.10 
Butternut 
4.42 
2.95 
12.80 
20.17 
6.72 
Acorn 
15.20 
5.78 
13.10 
34.08 
11.40 
Block totals→ 
40.72 
26.63 
47.1 
114.45 (grand total) 
As in Example 2, the hypothesis being tested is that the average yields (20.1, 6.72, and 11.4) are similar. Totals (of yield in each rep) and averages for each variety are the same as in Table 2A. Remaining steps are as follows:
 Obtain block and grand totals (see Table 4A). Under each block in Table 4A, add up the three yields (e.g. 21.1 + 4.42 + 15.2 = block total for block 1; do the same for blocks 2 and 3). Sum the numbers in the “Totals” column of Table 4A to obtain the grand total.
 Calculate total sum of squares (SS_{T}) in the same manner as in step 3 of Example 2.
Thus, SS_{T} = 387.78
 Calculate treatment sum of squares (SS_{Treatments}) in the same manner as in step 4 of Example 2.
Thus, SS_{Treatments} = 275.35
 Calculate “block sum of squares” (SS_{Blocks}).
A. Square each block total and sum the resulting values:
40.72^{2} + 26.63^{2} + 47.1^{2} = 4585.69
B. Divide the result of step 4A (4585.69) by the number of treatments (3).
4585.69/3 = 1528.56
C. Square the grand total (114.45) and divide the result by the total number of values (9).
114.45^{2}/9 = 1455.42
D. Subtract the result of step 4C (1455.42) from that of step 4B(1528.56).
SS_{Blocks} is equal to 1528.56 – 1455.42 = 73.14
 Calculate error sum of squares (SS_{E}) using the formula: SS_{E} = SS_{T} – SS_{Treatments} – SS_{Blocks}
SS_{E} = 387.78  275.35 – 73.14
= 39.29
 Calculate degrees of freedom for SS_{Treatments}, SS_{Blocks}, SS_{E}, and SS_{T}. As in Example 2, numbers identified as a and N are required. The number of levels of the factor (variety) is “a”. The total number of observations is “N”. Three varieties were tested, so “a” is equal to 3 (same as the number of averages). “N” is equal to 9 because there were a total of 9 measured values. In addition, the number of blocks (3 in this case) denoted as “b” is needed.
A. Degrees of freedom for SS_{Treatments} = a – 1
= 3 – 1 = 2
B. Degrees of freedom for SS_{Blocks} = b –1
= 3 –1 = 2
C. Degrees of freedom for SS_{E} = (a – 1) X (b – 1)
= (3 – 1) X (3 – 1)
= 2 X 2 = 4
D. Degrees of freedom for SS_{T} = N  1
= 9 – 1 = 8
 Calculate means squares abbreviated as MS_{Treatments} MS_{Blocks}, and MS_{E. }
A. MS_{Treatments }= SS_{Treatments }/Degrees of freedom for SS_{Treatments}
= 275.35/2 = 137.68
B. MS_{Blocks} = SS_{Blocks}/Degrees of freedom for SS_{Blocks}
= 73.14/2 = 36.57
C. MS_{E }= SS_{E} /Degrees of freedom for SS_{E}
= 39.29/4 = 9.82
 Obtain a calculated F value for treatments: F_{Calculated} = MS_{Treatments}/ MS_{E}
= 137.68/9.82 = 14.02
 Obtain a calculated F value for blocks: F_{Calculated} = MS_{Blockss}/ MS_{E}
= 36.57/9.82 = 3.72
Below is the ANOVA with formulas (Table 4B) and abovecalculated values (Table 4C).
Table 4B. Analysis of variance table describing information needed for a RCBD. 

Source of variation^{Z} 
Sum of Squares 
DF^{Y} 
Mean Square 
F_{Calculated} 
Between treatments 
SS_{Treatments} 
a  1 
MS_{Treatments} 
MS_{Treatments}/MS_{E} 
Between blocks 
SS_{Blocks} 
b  1 
MS_{Blocks} 
MS_{Blocks}/MS_{E} 
Error within treatments 
SS_{E} 
(a –1) (b – 1) 
MS_{E} 

Total 
SS_{T} 
N  1 

^{Z}A RCBD improves upon a CRD by accounting for variation between reps (grouped in blocks). 

^{Y}DF = degrees of freedom; see step 6 of Example 4 for definitions of “a”, “b”, and “N”. 
Table 4C. Analysis of variance table with information for Example 4. RCBD. 

Source of variation 
Sum of Squares 
DF^{Z} 
Mean Square 
F_{Calculated} 
Between treatments 
275.35 
2 
137.68 
14.02 
Between blocks 
73.14 
2 
36.57 
3.72 
Error within treatments 
39.29 
4 
9.82 

Total 
387.78 
8 

^{Z}DF = degrees of freedom; see step 6 of Example 4 for definitions of “a”, “b”, and “N”. 
 Obtain a critical F value for treatments with a significance level of 5%.
A. Determine degrees of freedom for the “numerator” (MS_{Treatments}) and “denominator” (MS_{E}) of F.
1.) Degrees of freedom for “numerator” = a – 1 with a equal to 3 (see step 6A).
3 – 1 = 2
2.) Degrees of freedom for the “denominator” = (a – 1) X (b – 1) with a equal to 3 and b equal to 3 (see step 6C)
(3 1) X (3 – 1) = (2 X 2) = 4
B. Find the critical F value (6.94) in Table 6 that is intersected by the column of numbers under 2 degrees of freedom for the numerator and the row of numbers across from 4 degrees of freedom for the denominator.
 Obtain a critical F value for blocks with a significance level of 5%.
A. Determine degrees of freedom for the “numerator” (MS_{Blocks}) and “denominator” (MS_{E}) of F.
1.) Degrees of freedom for “numerator” = b – 1 with b equal to 3 (see step 6B).
3 – 1 = 2
2.) Degrees of freedom for the “denominator” = (a – 1) X (b – 1) with a equal to 3 and b equal to 3 (see step 6C)
(3 1) X (3 – 1) = (2 X 2) = 4
B. Find the critical F value (6.94) in Table 6 that is intersected by the column of numbers under 2 degrees of freedom for the numerator and in the row of numbers across from 4 degrees of freedom for the denominator. In this instance, the critical F values for treatment and blocks are the same (6.94) because the number of treatments was the same as that of blocks.
 Decide whether or not there are differences between the averages. If F_{Calculated} is greater than F_{Critical}, for treatments or blocks, then the averages are not similar.
A. Significance of treatments for which F_{Calculated} was 14.02 and F_{Critical} was 6.94.
14.02 > 6.94 so average yields of treatments differed significantly.
B. Significance of blocks for which F_{Calculated} was 3.72 and F_{Critical} was 6.94
3.72 < 6.94 so average yields of blocks were similar. Assuming we blocked correctly (see Figures 4 and 5), this tells us that the field was relatively uniform. If yields had differed significantly between blocks, it would confirm that we had done a good job of blocking the reps.
 Make comparisons between averages by obtaining an LSD value as explained in step 11 of
Example 2. Notice that the MS_{E} in Example 4 is 9.82 instead of 18.74 as in Example 2. Consequently, use 4 (corresponding to MS_{E} in Table 4C) instead of 6 (corresponding to MS_{E} in Table 2C) degrees of freedom in obtaining the critical T statistic of 2.776. This reduces the LSD value from 8.64 (Example 2) to 7.11 (Example 4). Thus, with the RCBD, there is a greater chance that differences between means will be detected. As seen below, however, the slightly smaller LSD value of 7.11 did not change conclusions relative to those in Example 2.
La Primera vs. Butternut: 20.1 – 6.72 = 13.28; 13.28 > 7.11 so these averages differ.
La Primera vs. Acorn: 20.1 – 11.4 = 8.7; 8.7 > 7.11 so these averages differ.
Butternut vs. Acorn: 11.4 – 6.72 = 4.68; 4.68 < 7.11 so these averages do not differ.
NOTE: If the treatments in Example 4 had been “quantitative” as in Example 3, the same polynomial contrasts as in Example 3 would be used to test if the resulting line shape was linear or quadratic. These contrasts would be calculated in the same manner as in Example 3 with an MS_{E} value of 9.82 (Example 4) instead of 18.74 (Example 3).
For more information…
I suggest reading the VITA publication (ISBN: 0866190392) by G. Stuart Pettygrove entitled “How to Perform an Agricultural Experiment”. It is available in English or Spanish for $7.25 plus shipping. In addition to the CRD and RCBD designs discussed in this technical note, it also contains information on how to set up (not analyze results) experiments using “latin square” and “split plot” designs. To order, write to: Pact Publications, 1200 18^{th} Street NW, Washington, DC 20036. If you have internet access, you can also order the publication online. Go to “www.pactpublications.com” and enter “agricultural experiment” in the search bar. Below, I have listed several additional websites:
For calculating standard deviation: http://www.webmath.com/ (click on “standard deviation” at the bottom of the web page)
For a T test: http://home.clara.net/sisa/ (you will need the standard deviation for each of the two averages you are comparing the abovementioned website can be used for this)
For ANOVA: http://www.economics.pomona.edu/StatSite/SSP.htmlno financial cost http://members.aol.com/rcknodt/pubpage.htm $22.00 after a twomonth trial
For graphing quantitative data: see http://curveexpert.webhop.biz/ for a $40.00 software package that will allow you to fit and graph many line shapes.
Table 5. Critical T Values for conducting a Ttest. 

Significance Level 

0.4 
0.25 
0.1 
0.05 
0.025 
0.01 
0.005 
0.0005 

Degrees of Freedom 
1 
0.325 
1.000 
3.078 
6.314 
12.706 
31.821 
63.657 
636.619 
2 
0.289 
0.816 
1.886 
2.920 
4.303 
6.965 
9.925 
31.599 

3 
0.277 
0.765 
1.638 
2.353 
3.182 
4.541 
5.841 
12.924 

4 
0.271 
0.741 
1.533 
2.132 
2.776 
3.747 
4.604 
8.610 

5 
0.267 
0.727 
1.476 
2.015 
2.571 
3.365 
4.032 
6.869 

6 
0.265 
0.718 
1.440 
1.943 
2.447 
3.143 
3.707 
5.959 

7 
0.263 
0.711 
1.415 
1.895 
2.365 
2.998 
3.499 
5.408 

8 
0.262 
0.706 
1.397 
1.860 
2.306 
2.896 
3.355 
5.041 

9 
0.261 
0.703 
1.383 
1.833 
2.262 
2.821 
3.250 
4.781 

10 
0.260 
0.700 
1.372 
1.812 
2.228 
2.764 
3.169 
4.587 

11 
0.260 
0.697 
1.363 
1.796 
2.201 
2.718 
3.106 
4.437 

12 
0.259 
0.695 
1.356 
1.782 
2.179 
2.681 
3.055 
4.318 

13 
0.259 
0.694 
1.350 
1.771 
2.160 
2.650 
3.012 
4.221 

14 
0.258 
0.692 
1.345 
1.761 
2.145 
2.624 
2.977 
4.141 

15 
0.258 
0.691 
1.341 
1.753 
2.131 
2.602 
2.947 
4.073 

16 
0.258 
0.690 
1.337 
1.746 
2.120 
2.583 
2.921 
4.015 

17 
0.257 
0.689 
1.333 
1.740 
2.110 
2.567 
2.898 
3.965 

18 
0.257 
0.688 
1.330 
1.734 
2.101 
2.552 
2.878 
3.922 

19 
0.257 
0.688 
1.328 
1.729 
2.093 
2.539 
2.861 
3.883 

20 
0.257 
0.687 
1.325 
1.725 
2.086 
2.528 
2.845 
3.850 

21 
0.257 
0.686 
1.323 
1.721 
2.080 
2.518 
2.831 
3.819 

22 
0.256 
0.686 
1.321 
1.717 
2.074 
2.508 
2.819 
3.792 

23 
0.256 
0.685 
1.319 
1.714 
2.069 
2.500 
2.807 
3.768 

24 
0.256 
0.685 
1.318 
1.711 
2.064 
2.492 
2.797 
3.745 

25 
0.256 
0.684 
1.316 
1.708 
2.060 
2.485 
2.787 
3.725 

26 
0.256 
0.684 
1.315 
1.706 
2.056 
2.479 
2.779 
3.707 

27 
0.256 
0.684 
1.314 
1.703 
2.052 
2.473 
2.771 
3.690 

28 
0.256 
0.683 
1.313 
1.701 
2.048 
2.467 
2.763 
3.674 

29 
0.256 
0.683 
1.311 
1.699 
2.045 
2.462 
2.756 
3.659 

30 
0.256 
0.683 
1.310 
1.697 
2.042 
2.457 
2.750 
3.646 

∞ 
0.253 
0.674 
1.282 
1.645 
1.960 
2.326 
2.576 
3.291 

Used with permission from Statsoft Inc., http://www.statsoft.com/textbook/stathome.html 
Table 6. Critical “F” values for a significance Level of 5%^{a}. 

Degrees of Freedom for the Numerator 

1 
2 
3 
4 
5 
6 
7 
8 
9 
10 
12 
15 
20 

Degrees of Freedom for the Denominator 
1 
161.45 
199.50 
215.71 
224.58 
230.16 
233.99 
236.77 
238.88 
240.54 
241.88 
243.91 
245.95 
248.01 
2 
18.51 
19.00 
19.16 
19.25 
19.30 
19.33 
19.35 
19.37 
19.38 
19.40 
19.41 
19.43 
19.45 

3 
10.13 
9.55 
9.28 
9.12 
9.01 
8.94 
8.89 
8.85 
8.81 
8.79 
8.74 
8.70 
8.66 

4 
7.71 
6.94 
6.59 
6.39 
6.26 
6.16 
6.09 
6.04 
6.00 
5.96 
5.91 
5.86 
5.80 

5 
6.61 
5.79 
5.41 
5.19 
5.05 
4.95 
4.88 
4.82 
4.77 
4.74 
4.68 
4.62 
4.56 

6 
5.99 
5.14 
4.76 
4.53 
4.39 
4.28 
4.21 
4.15 
4.10 
4.06 
4.00 
3.94 
3.87 

7 
5.59 
4.74 
4.35 
4.12 
3.97 
3.87 
3.79 
3.73 
3.68 
3.64 
3.57 
3.51 
3.44 

8 
5.32 
4.46 
4.07 
3.84 
3.69 
3.58 
3.50 
3.44 
3.39 
3.35 
3.28 
3.22 
3.15 

9 
5.12 
4.26 
3.86 
3.63 
3.48 
3.37 
3.29 
3.23 
3.18 
3.14 
3.07 
3.01 
2.94 

10 
4.96 
4.10 
3.71 
3.48 
3.33 
3.22 
3.14 
3.07 
3.02 
2.98 
2.91 
2.85 
2.77 

11 
4.84 
3.98 
3.59 
3.36 
3.20 
3.09 
3.01 
2.95 
2.90 
2.85 
2.79 
2.72 
2.65 

12 
4.75 
3.89 
3.49 
3.26 
3.11 
3.00 
2.91 
2.85 
2.80 
2.75 
2.69 
2.62 
2.54 

13 
4.67 
3.81 
3.41 
3.18 
3.03 
2.92 
2.83 
2.77 
2.71 
2.67 
2.60 
2.53 
2.46 

14 
4.60 
3.74 
3.34 
3.11 
2.96 
2.85 
2.76 
2.70 
2.65 
2.60 
2.53 
2.46 
2.39 

15 
4.54 
3.68 
3.29 
3.06 
2.90 
2.79 
2.71 
2.64 
2.59 
2.54 
2.48 
2.40 
2.33 

16 
4.49 
3.63 
3.24 
3.01 
2.85 
2.74 
2.66 
2.59 
2.54 
2.49 
2.42 
2.35 
2.28 

17 
4.45 
3.59 
3.20 
2.96 
2.81 
2.70 
2.61 
2.55 
2.49 
2.45 
2.38 
2.31 
2.23 

18 
4.41 
3.55 
3.16 
2.93 
2.77 
2.66 
2.58 
2.51 
2.46 
2.41 
2.34 
2.27 
2.19 

19 
4.38 
3.52 
3.13 
2.90 
2.74 
2.63 
2.54 
2.48 
2.42 
2.38 
2.31 
2.23 
2.16 

20 
4.35 
3.49 
3.10 
2.87 
2.71 
2.60 
2.51 
2.45 
2.39 
2.35 
2.28 
2.20 
2.12 

21 
4.32 
3.47 
3.07 
2.84 
2.68 
2.57 
2.49 
2.42 
2.37 
2.32 
2.25 
2.18 
2.10 

22 
4.30 
3.44 
3.05 
2.82 
2.66 
2.55 
2.46 
2.40 
2.34 
2.30 
2.23 
2.15 
2.07 

23 
4.28 
3.42 
3.03 
2.80 
2.64 
2.53 
2.44 
2.37 
2.32 
2.27 
2.20 
2.13 
2.05 

24 
4.26 
3.40 
3.01 
2.78 
2.62 
2.51 
2.42 
2.36 
2.30 
2.25 
2.18 
2.11 
2.03 

25 
4.24 
3.39 
2.99 
2.76 
2.60 
2.49 
2.40 
2.34 
2.28 
2.24 
2.16 
2.09 
2.01 

26 
4.23 
3.37 
2.98 
2.74 
2.59 
2.47 
2.39 
2.32 
2.27 
2.22 
2.15 
2.07 
1.99 

27 
4.21 
3.35 
2.96 
2.73 
2.57 
2.46 
2.37 
2.31 
2.25 
2.20 
2.13 
2.06 
1.97 

28 
4.20 
3.34 
2.95 
2.71 
2.56 
2.45 
2.36 
2.29 
2.24 
2.19 
2.12 
2.04 
1.96 

29 
4.18 
3.33 
2.93 
2.70 
2.55 
2.43 
2.35 
2.28 
2.22 
2.18 
2.10 
2.03 
1.94 

30 
4.17 
3.32 
2.92 
2.69 
2.53 
2.42 
2.33 
2.27 
2.21 
2.16 
2.09 
2.01 
1.93 

40 
4.08 
3.23 
2.84 
2.61 
2.45 
2.34 
2.25 
2.18 
2.12 
2.08 
2.00 
1.92 
1.84 

60 
4.00 
3.15 
2.76 
2.53 
2.37 
2.25 
2.17 
2.10 
2.04 
1.99 
1.92 
1.84 
1.75 

120 
3.92 
3.07 
2.68 
2.45 
2.29 
2.18 
2.09 
2.02 
1.96 
1.91 
1.83 
1.75 
1.66 

∞ 
3.84 
3.00 
2.60 
2.37 
2.21 
2.10 
2.01 
1.94 
1.88 
1.83 
1.75 
1.67 
1.57 

^{a}Portion of table used with permission from StatSoft, Inc., http://www.statsoft.com/textbook/stathome.html 
Table 7. Linear and quadratic line coefficients for up to 7 levels of a factor such as rate^{a}. 

Number of levels 
Line 
Coefficients 

3 
Linear 
1 
0 
1 





Quadratic 
1 
2 
1 




4 
Linear 
3 
1 
1 
3 




Quadratic 
1 
1 
1 
1 



5 
Linear 
2 
1 
0 
1 
2 



Quadratic 
2 
1 
2 
1 
2 


6 
Linear 
5 
3 
1 
1 
3 
5 


Quadratic 
5 
1 
4 
4 
1 
5 

7 
Linear 
3 
2 
1 
0 
1 
2 
3 

Quadratic 
5 
0 
3 
4 
3 
0 
5 
http://www.psych.yorku.ca/lab/sas/sasanova.htm. 