17.2 Chi-Square Tests

The chi-square test for goodness of fit is a nonparametric hypothesis test used when there is one nominal variable.

The chi-square test for independence is a nonparametric hypothesis test used when there are two nominal variables.

Hot-hand research has moved from individual performance to team performance. For example, after opening the 2013 season by hanging on to win a wild game against Buffalo, Coach Urban Meyer of the Ohio State Buckeyes tried to explain that, “Momentum is an amazing thing in college football. The more mature your team is, momentum’s only about seven points…. We were up 22–0, I believe, when it turned upside-down” (Maks, 2013). But was Coach Meyer just seeing Elvis in a potato chip when he perceived a shift in his team momentum? It is a testable idea and this chapter describes what is being tested in two common kinds of chi-square statistical tests: (1) the chi-square test for goodness of fit, a nonparametric hypothesis test that is used when there is one nominal variable; (2) the chi-square test for independence, a nonparametric hypothesis test that is used when there are two nominal variables. Both chi-square tests involve the by-now familiar six steps of hypothesis testing.

Both chi-square tests use the chi-square statistic: χ2. The chi-square statistic is based on the chi-square distribution. As with t and F distributions, there are also several chi-square distributions, depending on the degrees of freedom. After we introduce chi-square tests, we’ll introduce several ways of determining the size of a finding—by calculating an effect size, graphing the finding, or determining relative risk.

465

MASTERING THE CONCEPT

17.2: When we only have nominal variables, we use the chi-square statistic. Specifically, we use a chi-square test for goodness of fit when we have one nominal variable, and we use a chi-square test for independence when we have two nominal variables.

Chi-Square Test for Goodness of Fit

The chi-square test for goodness of fit calculates a statistic based on just one variable. There is no independent variable or dependent variable, just one categorical variable with two or more categories into which participants are placed. In fact, the chi-square test for goodness of fit received its name because it measures how good the fit is between the observed data in the various categories of a single nominal variable and the data we would expect according to the null hypothesis. If there’s a really good fit with the null hypothesis, then we cannot reject the null hypothesis. If we hope to receive empirical support for the research hypothesis, then we’re actually hoping for a bad fit between the observed data and what we expect according to the null hypothesis.

EXAMPLE 17.1

For example, researchers reported that the best youth soccer players in the world were more likely to have been born early in the year than later (Dubner & Levitt, 2006a). As one example, they reported that 52 elite youth players in Germany were born in January, February, or March, whereas only 4 players were born in October, November, or December. (Those born in other months were not included in this study.)

The null hypothesis predicts that when a person was born will not make any difference; the research hypothesis predicts that the month a person was born will matter when it comes to being an elite soccer player. Assuming that births in the general population are evenly distributed across months of the year, the null hypothesis posits equal numbers of elite soccer players born in the first 3 months and the last 3 months of the year. With 56 participants in the study (52 born in the first 3 months and 4 in the last 3 months), equal frequencies lead us to expect 28 players born in the first 3 months and 28 in the last 3 months just by chance. The birth months don’t appear to be evenly distributed, but is this a real pattern, or just chance?

Like previous hypothesis tests, the chi-square goodness of fit test uses the six steps of hypothesis testing.

STEP 1: Identify the populations, distribution, and assumptions.

There are always two populations involved in a chi-square test: one population that matches the frequencies of participants like those we observed and another population that matches the frequencies of participants like those we would expect according to the null hypothesis. In this case, there is a population of elite German youth soccer players with birth dates like those we observed and a population of elite German youth soccer players with birth dates like those in the general population. The comparison distribution is a chi-square distribution. There’s just one nominal variable, birth months, so we’ll conduct a chi-square test for goodness of fit.

Are Elite Soccer Players Born in the Early Months of the Year? Based on data for elite German youth soccer players, a chi-square test for goodness of fit showed a statistically significant effect: Players were more likely to be born in the first 3 months than in the last 3 months of the year (Dubner & Levitt, 2006a).
© Gero Breloer/dpa/Corbis

466

The first assumption is that the variable (birth month) is nominal. The second assumption is that each observation is independent; no single participant can be in more than one category. The third assumption is that participants were randomly selected. If not, it may be unwise to confidently generalize beyond the sample. A fourth assumption is that there is a minimum number of expected participants in every category (also called a cell)—at least 5 and preferably more. An alternative guideline (Delucchi, 1983) is for at least five times as many participants as cells. In any case, the chi-square tests seem robust to violations of this last assumption.

Summary: Population 1: Elite German youth soccer players with birth dates like those we observed. Population 2: Elite German youth soccer players with birth dates like those in the general population.

The comparison distribution is a chi-square distribution. The hypothesis test will be a chi-square test for goodness of fit because we have one nominal variable only, birth months. This study meets three of the four assumptions: (1) The one variable is nominal. (2) Every participant is in only one cell (you can’t be born in January and November). (3) This is not a randomly selected sample of all elite soccer players. The sample includes only German youth soccer players in the elite leagues. We must be cautious in generalizing beyond young German elite players. (4) There are more than five times as many participants as cells (the table has two cells, and 2 × 5 = 10). We have 56 participants, far more than the 10 necessary to meet this guideline.

STEP 2: State the null and research hypotheses.

For chi-square tests, it’s easiest to state the hypotheses in words only, rather than in both words and symbols.

Summary: Null hypothesis: Elite German youth soccer players have the same pattern of birth months as those in the general population. Research hypothesis: Elite German youth soccer players have a different pattern of birth months than those in the general population.

STEP 3: Determine the characteristics of the comparison distribution.

Our only task at this step is to determine the degrees of freedom. In most previous hypothesis tests, the degrees of freedom have been based on sample size. For the chi-square hypothesis tests, however, the degrees of freedom are based on the numbers of categories, or cells, in which participants can be counted. The degrees of freedom for a chi-square test for goodness of fit is the number of categories minus 1:

MASTERING THE FORMULA

17-1: We calculate the degrees of freedom for the chi-square test for goodness of fit by subtracting 1 from the number of categories, represented in the formula by k. The formula is: dfχ2 = k − 1

dfχ2 = k − 1

Here, k is the symbol for the number of categories. The current example has only two categories: Each soccer player in this study was born in either the first 3 months of the year or the last 3 months of the year:

dfχ2 = 2 − 1 = 1

Summary: The comparison distribution is a chi-square distribution, which has 1 degree of freedom: dfχ2 = 2 − 1 = 1

STEP 4: Determine the critical value, or cutoff.

To determine the cutoff, or critical value, for the chi-square statistic, we use the chi-square table in Appendix B. χ2 is based on squares and can never be negative, so there is just one critical value. An excerpt from Appendix B that applies to the soccer study is given in Table 17-2. We look under the p level, usually 0.05, and across from the appropriate degrees of freedom, in this case, 1. For this situation, the critical chi-square statistic is 3.841.

467

Table : TABLE 17-2. Excerpt from the χ2 Table We use the χ2 table to determine critical values for a given p level, based on the degrees of freedom.
Proportion in Critical Region
df 0.10 0.05 0.01
1 2.706 3.841 6.635
2 4.605 5.992 9.211
3 6.252 7.815 11.345  

Summary: The critical χ2, based on a p level of 0.05 and 1 degree of freedom, is 3.841, as seen in the curve in Figure 17-1.

Figure 17-1

Determining the Cutoff for a Chi-Square Statistic We look up the critical value for a chi-square statistic, based on a certain p level and degrees of freedom, in the chi-square table. Because the chi-square statistic is squared, it is never negative, so there is only one critical value.

STEP 5: Calculate the test statistic.

To calculate a chi-square statistic, we determine the observed frequencies and the expected frequencies, as seen in Table 17-3 and in the second and third columns of Table 17-4. The expected frequencies are determined from the information we have about the general population. In this case, we estimate that, in the general population, about half of all births (only, of course, among those born in the first or last 3 months of the year) occur in the first 3 months of the year, a proportion of 0.50.

(0.50)(56) = 28

468

Table : TABLE 17-4. The Chi-Square Calculations As with many other statistics, we calculate the chi-square statistic using columns to keep track of our work. We list the observed frequencies, then calculate the expected frequencies, the difference between the observed frequencies and the expected frequencies, then we square the differences and then divide each square by its appropriate expected frequency. Finally, we add up the numbers in the sixth column to find the chi-square statistic.
Column 1 2 3 4 5 6
Category Observed (O) Expected (E) OE (OE)2
First 3 months 52 28   24 576 20.571
Last 3 months 4 28 −24 576 20.571

Of the 56 elite German youth soccer players in the study, we would expect to find that 28 were born in the first 3 months of the year (versus the last 3 months of the year) if these youth soccer players are no different from the general population with respect to birth date. Similarly, we would expect a proportion of 1 − 0.50 = 0.50 of these soccer players to be born in the last 3 months of the year:

(0.50)(56) = 28

These numbers are identical only because the proportions are 0.50 and 0.50. If the proportion expected for the first 3 months of the year, based on the general population, were 0.60, then we would expect a proportion of 1 − 0.60 = 0.40 for the last 3 months of the year.

The next step in calculating the chi-square statistic is to calculate a sort of sum of squared differences. We start by determining the difference between each observed frequency and its matching expected frequency. This is usually done in columns, so we use this format even though we have only two categories. The first three columns of Table 17-4 show us the categories, observed frequencies, and expected frequencies, respectively. The fourth column, using O for observed and E for expected, displays the differences. As in the other situations, if we sum the differences, we get 0; they cancel out because some are positive and some are negative. We solve this problem as we have others—by squaring the differences, as shown in the fifth column. Next, however, we have a step that we haven’t seen before with squared differences. We divide each squared difference by the expected value for its cell, as seen in the sixth column. The numbers in the sixth column are the ones we sum.

As an example, here are the calculations for the category “first 3 months”:

MASTERING THE FORMULA

17-2: The formula for the chi-square statistic is: . For each cell, we subtract the expected count, E, from the observed count, O. Then we square each difference and divide the square by the expected count. Finally, we sum the calculations for each of the cells.

Once we complete the table, the last step is easy. We just add up the numbers in the sixth column. In this case, the chi-square statistic is 20.571 + 20.571 = 41.14. We can finish the formula by adding a summation sign to the formula in the sixth column. Note that we don’t have to divide this sum by anything, as we’ve done with other statistics. We already did the dividing before we summed. This sum is the chi-square statistic. Here is the formula:

469

Summary:

STEP 6: Make a decision.

This last step is identical to that of previous hypothesis tests. We reject the null hypothesis if the test statistic is beyond the critical value, and we fail to reject the null hypothesis if the test statistic is not beyond the critical value. In this case, the test statistic, 41.14, is far beyond the cutoff, 3.841, as seen in Figure 17-2. We reject the null hypothesis. Because there are only two categories, it’s clear where the difference lies. It appears that elite German youth soccer players are more likely to have been born in the first 3 months of the year, and less likely to have been born in the last 3 months of the year, than members of the general population. (If we had failed to reject the null hypothesis, we could only have concluded that these data did not provide sufficient evidence to show that elite German youth soccer players have a different likelihood of being born in the first, versus last, 3 months of the year than those in the general population.)

Figure 17-2

Making a Decision As with other hypothesis tests, we make a decision with a chi-square test by comparing the test statistic to the cutoff, or critical value. We see here that 41.14 would be far to the right of 3.841.

Summary: Reject the null hypothesis; it appears that elite German youth soccer players are more likely to have been born in the first 3 months of the year, and less likely to have been born in the last 3 months of the year, than people in the general population.

We report these statistics in a journal article in almost the same format that we’ve seen previously. We report the degrees of freedom, the value of the test statistic, and whether the p value associated with the test statistic is less than or greater than the cutoff based on the p level of 0.05. (As usual, we would report the actual p level if we conducted this hypothesis test using software.) In addition, we report the sample size in parentheses with the degrees of freedom. In the current example, the statistics read:

χ2 (1, N = 56) = 41.14, p < 0.05

The researchers who conducted this study imagined four possible explanations: “a) certain astrological signs confer superior soccer skills; b) winter-born babies tend to have higher oxygen capacity, which increases soccer stamina; c) soccer-mad parents are more likely to conceive children in springtime, at the annual peak of soccer mania; d) none of the above” (Dubner & Levitt, 2006a). What’s your guess?

470

Dubner and Levitt (2006a) picked (d) and suggested another alternative. Participation in youth soccer leagues has a strict cutoff date: December 31. Compared to those born in December, children born the previous January are likely to be more physically and emotionally mature, perceived as more talented, chosen for the best leagues, and given better coaching—a self-fulfilling prophecy. All this from a simple chi-square test for goodness of fit!

Chi-Square Test for Independence

Clown Therapy Israeli researchers tested whether entertainment by a clown led to higher pregnancy rates after in vitro fertilization treatment. Their study had two nominal variables— entertainment (clown, no clown) and pregnancy (pregnant, not pregnant)—and could have been analyzed with a chi-square test for independence.
Image Source/Getty Images

The chi-square test for goodness of fit analyzes just one nominal variable. The chi-square test for independence analyzes two nominal variables.

Like the correlation coefficient, the chi-square test for independence does not require that we identify independent and dependent variables. However, specifying an independent variable and a dependent variable can help us articulate hypotheses. The chi-square test for independence is so named because it is used to determine whether the two variables—no matter which one is considered to be the independent variable—are independent of each other. Let’s take a closer look at whether pregnancy rates are independent of (that is, depend on) whether one is entertained by a clown after IVF treatment.

EXAMPLE 17.2

In the clown study, as reported in the mass media (Ryan, 2006), 186 women were randomly assigned to receive IVF treatment only or to receive IVF treatment followed by 15 minutes of clown entertainment. Eighteen of the 93 who received only the IVF treatment became pregnant, whereas 33 of the 93 who received both IVF treatment and clown entertainment became pregnant. The cells for these observed frequencies can be seen in Table 17-5. The table of cells for a chi-square test for independence is called a contingency table because it helps us see if the outcome of one variable (e.g., becoming pregnant versus not becoming pregnant) is contingent on the other variable (clown versus no clown). Let’s implement the six steps of hypothesis testing for a chi-square test for independence.

Table : TABLE 17-5. Observed Pregnancy Rates This table depicts the cells and their frequencies for the study on whether entertainment by a clown is associated with pregnancy rates among women undergoing in vitro fertilization.
Observed
Pregnant Not pregnant
Clown 33 60
No clown 18 75

471

STEP 1: Identify the populations, distribution, and assumptions.

Population 1: Women receiving IVF treatment like the women we observed. Population 2: Women receiving IVF treatment for whom the presence of a clown is not associated with eventual pregnancy.

The comparison distribution is a chi-square distribution. The hypothesis test will be a chi-square test for independence because we have two nominal variables. This study meets three of the four assumptions: (1) The two variables are nominal. (2) Every participant is in only one cell. (3) The participants were not, however, randomly selected from the population of all women undergoing IVF treatment. We must be cautious in generalizing beyond the sample of Israeli women at this particular hospital. (4) There are more than five times as many participants as cells (186 participants and 4 cells; 4 × 5 = 20). We have far more participants, 186, than the 20 necessary to meet this guideline.

STEP 2: State the null and research hypotheses.

Null hypothesis: Pregnancy rates are independent of whether one is entertained by a clown after IVF treatment. Research hypothesis: Pregnancy rates depend on whether one is entertained by a clown after IVF treatment.

STEP 3: Determine the characteristics of the comparison distribution.

For a chi-square test for independence, we calculate degrees of freedom for each variable and then multiply the two to get the overall degrees of freedom. The degrees of freedom for the variable in the rows of the contingency table is:

dfrow = krow − 1

The degrees of freedom for the variable in the columns of the contingency table is:

dfcolumn = kcolumn − 1

The overall degrees of freedom is:

MASTERING THE FORMULA

17-3: To calculate the degrees of freedom for the chi-square test for independence, we first have to calculate the degrees of freedom for each variable. For the variable in the rows, we subtract 1 from the number of categories in the rows: dfrowkrow = 1. For the variable in the columns, we subtract 1 from the number of categories in the columns: dfcolumn = kcolumn − 1. We multiply these two numbers to get the overall degrees of freedom: dfχ2 = (dfrow)(dfcolumn). To combine all the calculations, we can use the following formula instead: dfχ2 = (krow − 1)(kcolomn − 1).

dfχ2 = (dfrow)(dfcolumn)

To expand this last formula, we write:

dfχ2 = (krow − 1)(kcolomn − 1)

The comparison distribution is a chi-square distribution, which has 1 degree of freedom:

dfχ2 = (krow − 1)(kcolomn − 1) = (2 − 1)(2 − 1) = 1

STEP 4: Determine the critical values, or cutoffs.

The critical value, or cutoff, for the chi-square statistic, based on a p level of 0.05 and 1 degree of freedom, is 3.841 (Figure 17-3).

Figure 17-3

The Cutoff for a Chi-Square Test for Independence The shaded region is beyond the critical value for a chi-square test for independence with a p level of 0.05 and 1 degree of freedom. If the test statistic falls within this shaded area, we will reject the null hypothesis.

STEP 5: Calculate the test statistic.

The next step, determining the appropriate expected frequencies, is the most important in the calculation of the chi-square test for independence. Errors are often made in this step, and if the wrong expected frequencies are used, the chi-square statistic derived from them will also be wrong. Many students want to divide the total number of participants (here, 186) by the number of cells (here, 4) and place equivalent frequencies in all cells for the expected data. Here, that would mean that the expected frequencies would be 46.5.

472

But this would not make sense. Of the 186 women, only 51 became pregnant; 51/186 = 0.274, or 27.4%, of these women became pregnant. If pregnancy rates do not depend on clown entertainment, then we would expect the same percentage of successful pregnancies, 27.4%, regardless of exposure to clowns. If we have expected frequencies of 46.5 in all four cells, then we have a 50%, not a 27.4%, pregnancy rate. We must always consider the specifics of the situation.

In the current study, we already calculated that 27.4% of all women in the study became pregnant. If pregnancy rates are independent of whether a woman is entertained by a clown, then we would expect 27.4% of the women who were entertained by a clown to become pregnant and 27.4% of women who were not entertained by a clown to become pregnant. Based on this percentage, 100 − 27.4 = 72.6% of women in the study did not become pregnant. We would therefore expect 72.6% of women who were entertained by a clown to fail to become pregnant and 72.6% of women who were not entertained by a clown to fail to become pregnant. Again, we expect the same pregnancy and nonpregnancy rates in both groups—those who were and were not entertained by clowns.

Table 17-6 shows the observed data, and it also shows totals for each row, each column, and the whole table.

Table : TABLE 17-6. Observed Frequencies This table depicts the cells and their frequencies for the study on whether entertainment by a clown is associated with pregnancy rates among women undergoing in vitro fertilization. It also includes row totals (93, 93), column totals (51, 135), and the grand total for the whole table is 186.
Observed
Pregnant Not pregnant
Clown 33 60 93
No clown 18 75 93
51 135   186  

From Table 17-6, we see that 93 women were entertained by a clown after IVF treatment. As we calculated above, we would expect 27.4% of them to become pregnant:

(0.274)(93) = 25.482

Of the 93 women who were not entertained by a clown, we would expect 27.4% of them to become pregnant if clown entertainment is independent of pregnancy rates:

(0.274)(93) = 25.482

We now repeat the same procedure for not becoming pregnant. We would expect 72.6% of women in both groups to fail to become pregnant. For the women who were entertained by a clown, we would expect 72.6% of them to fail to become pregnant:

(0.726)(93) = 67.518

473

For the women who were not entertained by a clown, we would expect 72.6% of them to fail to become pregnant:

(0.726)(93) = 67.518

(Note that the two expected frequencies for the first row are the same as the two expected frequencies for the second row, but only because the same number of people were in each clown condition, 93. If these two numbers were different, we would not see the same expected frequencies in the two rows.)

The method of calculating the expected frequencies that we described above is ideal because it is directly based on our own thinking about the frequencies in the rows and in the columns. Sometimes, however, our thinking can get muddled, particularly when the two (or more) row totals do not match and the two (or more) column totals do not match. For these situations, a simple set of rules leads to accurate expected frequencies. For each cell, we divide its column total (Totalcolumn) by the grand total (N) and multiply that by the row total (Totalrow):

MASTERING THE FORMULA

17-4: When conducting a chi-square test for independence, we can calculate the expected frequencies in each cell by taking the total for the column that the cell is in, dividing it by the total in the study, and then multiplying by the total for the row that the cell is in: .

As an example, the observed frequency of those who became pregnant and were entertained by a clown is 33. The row total for this cell is 93. The column total is 51. The grand total, N, is 186. The expected frequency, therefore, is:

Notice that this result is identical to what we calculated without a formula. The middle step above shows that, even with the formula, we actually did calculate the pregnancy rate overall, by dividing the column total (51) by the grand total (186). We then calculated how many in that row of 93 participants we would expect to become pregnant using this overall rate:

(0.274)(93) = 25.482

The formula follows our logic, but it also keeps us on track when there are multiple calculations.

As a final check on the calculations, shown in Table 17-7, we can add up the frequencies to be sure that they still match the row, column, and grand totals. For example, if we add the two numbers in the first column, 25.482 and 25.482, we get 50.964 (different from 51 only because of rounding decisions). If we had made the mistake of dividing the 186 participants into cells by dividing by 4, we would have had 46.5 in each cell; then the total for the first column would have been 46.5 + 46.5 = 93, which is not a match with 51. This final check ensures that we have the appropriate expected frequencies in the cells.

Table : TABLE 17-7. Expected Frequencies with Totals This table includes the expected frequencies for each of the four cells. The expected frequencies should still add up to the row totals (93, 93), column totals (51, 135), and the grand total for the whole table (186).
Expected
Pregnant Not pregnant
Clown 25.482 67.518 93
No clown 25.482 67.518 93
51         135           186  

474

The remainder of the fifth step is identical to that for a chi-square test for goodness of fit, as seen in Table 17-8. As before, we calculate the difference between each observed frequency and its matching expected frequency, square these differences, and divide each squared difference by the appropriate expected frequency. We add up the numbers in the final column of the table to calculate the chi-square statistic:

Table : TABLE 17-8. The Chi-Square Calculations For the chi-square test for independence calculations, we use the same format as we did for the chi-square test for goodness of fit. We calculate the difference between each observed frequency and expected frequency, square the difference, then divide each square by its appropriate expected frequency. Finally, we add up the numbers in the last column, which gives us the chi-square statistic.
Category Observed (O) Expected (E) OE (OE)2
Clown; pregnant 33 25.482   7.518 56.520 2.218
Clown; not pregnant 60 67.518 −7.518 56.520 0.837
No clown; pregnant 18 25.482 −7.482 55.980 2.197
No clown; not pregnant 75 67.518 7.482 55.980 0.829

STEP 6: Make a decision.

Reject the null hypothesis; it appears that pregnancy rates depend on whether a woman receives clown entertainment following IVF treatment (Figure 17-4). The statistics, as reported in a journal article, would follow the format we learned for a chi-square test for goodness of fit as well as for other hypothesis tests in earlier chapters. We report the degrees of freedom and sample size, the value of the test statistic, and whether the p value associated with the test statistic is less than or greater than the critical value based on the p level of 0.05. (We would report the actual p level if we conducted this hypothesis test using software.) In the current example, the statistics would read:

χ2 (1, N = 186) = 6.08, p < 0.05

Figure 17-4

The Decision Because the chi-square statistic, 6.081, is beyond the critical value, 3.841, we can reject the null hypothesis. It is unlikely that the pregnancy rates for those who received clown therapy versus those who did not were this different from each other just by chance.

475

CHECK YOUR LEARNING

Reviewing the Concepts

  • The chi-square tests are used when all variables are nominal.
  • The chi-square test for goodness of fit is used with one nominal variable.
  • The chi-square test for independence is used with two nominal variables; usually one can be thought of as the independent variable and one as the dependent variable.
  • Both chi-square hypothesis tests use the same six steps of hypothesis testing with which we are familiar.

Clarifying the Concepts

  • 17-5 When do we use chi-square tests?
  • 17-6 What are observed frequencies and expected frequencies?

Calculating the Statistics

  • 17-7 Imagine a town that boasts clear blue skies 80% of the time. You get to work in that town one summer for 78 days and record the following data. (Note: For each day, you picked just one label.)
    Clear blue skies: 59 days
    Cloudy/hazy/gray skies: 19 days
    1. Calculate degrees of freedom for this chi-square test for goodness of fit.
    2. Determine the observed and expected frequencies.
    3. Calculate the differences and squared differences between frequencies, and calculate the chi-square statistic. Use the six-column format provided here.
    Category Observed (O) Expected (E) OE (OE)2
    Clear blue skies
    Unclear skies

Applying the Concepts

  • 17-8 The Chicago Police Department conducted a study comparing two types of lineups for suspect identification: simultaneous lineups and sequential lineups (Mecklenburg, Malpass, & Ebbesen, 2006). In simultaneous lineups, witnesses saw the suspects all at once, either live or in photographs, and then made their selection. In sequential lineups, witnesses saw the people in the lineup one at a time, either live or in photographs, and said yes or no to suspects one at a time. After numerous high-profile cases in which DNA evidence exonerated people who had been convicted, including many on death row, many police departments shifted to sequential lineups in the hope of reducing incorrect identifications. Several previous studies had indicated the superiority of sequential lineups with respect to accuracy. Over one year, three jurisdictions in Illinois compared the two types of lineups. Of 319 simultaneous lineups, 191 led to identification of the suspect, 8 led to identification of another person in the lineup, and 120 led to no identification. Of 229 sequential lineups, 102 led to identification of the suspect, 20 led to identification of another person in the lineup, and 107 led to no identification.
    1. Who or what are the participants in this study? Identify the independent variable and its levels as well as the dependent variable and its levels.
    2. Conduct all six steps of hypothesis testing.
    3. Report the statistics as you would in a journal article.
    4. Why is this study an example of the importance of using two-tailed rather than one-tailed hypothesis tests?

Solutions to these Check Your Learning questions can be found in Appendix D.

476