9.2 The Single-Sample t Test

  • A single-sample t test is a hypothesis test in which we compare a sample from which we collect data to a population for which we know the mean but not the standard deviation.

A single-sample t test is a hypothesis test in which we compare a sample from which we collect data to a population for which we know the mean but not the standard deviation. The logic of the single-sample t test is a model for other t tests that allow us to compare two samples—and all of the other more sophisticated statistical tests that will follow. You will soon be able to study just about anything you want.

The t Table and Degrees of Freedom

  • Degrees of freedom is the number of scores that are free to vary when we estimate a population parameter from a sample.

When we use the t distributions, we use the t table. There are different t distributions for every sample size and the t table takes sample size into account. However, we do not look up the actual sample size on the table. Rather, we look up degrees of freedom, the number of scores that are free to vary when we estimate a population parameter from a sample.

Language Alert! The phrase “free to vary” refers to the number of scores that can take on different values when a given parameter is known.

EXAMPLE 9.4

MASTERING THE CONCEPT

9-2: Degrees of freedom refers to the number of scores that can take on different values when a given parameter is known. For example, if we know that the mean of three scores is 10, only two scores are free to vary. Once we know the values of two scores, we know the value of the third. If we know that two of the scores are 9 and 10, then we know that the third must be 11.

For example, the manager of a baseball team needs to assign nine players to particular spots in the batting order but only has to make eight decisions (N − 1). Why? Because only one option remains after making the first eight decisions. So before the manager makes any decisions, there are N − 1, or 9 − 1 = 8, degrees of freedom. After the second decision, there are N − 1, or 8 − 1 = 7, degrees of freedom, and so on.

As in the baseball example, there is always one score that cannot vary once all of the others have been determined. For example, if we know that the mean of four scores is 6 and we know that three of the scores are 2, 4, and 8, then the last score must be 10. So the degrees of freedom is the number of scores in the sample minus 1. Degrees of freedom is written in symbolic notation as df, which is always italicized. The formula for degrees of freedom for a single-sample t test, therefore, is:

df = N − 1

MASTERING THE FORMULA

9-4: The formula for degrees of freedom for a single-sample t test is: df = N − 1. To calculate degrees of freedom, we subtract 1 from the sample size.

Table 9-1 is an excerpt from a t table; the full table is in Appendix B. Notice the relation between degrees of freedom and the critical value needed to declare statistical significance. As degrees of freedom go up, the critical values go down. In the column corresponding to a one-tailed test at a p level of 0.05 with only 1 degree of freedom (two observations), the critical t value is 6.314. With only 1 degree of freedom, the two means have to be extremely far apart and/or the standard deviation has to be very small to declare a statistically significant difference. But with 2 degrees of freedom (three observations), the critical t value drops to 2.920. It is easier to reach the critical t value because we’re more confident in making a reliable observation with three observations than with just two, just as we are more confident that a rumor is true when we hear the same story from three independent observers than from just two people.

image

220

The pattern continues when we have four observations (with df of 3). The critical t value needed to declare statistical significance decreases from 2.920 to 2.353. The level of confidence in the observations increases and the critical value decreases.

The t distributions become closer to the z distribution as sample size increases. After all, if we kept enlarging the sample, we would eventually study the entire population and wouldn’t need a pesky t test in the first place. But in the real world of research, the corrected standard deviation of a large enough sample is so similar to the actual standard deviation of the population that the t distribution is the same as the z distribution.

MASTERING THE CONCEPT

9-3: As sample size increases, the t distributions more and more closely approximate the z distribution. You can think of the z statistic as a single-blade Swiss Army knife and the t statistic as a multi-blade Swiss Army knife that includes the single blade that is the z statistic.

Check it out for yourself by comparing the z and t tables in Appendix B. For example, the z statistic for the 95th percentile—a percentage between the mean and the z statistic of 45%—is between 1.64 and 1.65. At a sample size of infinity, the t statistic for the 95th percentile is 1.645. Infinity () indicates a very large sample size; a sample size of infinity itself is, of course, impossible.

image
Nonparticipation in Therapy Clients missing appointments can be a problem for both clients and therapists. A t test can compare the consequences between those who do and do not sign a contract to attend a set number of sessions.
Zigy Kaluzny/Getty Images

Let’s remind ourselves why the t statistic merges with the z statistic as sample size increases. More participants in a study—if they are a representative sample—correspond to increased confidence that we are making an accurate observation. So don’t think of the t distributions as completely separate from the z distribution. Rather, think of the z statistic as a single-blade Swiss Army knife and the t statistic as a multi-blade Swiss Army knife that still includes the single blade that is the z statistic.

Let’s determine the cutoffs, or critical t values, for a research study using the full t table in Appendix B.

EXAMPLE 9.5

The study: A researcher knows the mean number of calories lab rats will consume in half an hour if unlimited food is available. She wonders whether a new food will lead rats to consume a different number of calories—either more or fewer. She studies 38 rats and uses a p level of 0.05.

The cutoff(s): This is a two-tailed test because the research hypothesis allows for change in either direction. There are 38 rats, so the degrees of freedom is:

df = N − 1 = 38 − 1 = 37

221

We want to look in the t table under two-tailed tests, in the column for 0.05 and in the row for a df of 37; however, there is no df of 37. In this case, we err on the side of being more conservative and choose the more extreme (i.e., larger) of the two possible critical t values, which is always the smaller df. Here, we look next to 35, where we see a value of 2.030. Because this is a two-tailed test, we will have critical values of −2.030 and 2.030.

The Six Steps of the Single-Sample t Test

Now we have all the tools necessary to conduct a single-sample t test. So let’s consider a hypothetical study and conduct all six steps of hypothesis testing.

EXAMPLE 9.6

Chapter 4 presented data that included the mean number of sessions attended by clients at a university counseling center. We noted that one study reported a mean of 4.6 sessions (Hatchett, 2003). Let’s imagine that the counseling center hoped to increase participation rates by having students sign a contract to attend at least 10 sessions. Five students sign the contract and attend 6, 6, 12, 7, and 8 sessions, respectively. The researchers are interested only in their university, so treat the mean of 4.6 sessions as a population mean.

STEP 1: Identify the populations, distribution, and assumptions.

Population 1: All clients at this counseling center who sign a contract to attend at least 10 sessions. Population 2: All clients at this counseling center who do not sign a contract to attend at least 10 sessions.

The comparison distribution will be a distribution of means. The hypothesis test will be a single-sample t test because we have only one sample and we know the population mean but not the population standard deviation.

This study meets one of the three assumptions and may meet the other two: (1) The dependent variable is scale. (2) We do not know whether the data were randomly selected, however, so we must be cautious with respect to generalizing to other clients at this university who might sign the contract. (3) We do not know whether the population is normally distributed, and there are not at least 30 participants. However, the data from the sample do not suggest a skewed distribution.

STEP 2: State the null and research hypotheses.

Null hypothesis: Clients at this university who sign a contract to attend at least 10 sessions attend the same number of sessions, on average, as clients who do not sign such a contract—H0: µ1 = µ2.

Research hypothesis: Clients at this university who sign a contract to attend at least 10 sessions attend a different number of sessions, on average, than do clients who do not sign such a contract—H1: µ1µ2.

222

STEP 3: Determine the characteristics of the comparison distribution.






µM = 4.6; sM = 1.114

Calculations:

µM = µ = 4.6

image
X XM (XM)2
6 −1.8 3.24
6 −1.8 3.24
12 4.2 17.64
7 −0.8 0.64
8 0.2 0.04

The numerator of the standard deviation formula is the sum of squares:

image

STEP 4: Determine the critical values, or cutoffs.





df = N − 1 = 5 − 1 = 4

For a two-tailed test with a p level of 0.05 and df of 4, the critical values are −2.776 and 2.776 (as seen in the curve in Figure 9-2).

image
Figure 9.3: FIGURE 9-2
Determining Cutoffs for a t Distribution
Figure 9.3: As with the z distribution, we typically determine critical values in terms of t statistics rather than means of raw scores so that we can easily determine whether the test statistic is beyond one of the cutoffs. Here, the cutoffs are −2.776 and 2.776, and they mark off the most extreme 5%, with 2.5% in each tail.

223

STEP 5: Calculate the test statistic.

image

STEP 6: Make a decision.

Reject the null hypothesis. It appears that counseling center clients who sign a contract to attend at least 10 sessions do attend more sessions, on average, than do clients who do not sign such a contract (Figure 9-3).

image
Figure 9.4: FIGURE 9-3
Making a Decision
Figure 9.4: To decide whether to reject the null hypothesis, we compare the test statistic to the critical t values. In this case, the test statistic, 2.873, is beyond the cutoff of 2.776, so we can reject the null hypothesis.

After completing the hypothesis test, we want to present the primary statistical information in a report. There is a standard American Psychological Association (APA) format for the presentation of statistics across the behavioral sciences so that the results are easily understood by the reader.

  1. Write the symbol for the test statistic (e.g., t).

  2. Write the degrees of freedom, in parentheses.

  3. Write an equal sign and then the value of the test statistic, typically to two decimal places.

  4. Write a comma and then indicate the p value by writing “p =” and then the actual value. (Unless we use software to conduct the hypothesis test, we will not know the actual p value associated with the test statistic. In this case, we simply state whether the p value is beyond the critical value by saying p < 0.05 when we reject the null hypothesis or p > 0.05 when we fail to reject the null hypothesis.)

In the counseling center example, the statistics would read:

t(4) = 2.87, p < 0.05

The statistics typically follow a statement about the finding. For example, “It appears that counseling center clients who sign a contract to attend at least 10 sessions do attend more sessions, on average, than do clients who do not sign such a contract, t(4) = 2.87, p < 0.05.” The report would also include the sample mean and standard deviation (not standard error) to two decimal points. Here, the descriptive statistics would read: (M = 7.80, SD = 2.49). By convention, we use SD instead of s to symbolize the standard deviation.

224

Calculating a Confidence Interval for a Single-Sample t Test

As it does with a z test, the APA recommends that researchers report confidence intervals and effect sizes, in addition to the results of hypothesis tests, whenever possible.

EXAMPLE 9.7

MASTERING THE CONCEPT

9-4: Whenever researchers conduct a hypothesis test, the APA encourages that, if possible, they also calculate a confidence interval and an effect size.

We can calculate a confidence interval with the single-sample t test data. The population mean was 4.6. We used the sample to estimate the population standard deviation to be 2.490 and the population standard error to be 1.114. The five students in the sample attended a mean of 7.8 sessions.

When we conducted hypothesis testing, we centered the curve around the mean according to the null hypothesis—the population mean of 4.6. Now we can use the same information to calculate the 95% confidence interval around the sample mean of 7.8.

STEP 1: Draw a picture of a t distribution that includes the confidence interval.

We draw a normal curve (Figure 9-4) that has the sample mean, 7.8, at its center (instead of the population mean, 4.6).

image
Figure 9.5: FIGURE 9-4
A 95% Confidence Interval for a Single-Sample t Test, Part I
Figure 9.5: To begin calculating a confidence interval for a single-sample t test, we place the sample mean, 7.8, at the center of a curve and indicate the percentages within and beyond the confidence interval.

STEP 2: Indicate the bounds of the confidence interval on the drawing.

We draw a vertical line from the mean to the top of the curve. For a 95% confidence interval, we also draw two much smaller vertical lines that indicate the middle 95% of the t distribution (2.5% in each tail, for a total of 5%). We then write the appropriate percentages under the segments of the curve.

STEP 3: Look up the t statistics that fall at each line marking the middle 95%.

For a two-tailed test with a p level of 0.05 and a df of 4, the critical values are −2.776 and 2.776. We can now add these t statistics to the curve, as seen in Figure 9-5.

image
Figure 9.6: FIGURE 9-5
A 95% Confidence Interval for a Single-Sample t Test, Part II
Figure 9.6: The next step in calculating a confidence interval for a single-sample t test is to identify the t statistics that indicate each end of the interval. Because the curve is symmetric, the t statistics have the same magnitude—one is negative, −2.776, and one is positive, 2.776.

225

STEP 4: Convert the t statistics back into raw means.

As we did with the z test, we can use formulas for this conversion, but first we identify the appropriate mean and standard deviation. There are two important points to remember. First, we center the interval around the sample mean, so we use the sample mean of 7.8 in the calculations. Second, because we have a sample mean (rather than an individual score), we use a distribution of means. So we use the standard error of 1.114 as the measure of spread.

Using this mean and standard error, we calculate the raw mean at each end of the confidence interval, and add them to the curve, as in Figure 9-6. The formulas are exactly the same as for the z test except that z is replaced by t, and σM is replaced by sM.

image
Figure 9.7: FIGURE 9-6
A 95% Confidence Interval for a Single-Sample t Test, Part III
Figure 9.7: The final step in calculating a confidence interval for a single-sample t test is to convert the t statistics that indicate each end of the interval into raw means, 4.71 and 10.89.

Mlower = −t(sM) + Msample = −2.776(1.114) +7.8 = 4.71

Mupper = t(sM) + Msample = 2.776(1.114) +7.8 = 10.89

The 95% confidence interval, reported in brackets as is typical, is [4.71, 10.89].

STEP 5: Verify that the confidence interval makes sense.

The sample mean should fall exactly in the middle of the two ends of the interval.

MASTERING THE FORMULA

9-5: The formula for the lower bound of a confidence interval for a single-sample t test is Mlower = −t(sM) + Msample. The formula for the upper bound of a confidence interval for a single-sample t test is Mupper = t(sM) + Msample. The only differences from those for a z test are that in each formula z is replaced by t, and σM is replaced by sM.

4.71 − 7.8 = −3.09; and 10.89 − 7.8 = 3.09

We have a match. The confidence interval ranges from 3.09 below the sample mean to 3.09 above the sample mean. If we were to sample five students from the same population over and over, the 95% confidence interval would include the population mean 95% of the time. Note that the population mean, 4.6, does not fall within this interval. This means it is not plausible that this sample of students who signed contracts came from the population according to the null hypothesis—students seeking treatment at the counseling center who did not sign a contract. We conclude that the sample comes from a different population, one in which students attend more mean sessions than does the general population. As with the z test, the conclusions from the hypothesis test and the confidence interval are the same, but the confidence interval gives us more information—an interval estimate, not just a point estimate.

Calculating Effect Size for a Single-Sample t Test

As with a z test, we can calculate the effect size (Cohen’s d) for a single-sample t test.

EXAMPLE 9.8

MASTERING THE FORMULA

9-6: The formula for Cohen’s d for a t statistic is:

image

It is the same formula as for the t statistic, except that we divide by the population standard deviation (s) rather than by the population standard error (sM).

Let’s calculate the effect size for the counseling center study. Similar to what we did with the z test, we simply use the formula for the t statistic, substituting s for sM (and µ for µM, even though these means are always the same). This means we use 2.490 instead of 1.114 in the denominator. Cohen’s d is based on the spread of the distribution of individual scores, rather than the distribution of means.

226

image

The effect size, d = 1.29, tells us that the sample mean and the population mean are 1.29 standard deviations apart. According to the conventions we learned in Chapter 8 (that 0.2 is a small effect, 0.5 is a medium effect, and 0.8 is a large effect), this is a large effect. We can add the effect size when we report the statistics as follows: t(4) = 2.87, p < 0.05, d = 1.29.

CHECK YOUR LEARNING

Reviewing the Concepts
  • A single-sample t test is used to compare data from one sample to a population for which we know the mean but not the standard deviation.

  • We consider degrees of freedom, or the number of scores that are free to vary, instead of N when we assess estimated t statistics against t distributions.

  • As sample size increases, confidence in the estimates improves, degrees of freedom increase, and the critical values for t drop, making it easier to reach statistical significance. In fact, as sample size grows, the t distributions approach the z distribution.

  • To conduct a single-sample t test, we follow the same six steps of hypothesis testing as we do for the z test, except that we estimate the standard deviation from the sample before we calculate standard error.

  • We can calculate a confidence interval and an effect size, Cohen’s d, for a single-sample t test.

Clarifying the Concepts 9-5 Explain the term degrees of freedom.
9-6 Why is a single-sample t test more useful than a z test?
Calculating the Statistics 9-7 Compute degrees of freedom for each of the following:
  1. An experimenter times how long it takes 35 rats to run through a maze with eight pathways.

  2. Test scores for 14 students are collected and averaged over four semesters.

9-8 Identify the critical t value(s) for each of the following tests:
  1. A two-tailed test with alpha of 0.05 and 11 degrees of freedom

  2. A one-tailed test with alpha of 0.01 and N of 17

Applying the Concepts 9-9 Let’s assume that according to university summary statistics, the average student misses 3.7 classes during a semester. Imagine that these are the data you have been working with (6, 3, 7, 6, 4, 5) for the number of classes missed by a group of students. Conduct all six steps of hypothesis testing, using a two-tailed test with a p level of 0.05. (Note: You completed the work for step 3 in Check Your Learning 9-2 and 9-3.)

Solutions to these Check Your Learning questions can be found in Appendix D.