Psychology 340 Syllabus
Statistics for the Social Sciences

Illinois State University
J. Cooper Cutting
Fall 2002



Hypothesis testing II: t-tests

  • Which test decision tree
  • One sample t-test
  • Matched samples t-test
  • Independent samples t-test
  • T-tests in SPSS


    Let's quickly recall the decision tree that we saw earlier in the semester.

    Find the string of decisions that lead to a 1-sample t-test.

    T-tests are basically the same as the one-sample z test from last time. With the earlier test statistic were concerned with situations in which we assumed that we knew the population standard devaition (s). As a result we could compute the actual standard error () associated with samples of size n.

    T-tests are used in situations in which we DON'T know the population standard devaition (s). This is a much more common situation (that is, we usually don't know the population values because we usually don't have access to the entire population). So instead, we will have to use an estimated standard error () using information that we get from our sample.

    What do we already know about that we can use as an estimate of s?

    If we know s: If we don't know s:
    standard error of = = estimated standard error of = s = =
    test statistic: z-score

      z =
    test statistic t-score

      t =
    estimating m

      m = + (Z-crit)()
    estimating m

      m = + (t-crit)()


    Even though the formulas in the two situations are very similar, there is an important conceptual difference between the two situations. Because we are using the sample standard deviation (s) to estimate the population standard deviation (s), then we need to take into account the fact that it is an estimate. If you think back to our earlier chapter in which we discussed standard deviations of samples, you'll remember that we must take the degrees of freedom into account.

    Okay, let's look at a few examples



    Using SPSS to compute a one-sample t-test


    Hypothesis tests analyzed with related samples t-tests

    In the prior lab we examined how to use a t-test to compare a treatment sample against a population (for which s isn't known). In this lab we'll consider the case where the null population m isn't known and must also be represented by a sample (like the treatment m was in the one-sample cases. Today we'll consider situations where the two samples means come from related samples. The are two ways the samples can be related. In one case, there are two separate but related samples. In the other case, there is a single sample of individuals, each of which gets measured on the dependent variable twice.

    Let's quickly recall the decision tree that we saw earlier in the semester.

    Find the string of decisions that lead to either a two-independent samples t-test, or a related samples t-test.

    Consider the following examples:

    In the first examples, the situation has been decided for you, there is a pre-existing relationship between the two samples.

    In the second and third examples, you, as the experimenter make a decision to make the two samples related. Why would you ever want to do that? To control for individual differences that might add more noise (error) to your data. In Example 2, each individual acts as their own control. In Example 3, the control group is made up of people as similar to the people in the experimental group as you could get them. Both of these designs are used to try to reduce error resulting from individual differences.

    Okay, so now we know that for repeated-measures and matched-subject designs we need a new t-test. So, what is the t statistic for related samples?

    Again, the logic of the hypothesis test is pretty much the same as it was for the one-sample cases we've already considered. Once again we'll go through the same steps. What changes are the nature of the hypothesis, and how the t is computed.

    All of the tests that we've looked at are examining differences. In the previous lab we were interested in comparing a known population with a treatment sample. Now we are beginning to consider cases when the null population m is unknown and must also be represented by a sample. The t-test for this chapter is also interested in the differences, but because the two samples are related, the differences are based on differences between each individual.

    Consider the following example:

    An instructor asks his statistics class, on the first day of classes, to rate how much they like statistics, on a scale of 1 to 10 (1 hate it, 10 love it). Then, at the end of the semester, the instructor asks the same students, the same question. The instructor wants to know if taking the stats course had an impact on the students feelings about statistics.

    The results of the two ratings are presented below. D stands for the difference between the pre- and post-ratings for each individual.

    Student Pre-test (first day) Post-test (end of semester) D D2
    1 1 4 3 9
    2 3 5 2 4
    3 4 6 2 4
    4 7 8 1 1
    5 2 3 1 1
    6 2 2 0 0
    7 4 6 2 4
    8 3 4 1 1
    9 6 6 0 0
    10 8 6 -2 4
    S 40 50 10 28

    mean difference = = 10/10 = 1.0


    Using SPSS to compute a related samples (paired samples) t-test


    Two independent samples t-tests

    The section above used a different computational formula to calculate the observed t for two more situations:

    The basic logic of the independent samples t-test should seem similar to the other tests that we've covered. We still use the t-distribution to find our critical values. However things get a little more complicated, because of the situation that we are interested in. Now we are going to look at a situation where we are interested in the potential difference between two different populations. And again, we'll deal with situations in which we don't know the s for either of these populations, so we'll have to use estimates.

    An experiment that uses a separate independent samples for each treatment condition (or each population) is called an independent-measures research design. Often you'll also see it referred to as a between-subjects or between-groups desgin.

    So we'll use the same logic and steps for hypothesis testing that we used in the previous labs, and fill in the details of the differences as we go.

    Let's start with step 1.

    Step 2.

    Step 3.

    Step 4.

    Step 5.

    step 6.

    step 7:

    Computing confidence intervals with two independent samples is very similar to what we've done in past labs, except that we use the estimate of the standard error that is derived from our pooled variance and both of the sample means are used. So the formula is:



    What are the assumptions of our independent measures t test?

    1) The observations are independent (both between and within groups)

    2) The two populations are normally distributed (also discussed in previous labs)

    ** new **

    3) The two populations have equal variances. This is referred to as homogeneity of variance. recall that in the formula we pool our sample variances. This is an okay thing to do if the variances are about the same. However, it isn't okay if they are very different. In our next lab we'll discuss a test to answer this question.


    Using SPSS to compute 2-independent samples t-tests






    If you have any questions, please feel free to contact me at jccutti@mail.ilstu.edu.