This page has moved to my new website.

Someone sent me an email describing a situation where an interaction effect in SPSS had a large p-value, but one of the individual components of that interaction had a small and statistically significant p-value.

This can occur in many statistical models where you are testing a factor or interaction that involves multiple degrees of freedom. You might be looking at a variable which indicates race/ethnicity. If there are five possible values for this variable (e.g., White, Black, Hispanic, Asian, and Other), then testing the effect of race in a statistical model involves four degrees of freedom. The hypothesis that you are testing is that the average outcome measure is the same for all five levels for race.

SPSS and other statistical software will then often present individual components of this hypothesis. They will select a reference category and then compare each of the remaining categories to that reference. SPSS choose the last or largest value of the category as the reference category, which may not always be exactly what you want. In the race/ethnicity example, that leads to four single degree of freedom comparisons:

- White versus Other;
- Black versus Other,
- Hispanic versus Other, and
- Asian versus Other.

It doesn't happen too often but sometimes the single degree of freedom comparisons are not consistent with the multiple degree of freedom test. What should you do in this situation?

First, make sure that you are looking at a logical reference category. The reference category might be a control group or it might be the largest group in your sample, or the youngest group. A lot depends on the context of the problem. One thing for sure is that the computer understands nothing about the context of your problem, so it only chooses a good reference category by dumb luck. In the example above, using Other as your reference category is probably the worst possible choice.

You should also apply a Bonferroni correction to the single degree of freedom components.
A p-value of 0.03 doesn't look quite as impressive if it is just one of five degrees of
freedom.

If you still get significance after applying Bonferroni, then I think you have to be honest
and mention the fact that your analysis produced ambiguous and conflicting results. Report
the p-values both for the multiple degrees of freedom test and for each of the individual
degrees of freedom test. Then let the reader decide.

Of course, with an interaction, you should also draw a graph, because certain types of
interactions, especially ones where the lines cross, are more interesting than others.

The reverse can happen as well. Suppose the test with multiple degrees of freedom has a
statistically significant p-value, but the p-values for all of the individual degree of
freedom tests are large. Here, you might want to do a bit of investigation. Maybe it is some
oddball combination (groups 1 and 4 combined versus groups 2, 3, and 5) that is showing a
large effect. This sort of thing might be missed by the particular single degree of freedom
tests that the software has chosen for you.

This page was written by Steve Simon while working at Children's Mercy Hospital. Although I do not hold the copyright for this material, I am reproducing it here as a service, as it is no longer available on the Children's Mercy Hospital website. Need more information? I have a page with general help resources. You can also browse for pages similar to this one at Category: Analysis of variance.