The library offers a range of helpful services. All of our appointments are free of charge and confidential.
Pearson correlation (sometimes referred to as Pearson’s r, bivariate correlation, or Pearson product-moment correlation coefficient) is used to determine the strength and direction of a relationship between two continuous variables. This test assumes both variables are approximately normally distributed.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Spearman’s rank-order correlation (i.e., the non-parametric alternative to a Pearson correlation). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a Correlations table that indicates the Pearson correlation (r) between the variables, the significance value (p), and the number of observations (n).
Pearson’s r can range from -1 (perfect negative) to +1 (perfect positive), and indicates the strength and direction of the relationship between the two variables; p indicates statistical significance, with < .05 generally considered statistically significant (i.e., indicating a significant correlation between the two variables). Here, we see a non-significant weak positive correlation between the two continuous variables.
Pearson correlations are best visualized using a scatterplot, which can be easily created by clicking Graphs, Chart Builder. Select Scatter / Dot, drag the desired graph type from the options to the blue text, drag and drop your variables to the x and y axis, and click OK. This plot allows you to visualize the linear relationship between your selected variables.
The one sample t-test is used to determine whether the mean of a single continuous variable differs from a specified constant. This test assumes that the observations are independent and that the data are normally distributed.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Wilcoxon signed rank test (i.e., the non-parametric alternative to the one sample t-test). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a One-Sample Statistics table that indicates some basic information about the variable you selected, a One-Sample Test table that indicates the result of the test (p < .05 is generally considered statistically significant, which would indicate that the variable mean differs from the test value), and a One-Sample Effect Sizes table that indicates both Cohen’s d and Hedges’ correction values.
The independent samples t-test (also referred to as a between subject’s t-test, student’s t-test, unpaired t-test, two-sample t-test…) is used to determine whether two groups’ means on the same continuous variable differ. This test assumes that the observations are independent, and are randomly sampled from normal distributions that have the same population variance.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Mann Whitney U test (i.e., the non-parametric alternative to the independent samples t-test). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a Group Statistics table that indicates some basic information about the variables you selected, a One-Sample Test table that indicates the result of the test (p < .05 is generally considered statistically significant, which would indicate that the variable mean differs from the test value), and a One-Sample Effect Sizes table that indicates Cohen’s d, Hedges’ correction, and Glass’s delta values.
Note that the One-Sample Test table includes a “Sig.” column for Levene’s test for equality of variances; if p < .05, the equality of variances assumption has been violated and you should use the “equal variances not assumed” row when interpreting the p-values found in the “Significance” column.
The paired samples t-test (also referred to as a dependent t-test, repeated samples t-test, matched pairs t-test…) is used to determine whether the means of two continuous variables (e.g., before and after treatment) from the same group of participants differ. This test assumes that the participants are independent from one another, while the two variables (e.g., measurements) are from the same participant. This test also assumes that the distribution of differences is normally distributed and there are no extreme outliers in the differences.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Wilcoxon signed-rank test (i.e., the non-parametric alternative to the paired-samples t-test). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. For this test, normality is often calculated on the DIFFERENCE SCORES between your two related groups. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a Paired Samples Statistics table that indicates some basic information about the variables you selected, a Paired Samples Correlations table that indicates the correlation between the variables you selected, a Paired Samples Test table that indicates the result of the test (p < .05 is generally considered statistically significant, which would indicate that the variable means of the groups differ), and a Paired Samples Effect Sizes table that indicates both Cohen’s d and Hedges’ correction values.
The one-way Analysis of Variance (ANOVA) is used to determine whether the means from three or more groups differ. This test assumes that each group is an independent random sample from normally distributed populations, and that the populations have equal variances.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Kruskal-Wallis H test (i.e., the non-parametric alternative to the one-way ANOVA). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. You will need to place your dependent variable in the “Dependent List” box, and your independent variable in the “Factor List” box. Ideally, your p-values should be > .05, your histograms should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plots should be fairly close to the line.
Running the above steps will generates a large output with multiple sections. In the Oneway section: a Test of Homogeneity of Variances table, an ANOVA table, an ANOVA Effect Sizes table, a multiple comparisons table, and profile plots. (NOTE: that there are other tables as well, but these are the critical tables).
The Tests of Homogeneity of Variances table shows you whether homogeneity has been violated (p < .05); if this is the case, a non-parametric test might be more appropriate.
The ANOVA table shows you whether you have a main effect of your dependent variable based on the independent variable groups. If your ANOVA is significant (p < .05), you need to run follow-up t-tests or multiple comparisons to determine which levels are significantly different.
The Multiple Comparisons table indicates which groups of the independent variable are different from each other (p < .05).
The Means Plots section displays a line graph or bar graph (based on your specification). Note: it’s always a great idea to generate a graph so you can visually try to interpret your data! Here, it looks like there might be a significant difference, but pay attention to the scale of the y-axis! If we adjust the scale, it’s easy to see why there is no significant difference indicated in the ANOVA or the multiple comparisons.
To generate the Tests of Normality table, use the Explore procedure (see the “Descriptive statistics” page for details) to generate the Shapiro-Wilk statistic. For a one-way ANOVA, you need to check normality for each group (i.e., for each level of the independent variable). If p < .05 for any of the groups, the assumption of normality has been violated. In this example, all of the groups pass the normality assumption (only one group shown).
The one-way repeated measures Analysis of Variance (ANOVA) is used to determine whether the means of three or more continuous variables OR measurements at three or more timepoints on the same continuous variable differ for one group. There are several assumptions of this test; two important ones you should consider are normality and sphericity.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Friedman test (i.e., the non-parametric alternative to the one-way repeated measures ANOVA). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. You will need to place each of your groups in the “Dependent List” box. Ideally, your p-values should be > .05, your histograms should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plots should be fairly close to the line.
Running the above steps will generate the following output: Mauchly’s Test of Sphericity table, Tests of Within-Subjects Effects table, Pairwise Comparisons table, and Profile Plots (NOTE: that there are other tables as well, but these are the critical tables).
The Sphericity table indicates which line(s) to read on the Within-Subject Effects table: If sphericity has been violated (Sphericity table p < .05), reading the “Greenhouse-Geisser” line is common; if sphericity has not been violated (Sphericity table p > .05), read the “Sphericity assumed” line in the Within-Subjects Effects table.
The Tests of Within-Subjects Effects table shows you whether you have a main effect of your repeated measures variable (p < .05). If your ANOVA is significant (p < .05), you need to run follow-up t-tests or pairwise comparisons to determine which levels are significantly different. The Pairwise Comparisons table indicates which levels of the repeated measures variable are different from each other (p < .05).
The Profile Plots section displays a line graph or bar graph (based on your specification). Note: it’s always a great idea to generate a graph so you can visually try to interpret your data! Here, it looks like there is a significant difference, which is indicated in the ANOVA.
To generate the Tests of Normality table, use the Explore procedure (see the “Descriptive statistics” page for details) to generate the Shapiro-Wilk statistic. For a repeated-measures ANOVA, you need to check normality for each group (i.e., for each group of the dependent variable). If p < .05 for any of the groups, the assumption of normality has been violated. In this example, one of the groups (shown) passes the normality assumption, but the other two groups do not.
A factorial Analysis of Variance (factorial ANOVA) is used to determine whether the means from two or more variables / factors with two or more levels each differ (e.g., main effects), and whether any factors interact (e.g., interactions). If there are two factors, this is sometimes called a “two-way” ANOVA; if there are three factors, this is sometimes called a “three-way ANOVA”.
There are three separate kinds of factorial ANOVA: fully between factorial ANOVA (where both or all factors have levels that are entirely between-group), fully within factorial ANOVA (where both or all factors have levels that are entirely within-group), and mixed factorial ANOVA (where one or more factors are entirely between-group AND one or more factors are entirely within-group).
Note that all three kinds of factorial ANOVA are parametric tests; if the assumptions of the test are not met, there is no non-parametric equivalent. You could consider transforming your dependent variable in some way and then running the parametric test (and all assumptions) again, though this does not guarantee normality. For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. You will need to place your within / repeated variable(s) in the “Dependent List” box, and your between / independent variable(s) in the “Factor List” box. Ideally, your p-values should be > .05, your histograms should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plots should be fairly close to the line. Additionally, your boxplots should indicate no outliers in any group.
This is an extension of a one-way ANOVA, including two or more factors with two or more levels each that are fully “between” (i.e., each participant / sample is in only one condition). This test assumes that each participant / sample is in only one condition (independence of observations), that there should be no outliers, that your dependent variable is approximately normally distributed in each condition (normality), and that each condition has approximately equal variances (homogeneity).
Running the above steps will generate a large output with multiple sections. In the Univariate Analysis of Variance section: a Descriptive Statistics table that indicates some basic information about the variables you selected, a Test of Homogeneity of Variances table that indicates Levene’s statistic (if p < .05, the homogeneity assumption has been violated), and a Tests of Between-Subjects Effects table (the ANOVA table; if p < .05 for the independent variables or interaction, this indicates that at least one of the means is different than the others. This does not tell you which means are different from each other, so you need to follow up with pairwise comparisons, multiple comparisons, or t-tests!).
Importantly, when p > .05 in the Tests of Between-Subjects Effects table, we cannot say that there is “no difference" between the level(s); we can only say that we failed to find a difference. If p < .05, we could conclude that there is a difference somewhere between the levels, though we would need to follow this up with additional testing to determine which levels are significantly different from each other. If you find a significant main effect for a factor that only has two levels, you can simply look at the descriptives to know which is higher / lower (as the main effect already tells you that these are different). If your factorial ANOVA reveals p < .05 for a main effect for a factor that has more than two levels, or an interaction, you need to run pairwise post hoc tests to determine the location of the difference(s). To do so, look at the Estimated Marginal Means section showing a Pairwise Comparisons table, and / or a Post Hoc Tests section showing a Multiple Comparisons table. For both of these, if p < .05, this indicates a difference in the means between conditions, which could help explain a significant main effect or interaction in the ANOVA.
Looking at a graph can also help visualize where the differences may be, so be sure to look at the Profile Plots section if you requested a graph.
This is an extension of a repeated measures ANOVA, including two or more factors with two or more levels each that are fully “within” (i.e., each participant / sample is in each condition). This test assumes that there should be no outliers, that your dependent variable is approximately normally distributed in each condition (normality), and that the differences between all conditions must be approximately equal (sphericity).
Running the above steps will generate a large output with multiple sections. In the General Linear Model section: a Descriptive Statistics table that indicates some basic information about the variables you selected, a Test of Sphericity table that indicates Mauchly’s statistic (if p < .05, the sphericity assumption has been violated), and a Tests of Within-Subjects Effects table (the ANOVA table; if p < .05 for the independent variables or interaction, this indicates that at least one of the means is different than the others. This does NOT tell you which means are different from each other, so you need to follow up with pairwise comparisons, multiple comparisons, or t-tests!).
Note. If the Sphericity assumption was violated (p < .05) according to Mauchly’s test, you should read the “Greenhouse-Geisser” lines in the “Tests of Within-Subjects Effects” table when interpreting your ANOVA.
Importantly, when p > .05 in the Tests of Within-Subjects Effects table, we cannot say that there is “no difference" between the level(s); we can only say that we failed to find a difference. If p < .05, we could conclude that there is a difference somewhere between the levels, though we would need to follow this up with additional testing to determine which levels are significantly different from each other. If you find a significant main effect for a factor that only has two levels, you can simply look at the descriptives to know which is higher / lower (as the main effect already tells you that these are different). If your factorial ANOVA reveals p < .05 for a main effect for a factor that has more than two levels, or an interaction, you need to run pairwise post hoc tests to determine the location of the difference(s). To do so, look at the Estimated Marginal Means section showing a Pairwise Comparisons table. If p < .05, this indicates a difference in the means between conditions, which could help explain a significant main effect or interaction in the ANOVA.
Looking at a graph can also help visualize where the differences may be, so be sure to look at the Profile Plots section if you requested a graph.
A mixed factorial ANOVA (also sometimes called a split-plot design) is a combination of a one-way ANOVA and a repeated measures ANOVA, where at least one of the factors is fully “between” (i.e., each participant / sample is in only one condition for this factor) and one of the factors is fully “within” (i.e., each participant / sample is in each condition for this factor). This test assumes that there should be no outliers, that your dependent variable is approximately normally distributed in each condition (normality), that each condition has approximately equal variances (homogeneity), and that the differences between all conditions must be approximately equal (sphericity).
Running the above steps will generate a large output with multiple sections, including the following relevant tables in the General Linear Model section: a Descriptive Statistics table that indicates some basic information about the variables you selected, a Mauchly’s Test of Sphericity table that indicates Mauchly’s statistic (if p < .05, the sphericity assumption has been violated), a Tests of Within-Subjects Effects table (the ANOVA table; if p < .05 for the independent variable(s) or interaction(s), this indicates that at least one of the means is different than the others. This does not tell you which means are different from each other, so you need to follow up!), a Test of Homogeneity of Variances table that indicates Levene’s statistic (if p < .05, the homogeneity assumption has been violated), and a Tests of Between-Subjects Effects table (the ANOVA table; if p < .05 for the independent variable(s) or interaction(s), this indicates that at least one of the means is different than the others. This does not tell you which means are different from each other, so you need to follow up!).
Note: If the Sphericity assumption was violated (p < .05) according to Mauchly’s test, you should read the “Greenhouse-Geisser” lines in the “Tests of Within-Subjects Effects” table when interpreting your ANOVA.
Importantly, when p > .05 in the Tests of Within-Subjects Effects table or the Tests of Between-Subjects Effects table, we cannot say that there is “no difference" between the level(s); we can only say that we failed to find a difference. If p < .05, we could conclude that there is a difference somewhere between the levels, though we would need to follow this up with additional testing to determine which levels are significantly different from each other. If you find a significant main effect for a factor that only has two levels, you can simply look at the descriptives to know which is higher / lower (as the main effect already tells you that these are different). If your factorial ANOVA reveals p < .05 for a main effect for a factor that has more than two levels, or an interaction, you need to run pairwise post hoc tests to determine the location of the difference(s). To do so, look at the Estimated Marginal Means section showing a Pairwise Comparisons table (within factors with > 2 levels OR interactions), and / or a Post Hoc Tests section showing a Multiple Comparisons table (between factors with > 2 levels). For both of these, if p < .05, this indicates a difference in the means between conditions, which could help explain a significant main effect or interaction in the ANOVA.
Looking at a graph can also help visualize where the differences may be, so be sure to look at the Profile Plots section if you requested a graph.
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.