The library offers a range of helpful services. All of our appointments are free of charge and confidential.
Pearson correlation (sometimes referred to as Pearson’s r, bivariate correlation, or Pearson product-moment correlation coefficient) is used to determine the strength and direction of a relationship between two continuous variables. This test assumes both variables are approximately normally distributed.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Spearman’s rank-order correlation (i.e., the non-parametric alternative to a Pearson correlation). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a Correlations table that indicates the Pearson correlation (r) between the variables, the significance value (p), and the number of observations (n).
Pearson’s r can range from -1 (perfect negative) to +1 (perfect positive), and indicates the strength and direction of the relationship between the two variables; p indicates statistical significance, with < .05 generally considered statistically significant (i.e., indicating a significant correlation between the two variables). Here, we see a non-significant weak positive correlation between the two continuous variables.
Pearson correlations are best visualized using a scatterplot, which can be easily created by clicking Graphs, Chart Builder. Select Scatter / Dot, drag the desired graph type from the options to the blue text, drag and drop your variables to the x and y axis, and click OK. This plot allows you to visualize the linear relationship between your selected variables.
The one sample t-test is used to determine whether the mean of a single continuous variable differs from a specified constant. This test assumes that the observations are independent and that the data are normally distributed.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Wilcoxon signed rank test (i.e., the non-parametric alternative to the one sample t-test). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a One-Sample Statistics table that indicates some basic information about the variable you selected, a One-Sample Test table that indicates the result of the test (p < .05 is generally considered statistically significant, which would indicate that the variable mean differs from the test value), and a One-Sample Effect Sizes table that indicates both Cohen’s d and Hedges’ correction values.
The independent samples t-test (also referred to as a between subject’s t-test, student’s t-test, unpaired t-test, two-sample t-test…) is used to determine whether two groups’ means on the same continuous variable differ. This test assumes that the observations are independent, and are randomly sampled from normal distributions that have the same population variance.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Mann Whitney U test (i.e., the non-parametric alternative to the independent samples t-test). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a Group Statistics table that indicates some basic information about the variables you selected, a One-Sample Test table that indicates the result of the test (p < .05 is generally considered statistically significant, which would indicate that the variable mean differs from the test value), and a One-Sample Effect Sizes table that indicates Cohen’s d, Hedges’ correction, and Glass’s delta values.
Note that the One-Sample Test table includes a “Sig.” column for Levene’s test for equality of variances; if p < .05, the equality of variances assumption has been violated and you should use the “equal variances not assumed” row when interpreting the p-values found in the “Significance” column.
The paired samples t-test (also referred to as a dependent t-test, repeated samples t-test, matched pairs t-test…) is used to determine whether the means of two continuous variables (e.g., before and after treatment) from the same group of participants differ. This test assumes that the participants are independent from one another, while the two variables (e.g., measurements) are from the same participant. This test also assumes that the distribution of differences is normally distributed and there are no extreme outliers in the differences.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Wilcoxon signed-rank test (i.e., the non-parametric alternative to the paired-samples t-test). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. For this test, normality is often calculated on the DIFFERENCE SCORES between your two related groups. Ideally, your p-value should be > .05, your histogram should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plot should be fairly close to the line.
Running the above steps will generate the following output: a Paired Samples Statistics table that indicates some basic information about the variables you selected, a Paired Samples Correlations table that indicates the correlation between the variables you selected, a Paired Samples Test table that indicates the result of the test (p < .05 is generally considered statistically significant, which would indicate that the variable means of the groups differ), and a Paired Samples Effect Sizes table that indicates both Cohen’s d and Hedges’ correction values.
The one-way Analysis of Variance (ANOVA) is used to determine whether the means from three or more groups differ. This test assumes that each group is an independent random sample from normally distributed populations, and that the populations have equal variances.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Kruskal-Wallis H test (i.e., the non-parametric alternative to the one-way ANOVA). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. You will need to place your dependent variable in the “Dependent List” box, and your independent variable in the “Factor List” box. Ideally, your p-values should be > .05, your histograms should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plots should be fairly close to the line.
Running the above steps will generates a large output with multiple sections. In the Oneway section: a Test of Homogeneity of Variances table, an ANOVA table, an ANOVA Effect Sizes table, a multiple comparisons table, and profile plots. (NOTE: that there are other tables as well, but these are the critical tables).
The Tests of Homogeneity of Variances table shows you whether homogeneity has been violated (p < .05); if this is the case, a non-parametric test might be more appropriate.
The ANOVA table shows you whether you have a main effect of your dependent variable based on the independent variable groups. If your ANOVA is significant (p < .05), you need to run follow-up t-tests or multiple comparisons to determine which levels are significantly different.
The Multiple Comparisons table indicates which groups of the independent variable are different from each other (p < .05).
The Means Plots section displays a line graph or bar graph (based on your specification). Note: it’s always a great idea to generate a graph so you can visually try to interpret your data! Here, it looks like there might be a significant difference, but pay attention to the scale of the y-axis! If we adjust the scale, it’s easy to see why there is no significant difference indicated in the ANOVA or the multiple comparisons.
To generate the Tests of Normality table, use the Explore procedure (see the “Descriptive statistics” page for details) to generate the Shapiro-Wilk statistic. For a one-way ANOVA, you need to check normality for each group (i.e., for each level of the independent variable). If p < .05 for any of the groups, the assumption of normality has been violated. In this example, all of the groups pass the normality assumption (only one group shown).
The one-way repeated measures Analysis of Variance (ANOVA) is used to determine whether the means of three or more continuous variables OR measurements at three or more timepoints on the same continuous variable differ for one group. There are several assumptions of this test; two important ones you should consider are normality and sphericity.
Note that this is a parametric test; if the assumptions of this test are not met, you could / should instead run a Friedman test (i.e., the non-parametric alternative to the one-way repeated measures ANOVA). For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. You will need to place each of your groups in the “Dependent List” box. Ideally, your p-values should be > .05, your histograms should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plots should be fairly close to the line.
Running the above steps will generate the following output: Mauchly’s Test of Sphericity table, Tests of Within-Subjects Effects table, Pairwise Comparisons table, and Profile Plots (NOTE: that there are other tables as well, but these are the critical tables).
The Sphericity table indicates which line(s) to read on the Within-Subject Effects table: If sphericity has been violated (Sphericity table p < .05), reading the “Greenhouse-Geisser” line is common; if sphericity has not been violated (Sphericity table p > .05), read the “Sphericity assumed” line in the Within-Subjects Effects table.
The Tests of Within-Subjects Effects table shows you whether you have a main effect of your repeated measures variable (p < .05). If your ANOVA is significant (p < .05), you need to run follow-up t-tests or pairwise comparisons to determine which levels are significantly different. The Pairwise Comparisons table indicates which levels of the repeated measures variable are different from each other (p < .05).
The Profile Plots section displays a line graph or bar graph (based on your specification). Note: it’s always a great idea to generate a graph so you can visually try to interpret your data! Here, it looks like there is a significant difference, which is indicated in the ANOVA.
To generate the Tests of Normality table, use the Explore procedure (see the “Descriptive statistics” page for details) to generate the Shapiro-Wilk statistic. For a repeated-measures ANOVA, you need to check normality for each group (i.e., for each group of the dependent variable). If p < .05 for any of the groups, the assumption of normality has been violated. In this example, one of the groups (shown) passes the normality assumption, but the other two groups do not.
A factorial Analysis of Variance (factorial ANOVA) is used to determine whether the means from two or more variables / factors with two or more levels each differ (e.g., main effects), and whether any factors interact (e.g., interactions). If there are two factors, this is sometimes called a “two-way” ANOVA; if there are three factors, this is sometimes called a “three-way ANOVA”.
There are three separate kinds of factorial ANOVA: fully between factorial ANOVA (where both or all factors have levels that are entirely between-group), fully within factorial ANOVA (where both or all factors have levels that are entirely within-group), and mixed factorial ANOVA (where one or more factors are entirely between-group AND one or more factors are entirely within-group).
Note that all three kinds of factorial ANOVA are parametric tests; if the assumptions of the test are not met, there is no non-parametric equivalent. You could consider transforming your dependent variable in some way and then running the parametric test (and all assumptions) again, though this does not guarantee normality. For a refresh on how to check normality, read the “Explore” procedure on the Descriptive Statistics SPSS LibGuide page. You will need to place your within / repeated variable(s) in the “Dependent List” box, and your between / independent variable(s) in the “Factor List” box. Ideally, your p-values should be > .05, your histograms should approximate a normal distribution (i.e., a standard “bell-shaped curve”), and the points on your Q-Q plots should be fairly close to the line.
This is an extension of a one-way ANOVA, including two or more factors with two or more levels each that are fully “between group”. This test assumes normality and homogeneity.
Running the above steps will generate a large output with multiple sections. In the Univariate Analysis of Variance section: a Descriptive Statistics table that indicates some basic information about the variables you selected, a Test of Homogeneity of Variances table that indicates Levene’s statistic (if p < .05, the homogeneity assumption has been violated), and a Tests of Between-Subjects Effects table (the ANOVA table; if p < .05 for the independent variables or interaction, this indicates that at least one of the means is different than the others. This does NOT tell you which mean, so you need to follow up with t-tests or pairwise / multiple comparisons!).
If appropriate, you may also have an Estimated Marginal Means section showing a Pairwise Comparisons table, and / or a Post Hoc Tests section showing a Multiple Comparisons table. For both of these, if p < .05, this indicates a difference in the means between conditions, which could help explain a significant main effect or interaction in the ANOVA. Generally, we only look at these tables if you found significance in the associated main effect or interaction in the ANOVA OR if you had a specific prediction in advance of running the statistics about an effect you wanted to look at regardless of the results of the ANOVA.
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.