By Benjamin Frimodig, published 2020
Chi-square (or χ2) tests draw inferences and test for relationships between categorical variables, that is a set of data points that fall into discrete categories with no inherent ranking.
There are three types of Chi-square tests, tests of goodness of fit, independence and homogeneity. All three tests also rely on the same formula to compute a test statistic.
All three function by deciphering relationships between observed sets of data and theoretical—or “expected”—sets of data that align with the null hypothesis.
The Chi-square goodness of fit test is used to compare a randomly collected sample containing a single, categorical variable to a larger population. This test is most commonly used to compare a random sample to the population from which it was potentially collected.
The test begins with the creation of a null and alternative hypothesis. In this case, the hypotheses are as follows:
Null Hypothesis (Ho): The collected data is consistent with the population distribution.
Alternative Hypothesis (Ha): The collected data is not consistent with the population distribution.
The next step is to create a contingency table that represents how the data would be distributed if the null hypothesis were exactly correct.
The sample’s overall deviation from this theoretical/expected data will allow us to draw a conclusion, with more severe deviation resulting in smaller p-values.
The Chi-square test for independence looks for an association between two categorical variables within the same population. Unlike the goodness of fit test, the test for independence does not compare a single observed variable to a theoretical population, but rather two variables within a sample set to one another.
The hypotheses for a Chi-square test of independence are as follows:
Null Hypothesis (Ho): There is no association between the two categorical variables in the population of interest.
Alternative Hypothesis (Ha): There is no association between the two categorical variables in the population of interest.
The next step is to create a contingency table of expected values that reflects how a data set that perfectly aligns the null hypothesis would appear.
The simplest way to do this is to calculate the marginal frequencies of each row and column; the expected frequency of each cell is equal to the marginal frequency of the row and column that corresponds to a given cell in the observed contingency table divided by the total sample size.
Contingency table (also known as two-way tables) are grids in which Chi-square data is organized and displayed. They provide a basic picture of the interrelation between two variables and can help find interactions between them.
In contingency tables, one variable and each of its categories are listed vertically and the other variable and each of its categories are listed horizontally.
Additionally, including column and row totals, also known as “marginal frequencies”, will help facilitate the Chi-square testing process.
In order for the Chi-square test to be considered trustworthy, each cell of your expected contingency table must have a value of at least five.
Each Chi-square test will have one contingency table representing observed counts (see Fig. 1) and one contingency table representing expected counts (see Fig. 2).
Figure 1. Observed table (which contains the observed counts).
To obtain the expected frequencies for any cell in any cross- tabulation in which the two variables are assumed independent, multiply the row and column totals for that cell and divide the product by the total number of cases in the table.
Figure 2. Expected table (what we expect the two-way table to look like if the two categorical variables are independent).
Calculate the chi square statistic (χ2) by completing the following steps:
The chi-square statistic tells you how much difference exists between the observed count in each table cell to the counts you would expect if there were no relationship at all in the population.
A very small chi square test statistic means means there is a high correlation between the observed and expected values. Therefore, the sample data is a good fit for what would be expected in the general population.
In theory, if the observed and expected values were equal (no difference) then the chi-square statistic would be zero — but this is unlikely to happen in real life.
A very large chi square test statistic means that the sample data (observed values) does not fit the population data (expected values) very well. In other words, there isn't a relationship.
To report a chi square output in an APA style results section, always rely on the following template:
χ2 (degrees of freedom, N = sample size) = chi-square statistic value, p = p value.
In the case of the above example, the results would be written as follows:
A chi-square test of independence showed that there was a significant association between gender and post graduation education plans, χ2 (4, N = 101) = 54.50, p < .001.
For a chi-square test, a p-value that is less than or equal to the .05 significance level indicates that the observed values are different to the expected values.
Thus, low p-values (p< < .05) indicate a likely difference between the theoretical population and the collected sample. You can conclude that a relationship exists between the categorical variables.
Remember that p-values do not indicate the odds that the null hypothesis is true, but rather provides the probability that one would obtain the sample distribution observed (or a more extreme distribution) if the null hypothesis was in fact true.
A level of confidence necessary to accept the null hypothesis can never be reached. Therefore, conclusions must choose to either fail to reject the null or accept the alternative hypothesis, depending on the calculated p-value.
The four steps below show you how to analyse your data using a chi-square goodness-of-fit test in SPSS (when you have hypothesised that you have equal expected proportions).
Step 1: Analyze > Nonparametric Tests > Legacy Dialogs > Chi-square... on the top menu as shown below:
Step 2: Move the variable indicating categories into the “Test Variable List:” box.
Step 3: If you want to test the hypothesis that all categories are equally likely, click “OK.”
Step 4: Specify the expected count for each category by first clicking the “Values” button under “Expected Values.”
Step 5: Then, in the box to the right of “Values,” enter the expected count for category 1 and click the “Add” button. Now enter the expected count for category 2 and click “Add.” Continue in this way until all expected counts have been entered.
Step 6: Then click “OK.”
The four steps below show you how to analyse your data using a chi-square test of independence in SPSS Statistics.
Step 1: Open the Crosstabs dialog (Analyze > Descriptive Statistics > Crosstabs).
Step 2: Select the variables you want to compare using the chi square test. Click one variable in the left window and then click the arrow at the top to move the variable. Select the row variable, and the column variable.
Step 3: Click Statistics (a new pop up window will appear). Check Chi-square, then click Continue.
Step 4: (Optional) Check the box for Display clustered bar charts.
Step 5: Click OK.
The Chi-square test for homogeneity is organized and executed exactly the same as the test for independence. The main difference to remember between the two is that the test for independence looks for an association between two categorical variables within the same population, while the test for homogeneity determines if the distribution of a variable is the same in each of several populations (thus allocating population itself as the second categorical variable).
The hypotheses for a Chi-square test of independence are as follows:
Null Hypothesis (Ho): There is no difference in the distribution of a categorical variable for several populations or treatments.
Alternative Hypothesis (Ha): There is a difference in the distribution of a categorical variable for several populations or treatments.
The difference between these two tests can be a bit tricky to determine especially in practical applications of a Chi-square test. A reliable rule of thumb is to determine how the data was collected.
If the data consists of only one random sample with the observations classified according to two categorical variables, it is a test for independence. If the data consists of more than one independent random sample, it is a test for homogeneity.
Ben is a senior at Harvard College studying History and Science with a minor in Global Health and Health Policy. Ben is most interested in the intersections between psychology and history and hopes to pursue a career in the field of mental healthcare.
Frimodig, B. (2020, Oct 20). Chi square test. Simply Psychology. https://www.simplypsychology.org/chi-square.html
Home | About | A-Z Index | Privacy Policy | Contact Us
This workis licensed under a Creative Commons Attribution-Noncommercial-No Derivative Works 3.0 Unported License.
Company Registration no: 10521846