21 Big ideas in POL 138

Political scientists can use quantitative research to learn about the political world, by recording data, analyzing these data, and making inferences based on these data.


The hypothesis that a political scientist tests is called the null hypothesis. The p-value is a measure of strength against this null hypothesis. If the p-value is under p=0.05, that means that the observed association is unlikely to be due to random chance. This low p-value merely indicates a correlation: it does not by itself indicate causation.


Sometimes political scientists analyze data from a randomized experiment, in which participants are randomly assigned to groups, so that the groups are similar to each other before being treated differently. If, after the difference in treatment, the groups differ in an outcome, then that difference in outcome is due to some combination of…

  • the effect of the difference in treatment

  • random assignment error making it look like the difference in treatment had an effect

Political scientists can use a p-value to assess the probability that, if the difference in treatment had no effect, the observed difference in outcome or a larger difference in outcome would have occurred due to random assignment error.


These p-values range from…

  • 1, for no evidence against the null hypothesis…

  • to as close to 0 as possible…

with lower p-values indicating more evidence against the null hypothesis.


For a randomized experiment, we start with only two plausible explanations for any difference between groups:

  1. the difference in treatment
  2. random assignment error

A p-value under p=0.05 lets us eliminate random assignment error as a plausible explanation, so that we can conclude that the treatment caused the difference.


But sometimes political scientists analyze data that are not from a randomized experiment. For these types of studies, we need to approximate the all-else-equal aspect of the experiment in some way.


For example, a discontinuity design compares groups that are as similar to each other as possible, with only one major difference. Suppose that we are interested in how, if at all, the letter grade a student earns in a course affects that student’s evaluation of the course. Instead of comparing, say, the mean evaluation from all students who earned an A to the mean evaluation from all students who earned a B, we could compare the mean evaluation from all students who earned 90% (and thus barely earned an A) to the mean evaluation from all students who earned 89% (and thus got a B but also barely missed an A). The 90/89 comparison will be of students who are a lot more similar to each other, compared to the A/B comparison, which might compare (on average) students who earned an 85 to students who earned a 95.


Perhaps the most common type of non-experimental research is an observational study with statistical control. For instance, a political scientist might test the null hypothesis that the mean length of criminal sentences given to men equals the mean length of criminal sentences given to women. In this case, a p-value under p=0.05 would be sufficient to conclude that these means differ, but this p-value would NOT be sufficient to conclude that gender bias caused these means to differ. For example, if men receive longer criminal sentences than women receive on average, that might be due to unfair bias against men, but that also might be due to men committing more serious crimes than women commit.

For such an analysis, political scientists can use control variables to eliminate alternate explanations one by one. For example, we can compare the mean sentence length for men convicted of murder to the mean sentence length for women convicted of murder, and then compare the mean sentence length for men convicted of robbery to the mean sentence length for women convicted of robbery, and so forth for all crimes. If – for each type of crime, or at least on average across all crimes – the mean sentence length for men is longer than the mean sentence length for women, then we have addressed the type of crime as an alternate explanation for why the mean sentence length for men is longer than the mean sentence length for women.

In an observational study with statistical control, researchers must remove alternate explanations one-by-one before making an inference about what causes what. Thus, a p-value of 0.05 or less can be properly interpreted as evidence of causation only after…

  • all plausible explanations are controlled for, or

  • explanations that are not controlled for are not sufficient to plausibly explain any residual difference

In our example, we might have addressed the type of crime as an alternate explanation, but – before inferring unfair gender bias – we should address other alternate explanations, such as past criminal history.


Political scientists can use a meta-analysis, which combines results from studies that have been conducted to test a given hypothesis. The Law of Large Numbers indicates that, the larger a representative sample, the closer the mean of the sample is expected to be to the true mean of the population. Thus, the larger sample that is used in a meta-analysis means that the meta-analysis estimate should be expected to be closer to the true value than the estimate of any given study in the meta-analysis, presuming that the studies included in the meta-analysis are representative of all studies that have been conducted on the topic.

For a meta-analyses, larger studies should on average receive more weight in the calculation of the overall estimate, because larger studies tend to provide more evidence than smaller studies provide.


Quantitative reasoning isn’t easy, and it’s even more difficult because researchers can make errors and can try to mislead you. Common methods to mislead include selective reporting of results and biased research designs.


It’s good to be skeptical of research: a lot of research is poorly done, is in error, or has a research design that does not permit valid inferences. But it’s also good to not be unduly skeptical. For example, instead of merely not believing a research result because there might be alternate explanations, see whether you can think of a plausible alternate explanation.

Life often involves a lot of decisions. Better use of quantitative reasoning can provide you more and better information to help you make better decisions.