Experimental Design
Experimental Design
In the past twenty years, experimental research designs have become increasingly
popular in disciplines in the social sciences, such as political science, that do not have a
longstanding tradition of experimentation. Advocates of experiments argue that these
methods are the gold standard in assessing causality: whether a variable of interest is
actually caused by another variable. However, critics push back that causality in the
context of an experiment does not always translate well to causal processes in the real
world.
There is no single research design that is perfectly suited to test every hypothesis a
social scientist might assert: the best research matches the most effective research
design to the particular puzzle to be explored. The purpose of this module is to help
you understand what differentiates experiments from other forms of research designs
to facilitate your informed evaluation of the benefits and drawbacks of
experimentation, in order to assess whether an experiment is well-suited for your
research question.
By the end of this module, you should be comfortable with the following terminology:
fundamental problem of causal inference, treatment, control, random assignment,
internal validity, external validity, and convenience sample. You should also be aware
of the avenues open to you to conduct experiments for an independent research
project.
1. What is an Experiment?
In common parlance, people often say “experiment” when they mean “study.” All
experiments are studies, but not all studies (in fact the vast majority of studies) are not
experiments. In social science research methods, labeling a study as an experiment
means something very precise.
Morton and Williams (2010) write that an experiment is conducted when “a researcher
intervenes in the data generating process by purposefully manipulating elements of
that process” (p. 42). Researchers want as much control as possible over their
manipulations to minimize the interference of confounding factors. The most common
way of controlling confounding factors is to compare experimental results to a baseline
where all observable conditions are identical except for the presence of the
manipulation. In within-subjects designs, subjects serve as their own baseline: all facets
of the study are held constant and the same subjects experience all manipulations.
However, between-subjects designs are more versatile and more common, and the
selection of an appropriate baseline comparison, a control group, is essential.
Consequently, when most political scientists use the term “experiment,” they refer to a
study where subjects (or whatever entity we are studying) are randomly assigned to
different treatments, where the treatments are causal interventions. Druckman et al.
(2011) write “In contrast to modes of research that address descriptive or interpretive
questions, researchers design experiments to address causal questions. A causal
question invites a comparison between two states of the world: one in which some sort
of intervention is administered (a treated state, i.e. exposing a subject to a stimulus)
and another in which it is not (an untreated state)” (pg. 16).
Second, if you are interested in human behavior, experiments often provide the most
straightforward way to understand the mechanisms that explain why people do what
they do. Demonstrating a relationship between two variables in observational data
does not tell you how that relationship actually operates, while experimental
techniques can isolate the process of influence.
Finally, some research puzzles are more conducive to experiments than others. We
would never want to randomly assign individuals to experience civil war, nor would we
want to randomly assign some people to be disenfranchised, for example. However, a
special kind of study, natural experiments (described more below) can help researchers
leverage the concept of random assignment even if they are not the force actually
assigning the treatment. A creative experimentalist can design a treatment to ethically
test a wide variety of political and social phenomenon. In recent years, there have been
papers published in political science journals utilizing a wide variety of experimental
stimuli to test questions like a “get out the vote” experiment on 61-million Facebook
users.
Lab Experiments: The basic requirements for a lab experiment are that subjects are
recruited to a common physical location where the study takes place, and a researcher
directs the behavior of the subjects. Lab experiments in political science have their root
in the experimental techniques developed in other disciplines. Experimental and social
psychology have had an enormous influence on the field of political psychology, and
political psychology experiments typically adhere to the norms established in these
disciplines. Subjects are paid a flat fee (or given class credit if student convenience
samples are used) and researchers assume that subjects will behave sincerely.
Researchers often go to great length to make their studies seem naturalistic and may
use deception. Conversely, in experiments modeled after techniques used in
behavioral economics, subjects are paid based on the choices they make in the
experiment under the assumption that the payment motivates subjects to behave in a
more natural fashion. These experiments tend to be abstracted away from reality and
are used to test theories and formal models. Deception is rarely used. Finally, “lab in
the field” experiments occur when researchers bring the controlled environment of the
lab into a field environment, in an effort to conduct experiments on a wider variety of
populations and to increase the external validity of the results of the experiment.
Field Experiments: Field experiments occur when a researcher’s intervention takes place
in subjects’ natural environment. One of the most common types of field experiments
in political science is “get out the vote” experiments, where registered voters are
randomly assigned to receive an encouragement to vote in some form (mailing, door
hanger, phone call, etc.) and the turnout rate of the treatment group is compared to a
group that did not receive the encouragement. However, there are many other clever
and interesting manipulations that take place in the natural environment.
Implementing an Experiment
8. Finding Subjects
All research designs pose unique challenges. One of the biggest in experimental
designs is the need to recruit subjects to take your study. In an ideal world, every
experiment would be conducted on a random sample of the population of interest to
achieve both strong internal and external validity. But one of the main benefits of
experimentation is that you can achieve strong internal validity on a convenience
sample. Therefore, experiments disproportionately rely on samples that are not
randomly selected and researchers make no claim that their sample is representative.
If you have used the above samples to collect pilot results that are promising tests of
your hypotheses, with the support of a faculty member, you can put together a
proposal for the Time Share Experiments in the Social Sciences. This project, sponsored
by the National Science Foundation, allows researchers to submit experimental
proposals for consideration to be fielded on a representative sample of American
adults on an Internet survey platform.
First and foremost, studies involving human subjects always require approval from the
Institutional Review Board (IRB). Applications to the IRB must include documentation
that you have successfully completed the ethics training mandated by the federal
government. Depending on the nature of your experiment, your study may be exempt
from full review of the board, but that is a decision made by the chair of the IRB, not by
you as the researcher. You should allow at least three weeks for your study to be
approved before you can collect data.
One of the drawbacks of survey experiments that subjects take in the comfort of their
own environment is a lack of control on the part of the researcher in controlling that
environment. For example, if you are conducting your survey on an Internet sample,
subjects are able to browse the Internet or walk away from the computer while they
are taking your study. This is most problematic if there is reason to think that some
facet of your treatment might make subjects more likely to get distracted or visit other
webpages, for example if your treatment is very long or if you ask post-test questions
about political knowledge, where subjects might feel inclined to seek out the correct
answers online. Programs like Qualtrics have some built in functionalities to be able to
detect this.
However, in some instances, you may want to conduct a survey experiment in the
laboratory, either to exert more control over the experiment or because you want to
deliver a treatment that is not well suited to the online or phone format. In that
instance you may want to use the research lab facilities of the SSRMC. If you are doing
a lab experiment and you are interested in delivering media (images, audio, or videos)
to subjects in a laboratory environment, one of the most popular software programs to
do so is SuperLab. The SSRMC currently has licenses for this program on one
computer. An example honors thesis that relies on SuperLab for stimulus delivery can
be found here.
If you are interested in doing an experiment in the field, academics frequently partner
with outside organizations—such as campaigns or advocacy organizations—because it
is often difficult to get access to the large subject pools necessary to conduct these
studies. William & Mary undergraduates have successfully done this, and this honors
thesis is a great example.
10. Analyzing the Data
There are more sophisticated ways of analyzing the data from experiments, and
methodologists are constantly developing new ways to extract less biased estimates of
the causal effects in a study. While these more advanced approaches are beyond the
scope of this introductory module, the resources listed at the end of this module
contain more in-depth information.
Studies involving human subjects must go through ethics review for good reason, and
this is especially important in the case of experiments, where researchers manipulate
the environment or stimuli to which participants are exposed.
Different experimental traditions in the social sciences have different norms. One of
the biggest differences is in the instructions that researchers give participants about
the nature of the experimental tasks. Experiments rooted in psychology often allow
researchers to use mild deception in their instructions if the researcher thinks that
knowing the true purpose of the study would alter the way that participants behave.
The SSRMC allows deception in studies, as long as that deception is approved by the
Institutional Review Board. However, deception is almost always avoided in
experiments rooted in economics. A second major difference between economic and
psychology experiments is whether (or how) subjects are incentivized for their
participation.
The classic tradeoff in experimental design is between internal validity and external
validity. While experiments have high internal validity, to varying degrees, they may lack
external validity, the ability of a researcher to make claims about how the results of the
study would generalize and hold up in different contexts. One particularly common
generalizability concern stems from differences in the sample used in your experiment
compared the population to which you want to generalize. When is it problematic to
generalize the findings of an experiment conducted on a convenience (often, student)
population? First, it is important to know how the student population differs from a
more representative population. The obvious answers are in age, education level, and
geographic location.
But other factors can matter as well. The key question to ask is “how else are college
students different in a way that should affect the strength or direction of the causal
relationship I am testing?” If your convenience sample is different in a way that makes it
harder to find the relationship you observe, then you can assert that your study likely
underestimates the relationship between the variables in a more representative
population (a testable proposition!) However, if your sample makes it easier to find
effects, then generalizability concerns become more serious. Sometimes, these
concerns are very large. For example, college students are particularly susceptible to
conformity (Sears, 1986), which could be important depending on the nature of your
study. The Omnibus Project draws from courses in the government and international
relations program, suggesting that our participants have a greater interest and level of
knowledge about politics. These factors may help or hurt your ability to make claims
about how general your findings are.
12. Resources
There are many excellent resources available online, through SWEM library, and
through the SSRMC research methods collection. The module below draws on
information from these sources, but there is much more detail available in the original
sources.
Experimental Design
Druckman, James N. et al. 2011. Cambridge Handbook of Experimental Political Science.
Rebecca Morton and Kenneth Williams. 2010. From Nature to the Lab: Experimental Political
Science and the Study of Causality
McDermott, Rose. 2002. "Experimental Methods in Political Science." Annual Review of
Political Science 5:31-61.
McDermott, Rose. 2002. "Experimental Methodology in Political Science." Political Analysis
10(4):325-342
Druckman, James, Donald Green, James Kuklinkski, and Arthur Lupia. 2006. "The Growth and
Development of Experimental Research in Political Science," American Political Science
Review 100:627-635
Gaines, Brian et al. 2006. "The Logic of the Survey Experiment Reexamined." Political
Analysis
Imai,Kosuke et al. 2011. "Unpacking the Black Box of Causality: Learning about Causal
Mechanisms from Experimental and Observational Studies." American Political Science Review
Field Experiments
Green and Gerber's Field Experiments book
Yales ISPS Field Experiments Initiative as well as there Data Archive