Validity of Research Results: How Strong Is The Relationship?
Validity of Research Results: How Strong Is The Relationship?
Internal Validity
When I hear the term "internal validity" the word cause always comes into my mind.
That's because internal validity is defined as the "approximate validity with which we
infer that a relationship between two variables is causal" (Cook and Campbell, 1979.
P.37).
A good synonym for the term internal validity is causal validity because that is
what internal validity is all about.
If you can show that you have high internal validity (i.e., high causal validity)
then you can conclude that you have strong evidence of causality; however, if you
have low internal validity then you must conclude that you have little or no
evidence of causality.
Types of Causal Relationships
There are two different types of causal relationships: causal description and causal
explanation.
Causal description involves describing the consequences of manipulating an
independent variable.
In general, causal description involves showing that changes in variable X (the
IV) cause changes in variable Y (the DV): X---->Y
Causal explanation involves more than just causal description. Causal explanation
involves explaining the mechanisms through which and the conditions under
which a causal relationship holds. This involves the inclusion (in your research
study) of mediating or intervening variables and moderator variables. Mediating
and moderator variables are defined in Chapter Two in Table 2.2 (on page 36). To
see Table 2.2, click here.
Criteria for Inferring Causation
There are three main conditions that are always required if you want to make a claim that
changes in one variable cause changes in another variable. We call these the three
necessary conditions for causality.
These three conditions are summarized below in Table 11.1:
If you want to conclude that X causes Y you must make sure that the three above
necessary conditions are met. It is also helpful if you have a theoretical rationale
explaining the causal relationship.
For example, there is a correlation between coffee drinking and likelihood of
having a heart attack. One big problem with concluding that coffee drinking
causes heart attacks is that cigarette smoking is related to both of these variables
(i.e., we have a Condition 3 problem). In particular, people who drink little coffee
are less likely to smoke cigarettes than are people who drink a lot of coffee.
Therefore, perhaps the observed relationship between coffee drinking and heart
attacks is the result of the extraneous variable of smoking. The researcher would
have to "control for" smoking in order to determine if this rival explanation
accounts for the original relationship.
In this design, a group is pretested, then a treatment is administered, and then the people
are post tested. For example, you could measure your students' understanding of history
at the beginning of the term, then you teach them history for the term, and then you
measure them again on their understanding of history at the end of the term.
The second weak design to remember for this chapter is called the posttest-only
design with nonequivalent groups. In this lecture, I will also refer to this design
as a two-group design and sometimes as a multigroup design (since it has more
than one group).
XTreatment
O2
---------------------XControl
O2
In this design, there is no pretest, one group gets the treatment and the other group gets
no treatment or some different treatment, and both groups are post tested (e.g., you teach
two classes history for a quarter and measure their understanding at the end for
comparison). Furthermore, the groups are found wherever they already exist (i.e.,
participants are not randomly assigned to these groups).
In comparing the two designs just mentioned note that the comparison in the one
group design is the participants' pretest scores with their posttest scores. The
comparison in the two group design is between the two groups' posttest scores.
Some researchers like to call the point of comparison the "counterfactual." In the
one-group pretest- posttest design shown above the counterfactual is the pretest.
In the two-group design shown above the counterfactual is the posttest of the
control group.
Remember this key point: In each of the multigroup research designs (designs that
include more than one group of participants), you want the different groups to be
the same on all extraneous variables and different ONLY on the independent
variable (e.g., such that one group gets the treatment and the other group does
not). In other words, you want the only systematic difference between the groups
to be exposure to the independent variable.
For example, the principal may come into the experimental classroom during the
research study which alters the outcome.
The history effect is a threat for the one group design but it is not a threat for the
multigroup group design.
You probably want to know why this it true. Well, in the one group design
(shown above) you take as your measure of the effect of the treatment the
difference in the pretest and posttest scores. In this case, this all or part of the
difference could be due to a history effect; therefore, you don't know whether the
change in scores is due to the treatment or to the history effect. They are
confounded.
The basic history effect is not a threat to the two group design (shown above)
because now you are comparing the treatment group to a comparison group, and
as long as the history effect occurs for both groups the difference between the
two groups will not be because of a history effect.
groups will not be due to testing. The two groups do differ on exposure to the
treatment (i.e., one group gets the treatment and the other group does not).
The fifth threat to internal validity is called instrumentation.
Instrumentation refers to any change that occurs in the way the dependent variable
is measured in the research study.
For example, let's say that one person does your pretest assessment of students'
racial stereotyping but you have a different person do your posttest assessment of
students' stereotyping. Also assume that the second person tends to overlook
much stereotyping but that the first person picks up on all stereotyping. The
problem is that perhaps much of the positive gain occurring from the pretest to the
posttest is due to the posttest assessment not picking up on the use of
stereotyping.
Therefore in the one group design, you will not know if their improvement from
pretest to posttest is due to your treatment for reducing stereotyping or if it is due
to an instrumentation effect.
Instrumentation is not a threat in the two group design because as long as the
people in both groups are affected equally by the instrumentation effect, the
difference between the two groups will not be due to instrumentation.
The sixth threat to internal validity is called regression artifacts (or regression to the
mean).
Regression artifacts refers to the tendency of very high pretest scores to become
lower and for very low pretest scores to become higher on post testing.
You should always be on the lookout for regression to the mean when you select
participants based on extreme (very high or very low) test scores.
For example, let's say that you select people who have extremely high scores on
your racial stereotyping test. Some of these scores are probably artificially high
because of transient factors and a lack of perfect reliability. Therefore, if
stereotyping goes down from pretest to posttest, some or all of the change may be
due to a regression artifact.
Therefore, in the one group design you will not know if improvement from pretest
to posttest is due to your treatment or if it is due to a regression artifact.
Regression artifacts is not a threat in the two group design because as long as the
people in both groups are affected equally by the statistical regression effect, the
difference between the two groups will not be due to regression to the mean.
The seventh threat to internal validity is called differential selection.
Differential selection only applies to multigroup designs. It refers to selecting
participants for the various groups in your study that have different
characteristics.
Remember, we want our groups to be the same on all variables except the
treatment variable; the treatment variable is the only variable that we want to be
systematically different for the groups.
Table 8.1 list a few of the may characteristics on which the students in the
different groups may differ (e.g., age, anxiety, gender, intelligence, reading
ability, etc.).
Unlike the previous five threats, selection is not an internal validity problem
for the one group design but it is a problem for the two or multigroup design.
Looking at the definition again, you can see that selection is defined for two or
multigroup designs. It is not relevant to the internal validity of the single group
design.
As an example, assume that you select two classes for your study on reducing
racial stereotyping. You use two fifth grade classes as your groups. One group
will get your treatment and the other will act as a control. The problem is that
these two groups of students may differ on variables other than your treatment
variable and any differences found at the posttest may be due to these "differential
selection" differences rather than being due to your treatment.
The eight threat to internal validity is called differential attrition (it is also sometimes
called mortality). Attrition simply refers to participants dropping out of your research
study.
Differential attrition is the differential loss of participants from the various
comparison groups.
Just like the last threat, differential attrition is a problem for two or multigroup
design but not for the single group design. (Notice the word differential in the
definition.)
For example, assume again that you are doing a study on racial stereotyping. Do
you see how your result would be compromised if the kind of children that are
most likely to have racial stereotypes drop out of one of your groups but not the
other group? Obviously, the difference observed at the post test may now be the
result of differential attrition.
The ninth threat to internal validity is actually a set of threats. This set is called additive
and interactive effects.
Additive and interactive effects refers to the fact that the threats to validity can
combine to produce a bias in the study which threatens our ability to conclude that
the independent variable is the cause of differences between groups on the
dependent variable. They only apply to two or multigroup designs; they do not
apply to the one-group design.
These threats occur when the different comparison groups are affected differently
(or differentially) by one of the earlier threats to internal validity (i.e., history,
maturation, testing, instrumentation, or statistical regression).
A selection-history effect occurs when an event occurring between the pretest and
posttest differentially affects the different comparison groups. You can think of
this as what could be called a differential history effect.
A selection-maturation effect occurs if the groups mature at different rates. For
example, first grade students may tend to naturally change in reading ability
during the school year more than third grade students. Hence, part of any
observed differences in the reading ability of the two groups at the posttest may
be due to maturation. You can think of this a what could be called a differential
maturation effect.
You now should be able to construct similar examples demonstrating the
following:
Selection-testing effect (where testing affects the groups differently)
Selection-instrumentation effect (where instrumentation occurs differentially)
Selection-regression artifacts effect (where regression to the mean occurs
differentially).
Remember that the key for the selection-effects is that the groups must be affected
differently by the particular threat to internal validity.
External Validity
External validity has to do with the degree to which the results of a study can be
generalized to and across populations of persons, settings, times, outcomes, and treatment
variations.
A good synonym for external validity is generalizing validity because it always
has to do with how well you can generalize research results.
The major types of external validity are population validity, ecological validity,
temporal validity, temporal validity, treatment variation validity, and outcome
validity. I will discuss each of these now...
Population Validity
The first type of external validity is called population validity.
Population validity is the ability to generalize the study results to individuals who
were not included in the study.
The issues are how well you can generalize your sample results to a population,
and how well you can generalize your sample results across the different kinds if
people in the larger population.
Generalizing from a sample to a population can be provided through random
selection techniques (i.e., a good sample lets you generalize to a population, as
you learned in the earlier chapter on sampling).
Generalizing across populations is present when the result (e.g., the effectiveness
of a particular teaching technique) works across many different kinds of people
(it works for many sub populations). This is the issue of "how widely does the
finding apply?" If the finding applied to every single individual in the population
then it would have full population validity. Research results that apply broadly are
welcome to practitioners because it makes their jobs easier.
Both of these two kinds of population validity are important; however, some
methodologists (such as Cook and Campbell) are more concerned about
generalizing across populations. That is, they want to know how widely a finding
applies.
Ecological Validity
Ecological validity is present to the degree that a result generalizes across different
settings.
For example, let's say that you find that a new teaching technique works in urban
schools. You might also want to know if the same technique works in rural
schools and suburban schools. That is, you would want to know if the technique
works across different settings.
Reactivity is a threat to ecological validity. Reactivity is defined as an alteration
in performance that occurs as a result of being aware of participating in a study. In
other words, reactivity occurs sometimes because research study participants
might change their performance because they know they are being observed.
Reactivity is a problem of ecological validity because the results might only
generalize to other people who are also being observed.
A good metaphor for reactivity comes from television. Once you know that the
camera is turned on to YOU, you might shift into your television behavior. This
can also happen in research studies with human participants who know that they
are being observed.
Another threat to ecological validity (not mentioned in the chapter) is called
experimenter effects. This threat occurs when participants alter their performance
because of some unintentional behavior or characteristics of the researcher.
Researchers should be aware of this problem and do their best to prevent it from
happening.
Temporal Validity
Temporal validity is the extent to which the study results can be generalized across time.
For example, assume you find that a certain discipline technique works well with
many different kinds of children and in many different settings. After many years,
you might note that it is not working any more; You will need to conduct
additional research to make sure that the technique is robust over time, and if not
to figure out why and to find out what works better. Likewise, findings from far in
the past often need to be replicated to make sure that they still work.
Treatment Variation Validity
Treatment variation validity is the degree to which one can generalize the results of the
study across variations of the treatment.
For example, if the treatment is varied a little, will the results be similar?
One reason this is important is because when an intervention is administered by
practitioners in the field, it is unlikely that the intervention will be administered
exactly as it was by the original researchers.
This is, by the way, one reason that interventions that have been shown to work
end up failing when they are broadly applied in the field.
Outcome Validity
Outcome validity is the degree to which one can generalize the results of a study across
different but related dependent variables.
For example, if a study shows a positive effect on self-esteem, will it also show a
positive effect on the related construct of self-efficacy?
Why do you think Rosenberg used 10 items to represent self-esteem? The reason
is because it would be very hard to tap into this construct with a single item.
Rosenberg used what is called multiple operationalism (i.e., the use of several
measures to represent a construct).
Think about it like this: Would you want to use a single item to measure
intelligence (e.g., how do you spell the word "restaurant")? No! You might even
decide to use more than one test of intelligence to tap into the different
dimensions of intelligence.
Whenever you read a research report, be sure to check out how they represent
their constructs. Then you can evaluate the quality of their representations or
"operationalizations."
Internal validity
Internal validity is the same as it was for quantitative research. It is the degree to which a
researcher is justified in concluding that an observed relationship is causal. It also refers
to whether you can conclude that one event caused another event. The issue of causal
validity is important if the qualitative researcher is interested in making any tentative
statements about cause and effect.
I have listed three strategies to use if you are interested in cause and effect in
qualitative research.
The first strategy is called researcher-as-detective (carefully thinking about cause
and effect and examining each possible "clue" and then drawing a conclusion).
The second is called methods triangulation (using multiple methods, such as
interviews, questionnaires, and observations in investigating an issue)
The third strategy is called data triangulation (using multiple data sources, such as
interviews with different types of people or using observations in different
settings). You do not want to limit yourself to a single data source.
External validity
External validity is pretty much the same as it was for quantitative research. That is, it is
still the degree to which you can generalize your results to other people, settings, and
times.
Note that generalizing has traditionally not a priority of qualitative researchers.
However, in many research areas today, it is becoming an important goal.
One form of generalizing in qualitative research is called naturalistic
generalization (generalizing based on similarity).
When you make a naturalistic generalization, you look at your students or clients
and generalize to the degree that they are similar to the students or clients in the
qualitative research study you are reading. In other words, the reader of the report
is making the generalizations rather than the researchers who produced the report.
Qualitative researchers should provide the details necessary so that readers will be
in the position to make naturalistic generalizations.
Another way to generalize qualitative research findings is through replication.
This is where you are able to generalize when a research result has been shown
with different sets of people, at different times, and in different settings.
Yet another style of generalizing is theoretical generalizations (generalizing the
theory that is based on a qualitative study, such as a grounded theory research
study. Even if the particulars do not generalize, the main ideas and the process
observed might generalize.
Here is a summary of the strategies used in qualitative research. (Note: they are also used
in mixed research and can be used creatively even in quantitative research.)
The bottom line of this chapter is this: You should always try to evaluate the research
validity of empirical studies before trusting their conclusions. And, if you are conducting
research you must use validity strategies if your research is going to be trustworthy and
defensible.