Understanding Language Tests and Testing Practices
Understanding Language Tests and Testing Practices
Understanding Language
Tests and Testing Practices
All ESOL teachers come into contact with language assessment in one form
or another, yet many find principles of assessment an aspect of professional
knowledge that is difficult to update and apply effectively. This chapter,
therefore, lays out the critical concepts required for an understanding of the
language tests and testing practices outlined in subsequent chapters.
We begin with a brief discussion of the importance of assessment in the
broader context of TESOL and education followed by an explanation of the
problem that has developed because of the intellectual division between
those concerned with language testing and those who teach. We argue that
changes in educational practice, measurement theory, and language testing
research necessitate bridging that division to meet current and future
classroom needs. This chapter begins to build the bridge by deconstructing
the dichotomy that narrows the knowledge and responsibility of the
language teacher, and by replacing it with a more robust set of concepts for
understanding a range of testing practices: test purpose, test method, and
test justification. These overarching concepts form the basis for introducing
key terms often used to describe language tests.
1
ESOL Tests and Testing: A Resource for Teachers and Administrators
2
Understanding Language Tests and Testing Practices
3
ESOL Tests and Testing: A Resource for Teachers and Administrators
for some purposes but not for others. But the more systemic result of the
division is the partial and fragile knowledge that teachers have about how to
collect information systematically for the purpose of making certain deter-
minations about learners, which has led to the perception that testing and
assessment are completely distinct educational processes. Ironically, this
perception further dichotomizes the classroom assessment that teachers
engage in and the testing that is the responsibility of researchers. Despite
the value of a strong theory and practice of classroom-based assessment,
maintaining the separation between testing and assessment keeps teachers
from applying their knowledge of assessment to high-stakes testing.
Educational Practices
The educational landscape in the United States changed dramatically in the
early 1980s. The U.S. educational reform movement was precipitated by the
publication of A Nation at Risk: The Imperative for Educational Reform
(National Commission on Excellence in Education, 1983). This 32-page
report to the U.S. secretary of education documented a decline in the
academic quality of U.S. educational institutions (public and private, from
kindergarten through university), and it recommended five major reforms to
correct the declines in achievement. Two recommendations with implica-
tions for testing and assessment were (a) to restore an academic core (called
new basics) to the curriculum that should reflect a decidedly applied
orientation and (b) to implement more rigorous and measurable standards
for academic performance. Instructional practices and assessment proce-
dures were modified to conform to the curriculum reforms that followed the
release of the report (Linn, 1994). The reform movement gained momentum
in the 1990s, when the federal government passed the Goals 2000: Educate
America Act (1994), which created a structure and empowered a body to
develop guidelines for national education standards and offer states exem-
plary standards and the assessments to use in achieving these new national
standards. Barton and Coley (1994) underscored the shift in assessment as a
result of these reforms.
4
Understanding Language Tests and Testing Practices
5
ESOL Tests and Testing: A Resource for Teachers and Administrators
6
Understanding Language Tests and Testing Practices
Q: Have the experts found this test to Q: Can appropriate types and levels of
be reliable? consistency be shown for this test
in my setting?
Q: Have the experts found strong Q: Can I demonstrate that this test
correlations between this test and performs as I would expect and
other measures? want it to in my situation?
Q: Have the experts found that the test Q: How can I show that the test is
has one or more of the three valid for my use?
validities?
typically seen as one type of validity evidence. In the past, validity was
largely established through correlations of a test with other tests, but now
validity is best argued on the basis of a number of types of rationales and
evidence, including the consequences of testing (e.g., its effect on teaching).
Construct validity was seen as one of three types of validity: content,
criterion related, and construct. But today, validity is a unitary concept with
construct validity as central; content and criterion-related evidence can be
used as evidence about construct validity.
These changes are interesting and important for language test users
particularly because of three themes that underlie them. First, the changes
have resulted in a view of validity as a context-specific argument rather than
a test characteristic that can be established in a universal way. As a conse-
quence, a second theme is the view that justifying the validity of test use is
the responsibility of all test users rather than a job solely within the purview
of testing researchers who develop large-scale, high-stakes tests. A third
theme is that one consideration of ESOL test users should be the effects of
tests on the teaching and learning of English.
7
ESOL Tests and Testing: A Resource for Teachers and Administrators
8
Understanding Language Tests and Testing Practices
choose tests to see that they are used fairly. This discussion builds on the
work of Canale (1987), who in the 1980s was an advocate for appreciating
that test specialists and practitioners alike have a responsibility to “ensure
that language tests are valuable experiences and yield positive consequences
for all involved” (Douglas & Chapelle, 1993, p. 3). One of the ongoing
issues of the postmodern period is to gain a greater understanding of how
test fairness should come into play in the testing process (Kunnan, 1997).
Clearly, recent developments in educational practices, measurement
theory, and language testing research offer compelling reasons for ESOL
professionals to be assessment literate, which means being able to choose and
use assessments for all of their purposes (Stiggins, 1997). At one time, the
roles of language teachers and testing specialists were highly differentiated,
leading many ESOL teachers and program administrators to become
increasingly disconnected from the technical developments and practices
associated with language tests and different types of assessments. However,
the postmodern period is placing more responsibility for selecting, devel-
oping, and justifying assessments in the hands of practitioners, many of
whom lack sufficient assessment literacy and confidence to fulfill these
responsibilities.
9
ESOL Tests and Testing: A Resource for Teachers and Administrators
Method Justification
10
Understanding Language Tests and Testing Practices
Figure 1.2. Types of Inferences That Can Be Drawn From Language Tests
Inference
Specific ➤ ➤ General
Ability Ability Specific- Specific- General-
connected connected purpose purpose purpose
to specific to class or language language language
material program ability ability ability
taught taught in
class
11
ESOL Tests and Testing: A Resource for Teachers and Administrators
12
Understanding Language Tests and Testing Practices
13
ESOL Tests and Testing: A Resource for Teachers and Administrators
Examinees’ Responses
Messick (1994) cautioned against making a dichotomous distinction
between multiple-choice items and open-ended performance tasks and argued
that they represent “different degrees of response structure” (p. 15). Simi-
larly, the dichotomy selected versus constructed (see Figure 1.6) is too clear-
cut a distinction to describe response types meaningfully. Messick submitted
that multiple-choice assessments constitute one end of a continuum whereas
“student-constructed products or presentations” (p. 15) form the other.
14
Understanding Language Tests and Testing Practices
Difficulty of TOEFL — — —
items correct Computer-
Based Test:
Grammar (SS)
15
ESOL Tests and Testing: A Resource for Teachers and Administrators
Validation criteria for evaluating language The specific practices of language test
tests should be based on work in evaluation are best guided by theory and
applied linguistics. research in language testing.
Construct validity is central in test The construct that the test is intended to
evaluation. measure must be clearly defined, and
other evaluative issues about a test
should be secondary.
16
Understanding Language Tests and Testing Practices
Essential Vocabulary
A full consideration of language tests and testing requires a working
knowledge of the terminology used to express and develop knowledge of
this domain. In this section we define the vocabulary used to discuss and
evaluate language tests and test results. Of course, these terms are also
associated with the three essential components of testing described above:
test purpose, test method, and test justification.
17
ESOL Tests and Testing: A Resource for Teachers and Administrators
18
Understanding Language Tests and Testing Practices
19
ESOL Tests and Testing: A Resource for Teachers and Administrators
the degree to which the test tasks, including the language, resemble those
that examinees will encounter beyond the test setting. Authenticity is
typically argued to be a desirable quality for a test, as we explain in chapter
4. Two other terms that have become widely used in describing language
test methods are discrete (point) and integrative. Discrete refers to test tasks
that aim to measure a single aspect of language knowledge, whereas
integrative refers to those that require examinees to call on multiple aspects
of language knowledge simultaneously. Authenticity is not necessarily
connected to discrete or integrative test methods.
20
Understanding Language Tests and Testing Practices
21
ESOL Tests and Testing: A Resource for Teachers and Administrators
22
Understanding Language Tests and Testing Practices
Conclusion
This chapter has laid the groundwork for examining current ESOL tests and
assessments. We have reviewed the historical division that exists between
language teachers and language testers as well as the changes over the past
20 years that make such a division untenable for both groups in the
postmodern period of language testing. In order to reconceptualize testing
and assessment in a more productive way, we have rejected the distinction
between the two and introduced concepts and terms for understanding the
notions within the domains of test purpose, test method, and test justification.
23