Attachment
Attachment
PANDORA'S BOX:
EVALUATING REMEDIAL/
DEVELOPMENTAL EDUCATION
W. Norton Grubb
David Gardner Chair in Higher Education
School of Education
University of California, Berkeley
February 2001
Supported by the Alfred T. Sloan Foundation
For additional copies please contact-
Community College Research Center
Teachers College, Columbia University
439 Thorndike Building
525 W. 120th Street, Box 174
New York, New York 10027
212-678-3091 (telephone)
212-678-3699 (fax)
Abstract
Introduction.............................................................................................................2
Endnotes.................................................................................................................42
References..............................................................................................................46
1
INTRODUCTION
for example, multiple-choice tests covering the facts and skills of a body of
content. Students are responsible for learning this content, and too bad if they fail.
in any event not the responsibility of the institution or its instructors. Like the
caricature of the college with a dropout rate of two thirds ("Look to your right;
look to your left . . ."), dropout rates are expected to be high — and in many two-
But over time this conception of student responsibility has been modified
by one that places greater responsibility for success on the institution itself.
Accepting such responsibility means identifying and then correcting the many
otherwise drop out for financial reasons; and providing child care, transportation,
two- and four-year colleges are testimony to a shift toward a greater institutional
2
complacent about high rates of non-completion, particularly since open access (in
income, minority, and immigrant students — whose high dropout rates are both
At the same time, colleges remain wedded to older ideals about their
purposes (e.g., Eaton, 1994), certainly critics deploring the dreadful state of
higher education (e.g., Traub, 1994; McGrath and Spear, 1991), and certainly
policymakers wanting to reduce funding for remediation. These groups find the
taking greater responsibility for their students' success but often ambivalent about
their efforts to do so. Most postsecondary institutions now provide some form of
remained marginal in most institutions, squeezed into the back pages of college
3
But remedial/developmental education is real education, of the most
difficult sort. Under the best circumstances, it tries to do more than simply filter
accomplished students from the others; it tries to educate all of them, including
those who seem not to have learned much in ten or twelve years of conventional
pressed into service, nor individuals untrained in its special challenges.ii The task
to achieving other goals including increased learning (of basic reading, writing, or
education freighted with multiple purposes (Labaree, 1997) — its goals are
relatively clear. But there have been relatively few evaluations of remedial
programs, and many existing evaluations are quite useless because, as I point out
in Section II, they fail to recognize what the program does — and therefore they
III a variety of evaluation approaches that can improve information about many
different aspects of remediation, including not only its effects but also the
4
instructional methods used, the progress of students, and the ways students find
because it is designed to open up the black box, to reveal the problems with
most central issues in American schooling, particularly those about equity and
with very different levels of preparation, and teach them all? Or are our schools
and colleges merely sorting mechanisms, sifting the competent from the
assignments, promoting one group and "cooling out" the other? The answers to
these questions may be partly ideological, but they are empirical as well — and
I. MULTIPLE APPROACHES TO
REMEDIAL/DEVELOPMENTAL EDUCATION
For all the debate over remedial education, there is almost no discussion
educative in any sense of the word, whether it stands any chance of bringing
activities that march under the banner of remedial education. Otherwise it is easy
5
to fall into the trap of assuming that developmental education is well defined, and
their peers — who are often literally down the street — are doing, even though
their instructional tasks are similar. In community colleges, what is often termed
time to time some efforts at state-level reform.v In four-year colleges, the same
takes Shaughnessy's (1997) Errors and Expectations as its starting point; she
argued that the errors in the writing of basic writers follow patterns rather than
simply being random mistakes, and that a skilled instructor could use these errors
Since then the field of compositional studies has further elaborated various
educators in two- and four-year colleges have virtually no contact with one
another; even though there are journals and associations to which the two groups
6
Adult education is a vast chaotic world of programs funded with state and
federal money, as well as a bewildering array of local funds (for example, for
characterize what happens in this sphere since there is so much variation. Because
based programs are more constructivist and student-centered in their teaching than
the rest of adult educationvi — but these cells of innovation often do not know
what the rest of adult education is like or how different they are, and there is not
(adult basic education distinct from adult secondary education) prepare students
to pass the GED, a credential of dubious valuevii that, because of its conventional
questions (Grubb and Kalman, 1994). In an era of insistent concern with the
higher-order skills for a flexible labor force, the complex area of literacy and
Finally, the equally vast and complex area of job training programs —
7
sometimes as a prerequisite to vocational skills training. It is impossible to know
how much of this goes on, because providers are not required to distinguish
remedial education from other services and because remedial education is usually
a local option; but there is general agreement that programs are forced to provide
more remediation than they would like. Often, these programs subcontract with
programmed workbooks for such clients (Grubb and Kalman, 1994). Sometimes
job training programs create remedial labs with computer-based programs — but
these programs are invariably just skills and drills conveyed to the computer, with
even shorter reading passages because of the small size of the screen, a rigid
progression through topics, and a lab "manager" (rather than a teacher) whose
jobs is to turn the machines on and off and monitor progress but who is not
trained to provide any instruction. Finally, some adult education and job training
programs have become enamored with functional context literacy training (Sticht
et al., 1987), an approach that uses the materials from a "functional context" (like
employment or the military) to teach multiple literacy skills (e.g., 186 "reading to
do" skills and 143 "reading to learn" skills, in Sticht, 1979). While the functional
lends itself to the most didactic and skills-oriented teaching. Functional context
with different histories, different students, and different goals. Certain pedagogies
8
more often, text-centered) approach I often call skills and drills. But because
programs are not in communication with one another, examples of innovation and
good practice cannot readily spread from one area to another, so the prospects for
reform often look gloomy — particularly in the spheres of adult education and job
training.
But there is substantial variation within each of these areas as well, and I
will illustrate this with examples from community colleges.viii These institutions
have certain advantages over the others that provide remedial education. As open
work, and so (unlike four-year colleges) the necessity for remedial education is
built into their basic structure. Community colleges also pride themselves on
being "teaching colleges"; even though this ideal is "honored more in the breech .
. . the tradition is there and can be called upon when warranted," as one English
instructor described it. Unlike adult education and job training, with their reliance
community colleges generally have master's degrees and are hired through
painstaking procedures (even though these usually have little to do with the
quality of teaching). Although these colleges have come to rely too much on part-
within community colleges is the approach I have labeled “skills and drills”
(Grubb and Associates, 1999, Ch. 5). This tends to focus on sub-skills — on
9
arithmetic procedures like multiplication and percentages, on grammar and
punctuation and vocabulary, on math "problems" of the most contrived sort and
reading passages from texts that have been simplified for low reading levels
instructors will bring in reading from outside the class — from newspapers, for
example — but in a typical heterogeneous remedial class there are few common
know one another, because of the common pattern of taking courses almost
randomly, and therefore do not serve as resources for one another; mastering
rather than a collective and social activity (Worthen, 1997). This approach to
students are referred if they score below a certain score on a basic skills test; it
also emerges in covert or hidden remediation, which takes place in some "college-
that are converted into remedial classes because the majority of students are not
ready for what the instructor considers "college level work" (Grubb and
Associates, 1999, Ch. 5). Thus the amount of remediation in most community
colleges almost surely exceeds the count of official remedial courses, and is
instruction. Invariably, the programs are simply drills transferred to the screen,
10
with short reading passages followed by questions of fact, fill-in-the-blank
word problems.ix They typically allow students to move to the next level only
when they have passed a short "test" on one subject, so they manage the student's
progress carefully — and often create records for the instructor. Often, students
instructor or "manager," but this individual typically has neither the time nor the
training for instruction: if a student gets stuck, he or she has to go back in the
computer program to try to work out the problem, but there is no teaching in the
conventional sense of the term. Some of these programs are quite elaborate,
covering many different topics, and some are quite expensive; they are often
Unfortunately, the majority of these are simply repackaging and peddling the
The problem with this approach is not just that these classes are deadly,
with low levels of student engagement. They also violate all the maxims for good
teaching in adult education (Grubb and Kalman, 1994). And their tactic is simply
"more of the same": they take students who have not learned well in ten or twelve
years of standard didactic instruction, and then put them through an additional 15
weeks of similar instruction. There may be some success stories,x but overall there
is little chance that this dominant pedagogy can be very effective. It is foolish to
think that students who have never learned to read for meaning, or who have no
real understanding of numerals, can suddenly learn quickly from another round of
11
In our observations, substantial numbers of community college instructors
develop approaches to teaching — largely through trial and error — that are more
Ch. 5). They are quite aware that community college students have suffered a
of poor teaching; they are likely to blame urban school districts for the low levels
of their students. These instructors are more likely to bring in reading materials
from work, or from newspapers and political debates; they tend to spend
considerable time probing the interests of their students and their purposes for
attending college, so they can mold reading and writing to these interests. Here is
12
begin to teach what they are interested in, what they need, what makes
sense to them.
These instructors also foster work in groups rather than individual drill, partly in
conventional grammar and punctuation drills, but they stress above all having
students learn to create meaning from and with texts, and they subordinate drills
to that kind of reading. Here, for example, is one instructor talking about her
The topics these classes cover are idiosyncratic, because a student-centered class
of students. This creates problems for evaluation, because the outcomes are not
does mean that such classes are livelier than skills-oriented remediation. Students
are much more engaged, with each other as well as the instructor; the activities
and materials of the class are generally adult, rather than the childish drills of the
behaviorist classroom; and there seems much greater chance that this approach
13
can finally teach students about the complexities of language and mathematical
thinking.
teaching seems random and idiosyncratic, because the odyssey from didactic to
constructivist teaching is usually one that instructors make on their own, through
trial and error, with at best a little help from their friends. In most community
colleges, there are few institutional resources to help instructors make this
division complained that existing basal readers generally follow either a "phonics"
polarized: "We're back to the same old thing — top-down or bottom-up; and that's
prewriting, rewriting, and revision, and includes "strategies for invention and
discovery whereby instructors help students to generate content and purpose." The
approach includes grammar, spelling, and other mechanics, but only in the final
stages of writing since its use early in writing has been so counter-productive.
14
research and practice. The rest of the volumes contain examples, applications, and
syllabuses in great profusion. The purpose of the "basic writing curriculum book"
was partly to prepare new and part-time instructors to the division's methods. The
them with their initial stages of teaching. In this division — and in a small number
when students take two or more classes jointly; then, if instructors spend
sufficient time planning together, each course can support and complement the
others. LCs are infinitely flexible and can be used for a variety of purposes
central academic course — is matched with an English and/or a math course. For
example, one institution found that a particular biology course was blocking the
their problems in reading and math. The biology course was then joined with
supportive math and English courses, which in turn modified their content to
15
provide the kinds of academic competencies necessary in biology. An automotive
instructor who discovered the problems his students had with reading devised a
The benefits of LCs are multiple, at least when they work well. Most
obviously, students find themselves making progress in subjects that they care
contextualize their teaching: examples and applications in English and math can
come from the "lead" course, and the lead instructor can develop writing exercises
and problem sets that are used in the other courses. Students within learning
communities get to know one another much better than most community college
students do, and they universally report forming study groups as a result.xiii And
instructors report benefits too, since learning communities break down the
teachers. With all these apparent benefits, the evidence so far indicates that
students in learning communities tend to persist longer and earn higher grades
16
There are a number of other approaches to improving developmental
Weinstein and her colleagues (Weinstein et al., 1998) have suggested strategic
are sometimes used as the basis for staff development for all faculty in a
community college.
communities are vastly different from the conventional tendency to fragment the
have come to college with many years of formal schooling but without adequate
command of language and math; others promise new approaches that might
correct these problems. Even within one institution like the community college,
17
move out into the world of employment or onto further education? The evidence
is sparse, and partly it is for lack of trying: most states and most colleges that
provide remediation have not yet started to evaluate their programs in any way.xv
decisions to expand or abolish programs might be based on such evidence: if, for
example, there were many years of evidence that no form of remediation benefits
students, then we might be tempted to eliminate all funding for such programs.
But given the educational imperatives behind remediation and the lack of
evidence, very few educators would be willing to take such a step — so I presume
the purpose of evaluation is both to learn more about the conditions of success
developmental courses. However, such an approach fails to see whether there are
a good thing, it cannot by itself help people find jobs or help them vote, or
comparison of pre-tests and post-tests, usually on some test of basic skills like the
18
TABE or the CASAS. For example, the Learning Assessment and Retention
volumes of such figures (e.g., LARC, 1989a and b); they enable one to determine,
for example, that increases in reading scores were higher in one college than
another. But such results are almost useless, for a number of reasons. Most
obviously, pre-test/post-test comparisons are available only for students who have
stayed with a course until the end, when the post-test is given; if weaker students
drop out, or if only "brush-up" students survive until the end, then the test
increases will badly overstate the results for the average or random student. In
might be possible to draw some conclusions, either if gains were close to zero or
if they seem high, by some unarticulated and possibly idiosyncratic standard; but
even that is risky. The meaning of such results improves a little when they include
the proportion of students who complete the course. For example, recent figures
from the CUNY system reveal that in 1990, 64.7 percent of post-tested students
gained at least one year on the TABE; since only 61.7 percent of students were
post-tested, this means that — under certain shaky assumptions — perhaps as few
year.xvi
19
But other problems with pre-test/post-test results cannot be resolved
simply by collecting more data. From some perspectives, the tests themselves are
objectionable: they ask for the kinds of responses, about grammar and vocabulary,
arithmetic operations and simple word problems, that fuel skills and drills
communities, would accept the results of such tests; the vitriolic debate between
McKenna, Miller, and Robinson (1990) and Edelsky (1990) about the evaluation
scores may be better than the opposite, but it may still not lead to further progress,
the completion of meaningful degrees, or other outcomes. Finally, and worst of all
from the perspective of improving programs, these statistical results say nothing
about why test scores are what they are: they provide literally no information
about why students fail to complete courses in such large numbers, about whether
some approaches are better than others, about whether stand-alone remedial
learning communities). Such figures, if they are discouraging enough, may lend
some urgency to the problem of reform, but they do not provide any guide about
what to do next.
evaluation designs, but many of the problems with such results are repeated in the
20
most sophisticated approach to evaluation — exemplified by the random-
assignment evaluation of the remedial program in the San Diego GAIN program
a control group of welfare recipients was not assigned to such a program, though
8.4 percent of them participated in adult basic education or some kind of GED
enrolled in GAIN). Participating in the GAIN program increased the rate at which
individuals received the GED — since 9.1 percent of the GAIN group but only
2.0 percent of the control group received a GED. However, the GAIN group as a
whole did not improve their scores on the Test of Adult Literacy Skills
(TALS),xvii though individuals who scored the highest on an initial screening test
from 454 among the experimental group to 488 among the GAIN group, on a
scale ranging up to 1,000. There are a few clues about the distinctiveness of the
San Diego program: it was designed specifically for the GAIN program, "built on
the premise that existing adult education services were not appropriate for the
21
classroom instruction, integrated academic and life-skills instruction," and "a new
teaching staff." The results are not particularly encouraging even for San Diego,
then, because even there the increase in TALS is trivial.xviii But the TALS is not a
programs, there is not much guidance about how to reform the remedial programs
that were part of GAIN — and there is no information whatsoever on the nature
of the remedial programs that caused even more dismal results in other counties.
others. For a number of years Miami-Dade Community College has evaluated its
remedial programs in the format of Table 1: completion rates are calculated for
students who are judged "below standard" in one, two, or three subjects, and who
have successfully completed all appropriate remedial courses versus those who
have not.xx (Table 2 provides some earlier results, in a slightly more detailed
example, that of the 6,324 students who entered, 59 percent were judged to need
some kind of remedial education. Students with three deficiencies had a much
harder time than students with one deficiency: only 42 percent of the former
group corrected all three deficiencies and only 9 percent of these students
graduated within three years, while 63 percent of students with one deficiency
22
corrected it and 28 percent of these graduated. (There are no surprises here, but
there is a substantial warning to high school students who think they can easily
make up during college the learning they have failed to do in high school.xxi) And
even students who take the full complement of remedial courses they presumably
need, graduated at much lower rates than those who entered needing no
outcomes for those students still enrolled after five years: some of these will
reasonable conclusion from these results is that remedial courses help a great deal,
but they cannot eliminate the gap between students with and without some need
groups; they neglect maturation effects, test effects, regression to the mean (which
doubt these results overstate the effects of remedial courses, and more
sophisticated statistical analysis could improve the results. But they are a vast
improvement over some of the evaluation results already presented: the outcome
measure is one of intrinsic value,xxii they clarify that the amount of remedial
23
education completed matters a great deal, and they compare many different
groups of students with varying needs for remediation. However, they still fail to
investigate what remediation is: they provide no clue about why so many students
fail to complete remedial courses, they have not examined what about them
attracts and repels students, and they do not investigate what these courses are like
and whether some of them are more effective than others. Interesting as they are,
24
Table 1
Below
standard Completed all Did not complete
in: remedial courses all remedial courses
__________________________________________________________________
"Still enrolled" refers to those still enrolled with a GPA of 2.00 or better.
Source: Morris (1994), Table 6.
25
Table 2
"Still enrolled" refers to those still enrolled with a GPA of 2.00 or better.
Source: Losak and Morris (1985), Table 1.
26
I know of no outcome evaluations that try to compare different approaches
(endnote 15). And a large testimonial literature indicates that students like
evaluation, perhaps because the dominant focus has been on measuring the effects
of a program on those enrolled in it. But what I will call, for want of a better term,
remediation based on an assessment of some sort — usually a basic skills test like
the TABE or the CASAS, sometimes with a writing sample, sometimes with some
fare better than those who did not. It is possible that some students assigned to
remediation do not need it — for example, students only needing brush up — and
students who did not "fail" the assignment test — who scored just above the cut-
off point, and did not enroll in remedial programs. It is quite possible that these
students would still benefit from remediation, or that some students score well on
basic skills tests but still cannot write or reason well — because these tests do not
measure such higher-level competencies. Any assignment procedure runs the risk
not need it — as well as allowing students who need some kind of remedial work
27
to progress to college-level work (Type II errors). Both Type I and Type II errors
may reduce rates of completing programs, and Type II errors have the added cost
converted into remedial courses. The assignment problem is a difficult one, and I
These issues about errors in assignment are at the heart of debates about
community colleges and whether they are egalitarian, advancing students who
they "cool out" students who otherwise would go further in four-year colleges.xxiii
implemented matriculation poorly, using tests that had not been validated for the
irrelevant tests, and using test results by themselves whereas the education code
stated that they were to be advisory only. To improve the quality of matriculation,
MALDEF claimed that there were too many Type I errors, of Mexican-American
28
regulations requiring that any prerequisites for any courses be justified through a
validation study, a burdensome procedure that has all but eliminated prerequisites
in California community colleges. Now there may be fewer Type I errors — and
if there are, they are presumably the choice of students rather than the result of
college assignment — but there may be more Type II errors, of students who need
Indeed, the lawyer involved in the case admitted that the remedy was imprecise
matriculation process, but instead got a series of crude rulings and bureaucratic
procedures, as is typical in legal cases.xxiv But the result of the lawsuit is unclear:
hurt them, or whether it benefited or harmed other students who have been
equally affected by the ruling and the response of the Chancellor's Office. Indeed,
no one has even thought to ask the question, never mind answer it.
infancy, and no one knows much about what works and what does not.xxv What
little evidence there is indicates that completion rates in remedial courses are low,
that the amount of remediation does matter to important outcomes like persistence
probably more effective than stand-alone classes. There is some suggestion (from
the GAIN evaluation) that "school-like" programs are less effective, but it is not
entirely clear what this means. The observational evidence (such as that in Section
29
learning — might not be widely accepted and might not even be correct; there is
particularly innovative and yet they have substantial effects on graduation and
retention (Table 1). But, aside from the possible recommendation to teach all
possible — is no more useful than the most basic pre-test/post-test designs, as the
GAIN results illustrate. There are too many dimensions of remedial education that
are poorly understood, and investigating them requires several different methods.
One is the program level, where information about a particular course and a
specific instructor would be useful in diagnosing what is going well and badly.
and subsequent progress through the college, but it should also include peer
30
useful to instructors themselves, they cannot be particularly complex — they
surely cannot use comparison groups, or follow students over long periods of
A second level includes more formal evaluations carried out at the institution
— or at the state or national level. These can be more complex, with control or
comparison groups, and can follow students over longer periods of time; their
purpose is not to improve the practice of specific instructors, but rather to assess
institutional and state policy, the overall effects of remediation, and the
approaches, each of which has the potential for illuminating a different aspect of
While it is plausible that dreary teaching is the reason, the difficult lives of many
combinations of reasons are responsible, and that even students themselves cannot
articulate why they stay with or leave a particular program. As one student
It was not even a decision. I just did not go. Sometimes you decide on
certain things. It was not a decision at all. Just like you go home, tired
31
from work, you don't decide about "Oh, I'm just going to go to sleep now."
You just doze off and go to sleep. It wasn't a plan. That's the way
[dropping] the class was: it wasn't a plan.
2. Outcome measures need to include more than test scores of basic skills.
(Indeed, it is an open question whether such test scores mean anything at all.)
like writing portfolios, and other measures emerge from a college's intentions for
the most valuable outcomes in some cases. More systematic collection of outcome
remedial courses can achieve. Some students may declare other purposes —
studies can clarify these goals and the contribution of courses toward achieving
them.
courses produces benefits compared to students who do not take such courses.xxx
But while it might be possible to design a random-assignment study under certain
32
conditions — for example, comparing the effects of learning communities to
Table 1 — where some students are thought to need remediation, but do not take
Evaluation studies can still collect other information like prior test scores and
evaluating — and this means observing and describing the classroom practices in
remedial courses. The conventional "black box" evaluation, in which the nature of
"Pandora's box" approach that clarifies both the triumphs and the troubles of
and didactic practices on the one hand, and constructivist and student-centered
practices on the other, are dimensions of teaching that emerge over and over
teaching and learning. Although there is endless debate about what dimensions of
teaching are important, these two polar approaches can be operationalized for
33
However, there may be other ways to think about the power of different
approaches, particularly since student motivation arises for many complex reasons
occupational class where the academic material would be applied. Some teachers
rapport with students, that overcome the limits of drill, and some students —
particularly ESL students, who seem to be able to sit through anything, and older
students with clear and passionate goals — are able to learn from even the most
dreary teaching. Some of these possible successes may be replicable and others
may not, but understanding them better is a necessary first step to improving the
quality of instruction.
some students who might benefit do not attend them — either because the
students judged in need of remediation who did not enroll in such courses — like
some of the groups whose progress is measured in Table 1; (b) the "near misses,"
34
or those who barely passed the assignment test, compared both to those who
enrolled in remediation and those who clearly do not need remediation, at least
based on the initial basic skills test.xxxvi Finally, some consideration of alternative
empirical issues of this kind. The expansion of postsecondary education since the
provided opportunities for some students where none existed before, and the
35
APPENDIX:
some kind. In remedial education, the assessment is usually a basic skills exam,
though a few colleges add a holistically graded writing sample; the assessment
process could be a more complex procedure in which multiple tests (and more
education, and the like. Then a person is assigned to remedial education based on
event some students enroll in the remedial course and others do not. Of those who
enroll, some fraction complete and others (often a very large percent) do not.
Sometimes there is an exit exam to move to the next level of education (for
example, the first college-level course, or the next remedial course in a sequence),
college.
36
reading, writing, math, or any other subject can be measured on a 100-point scale,
ascertained by a conventional test, and that (arbitrarily, perhaps, but with the
weight of tradition) a certain point on this scale (say 70) is considered necessary
in remediation until they achieve 70 on an exit exam; there may be different levels
of remedial courses depending on the scores students attain (e.g., one for those
scoring 50-60, another for those scoring 60-70). The effectiveness of remedial
programs is simply measured by the absolute increase in the score for the average
Finally, there might be a different standard (say 85) for graduation from this
exam for graduation (as Florida has in its rising junior exam, or high schools now
graduating.
In this simple world, there are not any special problems once the test has
been created and the various cut-points established: the test used for initial
adequate, and the required exit exam is obvious. Indeed, the conventional pattern
this simple model. But of course facility in various subjects does not follow this
model at all because there are multiple dimensions to reading, writing, or math.
37
different from those in a standard English course; the dimensions tested in a
diagnostic basic skills test may be different from those required in any subsequent
college-level courses; and the content of any specific remedial course may be
different from both the diagnostic exam and subsequent courses. Thus the initial
assessment may not be aligned with the remedial program (which is a problem of
predictive validity); the remedial program may not be aligned with the exit exam,
if the exit exam is established independently of the program (though in this case
one might assume that instructors would start teaching to the test); and neither the
program nor the exit exam may be aligned with subsequent "college-level"
enhance subsequent outcomes. Of course, the college program as a whole may not
points at which the multiple dimensions of any particular subject can create
benefit (Type I errors). When a college has a series of remedial courses, the
time and (usually) money. The additional time required to complete a remedial
sequence may itself lead to non-completion. Distinguishing the time and money
By the same token, students may not enroll in remedial courses even if
they would benefit — the problem of Type II errors. This may happen (as in
38
California and some other states) because a placement test is advisory only,
because students dislike the additional time and money costs of remediation, or
because college-level courses stress competencies that are not measured by the
that subsequent college-level work requires facility with analytic thought and try
to teach that, while the initial basic skills assessment measures facility with
grammar and vocabulary, then students will pass the initial assessment but may
Now we can see a little more clearly where conventional evaluations fail
family support or family responsibilities, etc.) can identify the increase in the
need of them. Then one minus this probability is the probability of Type I error;
for example, using the simplified figures in Table 1 (instead of the logit or probit
subject benefited from completing the appropriate remedial course, in the sense
that they graduated rather than not graduating, but 45 percent did not.xxxvii Using
the figures on those who graduated or were still enrolled, the results are much
better: 83 percent benefited and only 17 percent did not. However, the likelihood
39
of finally graduating surely decreases as the period of time enrolled increases, and
function of time, where the time necessary to graduate is increased (and the
the nature of alignment or misalignment among the different aspects of the overall
exam, the various remedial courses, exit exams (if any) and the courses students
complain about misalignment (Grubb and Associates, 1999, Ch. 5), careful
the lack of a particular kind of information. When students are judged in need of
can be determined by comparing the two groups (as in Tables 1 and 2). But if
students are judged not in need of remediation, then none of these students enroll
40
in remedial courses — so it is impossible to tell if some of them would have
group: if we assume that students within one standard deviation of the critical cut
score are statistically indistinguishable from one another, then analysis of this
restricted group and the effects of remediation might provide an estimate of Type
complaining that students are not prepared for the specific uses of reading,
with the basic skills necessary to particular occupational areas. But in the absence
or to some of both.
Many developmental instructors are quite aware of many of these problems, and
Associates, 1999, Ch. 5, especially section 1). But most writing about
whatsoever. Until these issues can be more carefully examined and understood,
evaluation be incomplete.
41
ENDNOTES
i
I will not repeat the various debates over the terms "remedial", "developmental", and "basic
skills" education; see Goto (1995) for an excellent review of these issues and the compromise of
"remedial/developmental" education.
ii
A full appreciation of the difficulty of remedial/developmental teaching can be found in Goto
(1998), who followed a number of students in two community college classrooms. Such
understanding can come only from examining the lives of students as well as activities within the
classroom and pedagogical strategies.
iii
In reality, most instructors and institutions interpret remediation in instrumental terms, but
students may not. A more student-centered conception could accept education that looks to be
relatively basic to be valuable in its own right, for students whose purposes may not include
completing advanced degrees. See Goto's (1998) description of basic writing students in a
community college, some of whom view it as valuable regardless of its instrumental purposes.
This way of looking at remedial education is more common in certain branches of adult
education; see for example, Gowen and Bartlett's (1997) description of several women able to
confront domestic abuse through a remedial writing program.
iv
Traub (1994) includes some descriptions of one College Skills class at CCNY with many
comments from the instructor about the lack of academic preparation among his students. These
descriptions convey an example of a disequilibrium between instructor and students — where the
instructor has an expectation of what students should be able to do that is not matched by their
preparation. For other descriptions of remedial/developmental classes in community colleges, see
Grubb and Associates (1999), Ch. 5, and Goto (1998). But most examinations of remedial or
developmental education contain no analysis of classrooms whatsoever; see, for example,
Roueche and Roueche (1993) and McCabe and Day (1998).
v
See the National Center for Developmental Education at Appalachian State University in
Boone, NC, which publishes the Journal of Developmental Education and a newsletter, the
Review of Research in Developmental Education. The national association is the National
Association for Developmental Education.
vi
I based this statement on recent observations in several library and adult programs by Caleb
Paull.
vii
Cameron and Heckman (1993) found no employment value to the GED, using sophisticated
statistical techniques; reworking the same data, Murnane, Willett, and Boudett (1995) found a
small effect, though they noted that it might not be enough to overcome the pedagogical
disadvantages of the test. Nor does the GED appear to enhance subsequent education attainment;
see Quinn and Haberman (1986).
viii
This section draws heavily on a book about teaching in community colleges (Grubb and
Associates, 1999) based on observations of and interviews with about 280 instructors (including
27 English instructors and 36 remedial/developmental instructors) and about 60 administrators.
See also Worthen (1997), drawn from the same data. This is, amazing to say, almost the only
empirical work on teaching in community colleges since Richardson, Fisk, and Okun (1983).
42
ix
See Grubb and Kalman (1994), and the earlier review of computer programs for job training
programs by Weisberg (1988). While the latter review is by now several computer generations
old, we saw only drill-oriented computer programs in our observations during 1993 - 1997.
There are some interesting constructivist uses of computers by a few community college
instructors, but they are all individual efforts by instructors developing computer applications on
their own; see Grubb and Associates (1999), Ch. 7.
x
Community college and adult instructors sometimes tell stories of students, invariably older,
who breeze through a programmed text or workbook. I interpret some of these as "brush-up"
students, who have been out of school for a decade or more and have failed an initial placement
exam because they have forgotten the trivia involved in such tests. If they have learned basic
English and math in their earlier schooling, one additional exposure is sufficient to brush up on
these skills.
xi
Learning communities have generated a great deal of interest; see especially Matthews (1994a
and 1994b).
xii
Learning communities have also been used for ESL classes — for example, by pairing a
computer class with an ESL class concentrating on computer-related literature and vocabulary —
but I will not discuss these here.
xiii
Such study groups are reminiscent of those that are at the heart of Uri Treisman's approach to
teaching math, though they are more informal.
xiv
Gudan, Clack, Tang, and Dixon (1991); Tokina (1993); Tinto, Goodsell-Love, and Russo
(1994); MacGregor (1991).
xv
Boylan, Bliss, and Bonham (1997) found that only 14 percent of community colleges had any
systematic evaluation. A recent SHEEO survey found very few states able to comment on the
impact of remedial policies on student success, and most studies seem to be somewhere in the
planning stages (Russell, 1998, p. 26 and Appendix G). Most studies that purport to describe
effective programs rely on nominations of programs by various observers, not on outcome
measures; see, for example, the programs profiled in McCabe and Day (1998) or in Roueche and
Roueche (1993).
xvi
This is true if none of the students who failed to survive to the post-test gained at least a year.
Since some of these students may have benefited from their period in adult ed, one might say that
the results support gains among 40 percent to 64.7 percent of students — too broad a range to
have much confidence about the outcomes. Other results indicate that students with less than 21
contact hours gained an average of .76 years, while those with more than 120 hours gained .96
years — a difference that, whether it is statistically significant or not, strikes me as being trivial
in practical terms. See Student Outcomes Research Project (1996), Tables 2 and 4. This report,
to its great credit, spends a great deal of time clarifying the limitations of pre-test/post-test
comparisons and uses various other measures of success including interviews with teachers and
students.
xvii
The TALS, developed by the Educational Testing Service, includes document, quantitative,
and prose literacy components; only the first two were used.
xviii
The gain from 454 to 488 in San Diego is equivalent of a gain from 227 to 244 on each
component. ETS divides the TALS scores into five levels of proficiency: scores below 225 are
considered the most basic; then levels 2, 3, and 4 are distinguished by 50-point increases. An
43
average increase of 17 points is therefore only one-third of the variation within these levels, and
is unlikely to move an individual from one level to another.
xix
This finding is based on an article sent to me for refereeing, and therefore anonymous.
xx
Similar results are available for the CUNY system. For example, they indicate that 36.5
percent of Associate degree students who passed all the basic skills tests they needed had
graduated eight years later, compared to 14.9 percent of those who did not pass all of them and
33.9 percent of those who took no remediation (CUNY Office of Institutional Research 1998,
Table 12).
xxi
See also Rosenbaum (1998), who clarifies the much slower progress through postsecondary
education for students who have done poorly in high school.
xxii
The results in Table 2 are also available for CLAS test scores. The CLAS test is a "rising
junior" exam required of all students in Florida before they start their junior year; while it is only
a test score, it is one of great importance to students who want to transfer to four-year colleges.
xxiii
I have summarized these debates and the evidence in Grubb (1996), Ch. 2. See also Lavin and
Hyllegard (1996) and Rouse (1995). The empirical evidence on balance is against the hypothesis
of cooling out, since most community college students would not otherwise have gone to
postsecondary education at all. In addition, the critics of community colleges tend to rely on
ancient "evidence" about the role of counselors in Clark (1960), although approaches to
counseling have changed dramatically since the 1950s. If there is any truth to the charge of
"cooling out," my argument is that it occurs by accepting "non-traditional" students and then
teaching them in traditional ways; see Grubb and Associates (1999), especially Ch. 10.
xxiv
Oral communication, Susan Brown, Council for Latino Issues Forum, San Francisco,
December 1997.
xxv
There is, to be sure, a large advice literature about how best to teach adults, but this is based
largely on experience rather than empirical evidence of any sort. Similarly, the synthetic lists of
recommendations about good teaching, like the widely cited "seven principles for good practice"
(Chickering and Gamson, 1991), are based on a mixture of evidence from the K-12 literature,
experience, and student ratings.
xxvi
For a more extended argument about the creation of such communities of practice through
peer observation, see Grubb and Associates (1999).
xxvii
This is taken from interviews with community college students, in Grubb (1996), Ch. 2.
xxviii
The dominant approach to dropouts in higher education has followed Tinto's (1987) model,
which assumes that the extent of academic and social integration into a college explains dropping
out. But this model is much too restrictive for community colleges since it fails to include the
many external factors — fiscal reasons, complex lives, issues of identity and commitment — that
affect community college students. Therefore quantitative analysis should look for causes
beyond Tinto's model. Some of these, like the reasons for experimenting, are difficult to quantify
and are probably best examined through interviews.
xxix
See the battle between Kane and Rouse (1995a, 1995b) and me (Grubb, 1993, 1995) about
whether program completion is necessary for economic benefits to materialize. Other work with
the SIPP data (Grubb, 1997) and a survey of the available literature (Grubb, 1998) indicates that
the benefits of taking courses without completing credentials is on the average quite low and
quite variable.
44
xxx
For an example of a large study without any comparison group, the National Study of
Developmental Education surveyed results for developmental students only in a variety of
postsecondary institutions. The findings — e.g., that 24 percent of developmental students in
community colleges persisted until graduation — are impossible to interpret without knowing
more about persistence of other students at the same institutions. See Boylan and Bonham
(1992); other results from this study came out in subsequent issues of Research in
Developmental Education.
xxxi
Community college students often take courses according to the time of day they are taught,
to fit into their complex schedules. One possible design, therefore, would be to randomly assign
different pedagogical approaches to different times of day. Then students could either choose a
course or be assigned randomly according to the time of day they prefer; any particular class
would have a combination of self-selected and randomly assigned students, and these two groups
could be compared to see if the students who select a particular course are different on any
dimensions from randomly assigned students.
xxxii
In some states, placement in remedial courses is mandatory if students score below certain
levels on diagnostic tests; in other states remediation is voluntary.
xxxiii
For a similar recommendation in the context of job training programs see Friedlander,
Greenberg, and Robins (1997).
xxxiv
See, for example, the work of Knapp & Turnbull (1990) and Knapp & Associates (1995).
They defined "the conventional wisdom" and "alternatives to conventional practice" almost
precisely as I have defined skills and drills versus meaning-centered approaches. They then
compared the effects of classes with different numbers of practices drawn from the list of
"alternatives."
xxxv
One instructor we observed had devised a way of reinforcing material in four different ways,
and was highly conscious of using different materials — written materials, oral instruction, films,
computer/based materials, etc. — to fit different "learning styles." While the skills presented in
her class were quite basic, students seemed more engaged than in most remedial classes.
xxxvi
Where individuals are assigned to remediation based on a basic skills tests with continuous
results, information is available on the subsequent education experiences of those just above and
just below the cut-off score for assignment to remediation. Alternatively, analysis of completion
as a function of scores at entry plus remediation could reveal whether remediation is effective for
different groups of entering students. Most community colleges have these data in their files,
though they are often scattered in different data systems and research of this kind seems always
to be a low institutional priority.
xxxvii
From those below standard in one subject, 7 percent who did not complete the required
course graduated, while 28 percent who passed the required course graduated. If we assume that
remedial courses ought to bring students to the level of those who did not need remediation, then
45 percent should have graduated; therefore (28-7)/(45-7)=55 percent benefited. Other
assumptions of what remediation can hope to achieve obviously generate different conclusions.
45
REFERENCES
Boylan, H., Bliss, L., & Bonham, B. (1997). Program components and their relationship to
student success. Journal of Developmental Education, 20, 2-9.
Boylan, H., & Bonham, B. (1992). The impact of developmental education programs. Review of
Research in Developmental Education, 9.
Cameron, S., & Heckman, J. (1993, January). The Non-equivalence of High School Equivalents.
Journal of Labor Economics, 11.
Chickering, A.W., & Gamson, Z.F. (1991). Applying the seven principles for good practice in
undergraduate education. San Francisco, CA: Jossey-Bass Publishers.
Clark, B. (1960). The open door college: A case study. New York: McGraw-Hill.
Cross, K.P. & Angelo, T. (1993). Classroom assessment techniques: A handbook for college
teachers, (2nd ed.). Ann Arbor: National Center for the Improvement of Postsecondary
Teaching and Learning.
CUNY Office of Institute Research and Analysis (1998, February). Basic skills and ESL at the
City University of New York: An overview. New York: CUNY.
Friedlander, D., Greenberg, D., & Robins, P. (1997). Evaluating government training programs
for the economically disadvantaged. Journal of Economic Literature, 35,1809-1855.
Goto, S. (1998, October). The threshold: Basic writers and the open door college. Unpublished
doctoral dissertation. Berkeley School of Education, University of California.
Gowan, S. (1992). The politics of workplace literacy: A case study. New York: Teachers College
Press.
46
Gowen, S., & Bartlett, C. (1997). Friends in the kitchen: Lessons from survivors. In G. Hull
(Ed.), Changing Work, Changing Workers: Critical Perspectives on Language, Literacy,
and Skills. (pp. 36-51) Albany: State University of New York Press.
Grubb, W. N. (1993). The varied economic returns to postsecondary education: New evidence
from the class of 1972. Journal of Human Resources, 28, 365-382.
Grubb, W.N. (1995). Response to comment. Journal of Human Resources, 30, 222-228.
Grubb, W. N. (1996). Working in the middle: Strengthening education and training for the mid-
skilled labor force. San Francisco: Jossey-Bass.
Grubb, W.N. (1997). The returns to education and training in the sub-baccalaureate labor market,
1984-1990. Economics of Education Review, 16,231-246.
Grubb., W.N. (1998, August). Learning and earning in the middle: The economic benefits of
sub-baccalaureate education. Occasional Paper, Community College Research Center,
Teachers College, Columbia University.
Grubb, W.N., & Associates (1999). Honored but invisible: An inside look at teaching in
community colleges. New York and London: Routledge.
Grubb, W.N., & Kalman, J. (1994). Relearning to earn: The role of remediation in vocational
education and job training. American Journal of Education, 103, 54-93.
Gudan, S., Clack, D., Tang, K., & Dixon, S. (1991). Paired classes for success. Livonia, MI:
Schoolcraft College.
Hull, G. (undated). Alternatives to remedial writing: Lessons from theory, from history, and a
case in point. Paper prepared for the Conference on Replacing Remediation in Higher
Education, National Center for Postsecondary Improvement, Stanford University.
Berkeley: School of Education, University of California.
Kane, T., & Rouse, C. (1995a). Comment on W. Norton Grubb, The varied economic returns to
postsecondary education: New evidence from the class of 1972. Journal of Human
Resources, 30, 205-221.
Kane, T., & Rouse, C. (1995b). Labor market returns to two- and four-year colleges. American
Economic Review, 85, 600-614.
Knapp, M. S., & Turnbull, B. J. (1990, January). Better schooling for the children of poverty:
Alternatives to conventional wisdom. Volume I: Summary. Washington, DC: U.S.
Department of Education, 1990. (ERIC Document Reproduction Service No. ED 314
548)
47
Knapp, M., & Associates. (1995). Teaching for meaning in high-poverty classrooms. New York:
Teachers College Press.
Knapp, M., Shields, P., & Turnbull, B. (1993). Academic challenge for the children of poverty.
Volume 1: Findings and conclusions. Washington, DC: U.S. Department of Education.
Labaree, D. (1997, Spring). Private goods, public goods: The American struggle over
educational goals. American Educational Research Journal, 34, 39-81.
Learning Assessment and Retention Consortium (LARC). (1989a, August). Student outcomes
study: Fall 1988 follow-up study of students enrolled in remedial writing courses in fall
1986 and remedial reading courses in fall 1987. Santa Ana, CA: Rancho Santiago
College.
Learning Assessment and Retention Consortium (LARC). (1989b, October). Student outcomes
study: Mathematics, year 3. Santa Ana, CA: Rancho Santiago College.
Lavin, D., & Hyllegard, D. (1996). Changing the odds: Open admissions and the life chances of
the disadvantaged. New Haven: Yale University Press.
Losak, J., & Morris, C. (1985). Comparing treatment effects for students who successfully
complete college preparatory work (Research Report No. 85-45). Miami, FL: Miami-
Dade Community College, Office of Institutional Research.
Martinson, K., & Friedlander, D. (1994, January). GAIN: Basic education in a welfare-to-work
program. New York: Manpower Demonstration Research Corporation.
Matthews, R.S. (1994a, Spring). Notes from the field: Reflections on collaborative learning at
LaGuardia. Long Island City, NY: Office of the Associate Dean for Academic Affairs,
LaGuardia Community College.
Matthews. R.S. (1994b). Enriching teaching and learning through learning communities. In T.
O’Banion & Associates (Eds.), Teaching and learning in the community college. (pp.
179-200). Washington DC: American Association of Community Colleges.
McCabe, R., & Day, P. (1998, June). Developmental education: A twenty-first century social and
economic imperative. Mission Viejo, CA: League for Innovation in the Community
College and The College Board.
48
McGrath, D., & Spear, M.B. (1991). The academic crisis of the community college. New York:
State University of New York Press.
McKenna, M.C., Miller, J.W., & Robinson, R.D. (1990). Whole language: A research agenda for
the nineties. Educational Researcher, 19, 3-6.
Morris, C. (1994, November). Success of students who needed and completed college
preparatory instruction. (Research Report No. 94-19R). Miami: Institutional Research,
Miami-Dade Community College.
Murnane, R., Willett, J., & Boudett, K.P. (1995). Do high school dropouts benefit from obtaining
a GED? Educational Evaluation and Policy Analysis, 17, 133-148.
Quinn, L., & Haberman, M. (1986, Fall). Are GED certificate holders ready for postsecondary
education? Metropolitan Education, 2, 72-82.
Richardson, R.C., Jr., Fisk, E.C., & Okun, M.A. (1983). Literacy in the open-access college.
San Francisco: Jossey-Bass.
Romberg, T., & Carpenter, T. (1986). Research on teaching and learning mathematics: Two
disciplines of scientific inquiry. In M.C. Wittrock, (Ed.), Handbook of Research on
Teaching, (Third edition, pp.850-873). New York: McMillan.
Rosenbaum, J. (1998, October). Unrealistic plans and misdirected efforts: Are community
colleges getting the right message to high school students? Occasional Paper. New York:
Community College Research Center, Teachers College, Columbia University.
Roueche, J. E., & Roueche, S. D. (1993). Between a rock and a hard place: The at-risk student in the
open-door college. Washington, DC: American Association of Community Colleges.
Russell, A. (1998, January). Statewide college admissions, student preparation, and remediation
policies and programs. Denver: State Higher Education Executive Officers.
Shaughnessy, M. (1977). Errors and expectations : A guide for the teacher of basic writing.
New York: Oxford University Press.
49
Sticht, T. (1979). Developing literacy and learning strategies in organizational settings. In H.
O'Neill, (Ed.), Cognitive and Affective Learning Strategies. New York: Academic Press.
Sticht, T., Armstrong, W. B., Caylor, J. S., & Hickey, D. T. (1987). Cast-off youth: Policy and
training methods from the military experience. New York, NY: Praeger Publishers.
Student outcomes research project, final report. (1996, December). Division of Adult and
Continuing Education, City University of New York.
Tinto, V. (1987). Leaving college: Rethinking the causes and cures of student attrition. Chicago:
University of Chicago Press.
Tinto, V., Goodsell-Love, A., & Russo, P. (1994). Building learning communities for new
college students: A summary of research findings of the Collaborative Learning Project.
Washington D.C.: National Center on Postsecondary Teaching, Learning, and
Assessment, Office of Educational Research and Improvement, U.S. Department of
Education.
Tinto, V., Russo, P., & Kadel, S. (1994). Constructing educational communities: Increasing
retention in challenging circumstances. AACC Journal, 64, 26-29.
Tokina, K. (1993). Long-term and recent student outcomes of freshman interest groups. Journal
of the Freshman Year Experience, 5, 7-28.
Tokina, K., & Campbell. F. (1992). Freshman interest groups at the University of Washington:
Effects on retention and scholarship. Journal of the Freshman Year Experience, 4, 7-22
Traub, J. (1994, Sept. 19). Class struggle. New Yorker, 70, 76-90.
Weisberg, A. (1988). Computers, basic skills, and job training programs: Advice for
policymakers and practitioners. New York, NY: Manpower Demonstration Research
Corporation.
Weinstein, C, Dierking, D., Husman, J., Roska, L., & Powdrill, L. (1998). The impact of a course
on strategic earning on the long-term retention of college students. In J. Higbee and P.
Dwinnel (Eds.), Developmental Education: Preparing Successful College Students, pp.
85-96. Columbia, SC: National Research Center for the First-Year Experience and
Students in Transition.
50
Worthen, H. (1997). Signs and wonders: The negotiation of literacy in community college
classrooms. Unpublished doctoral dissertation, School of Education, University of
California, Berkeley.
51