Classroom Basedassessment
Classroom Basedassessment
Assessment in L2
Contexts
Classroom-based
Assessment in L2
Contexts
Edited by
Dina Tsagari
Classroom-based Assessment in L2 Contexts
All rights for this book reserved. No part of this book may be reproduced,
stored in a retrieval system, or transmitted, in any form or by any means,
electronic, mechanical, photocopying, recording or otherwise, without
the prior permission of the copyright owner.
INTRODUCTION ............................................................................................. 1
Dina Tsagari
CHAPTER FIVE............................................................................................ 96
Examining the Quality and Quantity of the Washback Effect of the FCE
Test in Greece
Lambrini Loumbourdi
DINA TSAGARI
1
See http://www.ealta.eu.org/resources.htm.
2 Introduction
Works Cited
Alderson, J. Charles and Dianne Wall. 1993. “Does Washback Exist?”
Applied Linguistics 14:115–29.
Black, Paul, and Dylan Wiliam. 2009. “Developing the Theory of
Formative Assessment.” Educational Assessment, Evaluation, and
Accountability 21:5–31.
Black, Paul, and Dylan William. 1998. “Assessment and Classroom
Learning.” Assessment in Education 5:7–74.
Council of Europe. 2001. Common European Framework of Reference for
Languages: Learning, Teaching, Assessment. Cambridge: Cambridge
University Press.
Purpura, James E. and Carolyn E. Turner. forthcoming. Learning-oriented
Assessment in Language Classrooms: Using Assessment to Gauge and
Promote Language Learning. New York, NY: Routledge, Taylor &
Francis.
Rea-Dickins, Pauline. 2004. “Understanding Teachers as Agents of
Assessment.” Language Testing 21(3):249–58.
Rea-Dickins, Pauline. 2008. “Classroom-based Assessment.” In
Encyclopedia of Language and Education, edited by Elana Shohamy
and Nancy H. Hornberger, 1–15. USA: Springer.
Tsagari, Dina and Ildikó Csépes (eds.) 2011. Classroom-based Language
Assessment. Frankfurt am Main: Peter Lang GmbH.
Tsagari, Dina. 2009. The Complexity of Test Washback. Frankfurt: Peter
Lang.
Turner, E. Carolyn. 2012. “Classroom Assessment.” In The Routledge
Handbook of Language Testing, edited by Glenn Fulcher and Fred
Davidson, 65–78. London and New York: Routledge.
PART I:
ASSESSMENT LITERACY
FOR LANGUAGE TEACHER
ANTHONY GREEN
his greatest concern over the shortcomings of teacher training, noting that
“the majority of teacher education programmes require no training in
assessment for graduation” (Stiggins 1991, 535). He suggested that
training teachers in assessment was essential to effective student learning,
and called for a redirection of research efforts towards assessment in the
classroom rather than standardised testing. The need to shift the focus of
attention towards teacher education and the role of assessment in the
classroom has been a persistent theme for advocates of assessment
literacy.
While Stiggins (1991) was writing about general education in US
schools, recent years have seen similar concerns being voiced about levels
of assessment literacy in international language education. As in Stiggins’
article, the difficulties that a wide range of groups experience in
interpreting and using assessment data have been discussed. These groups
include university admission officers, policy makers, and the general
public. However, the topic attracting the greatest attention has been “the
assessment skills and understandings currently perceived as vital for
conducting language assessment in educational settings” (Inbar-Lourie
2008, 387): the core competences in assessment required for effective
language teaching.
Jeong (2013) also followed Brown and Bailey (2008) in her survey of
140 language testing course leaders around the world. Her goal was to
involve instructors who lacked specialist language assessment qualifications
in order to correct a bias towards language testing specialists in Brown and
Bailey’s data (their respondents had been recruited via a language testing
discussion list). Although her data suggested that courses taught by both
groups tended to follow similar patterns, it also pointed to a greater
diversity of provision. Non-specialists tended to give greater attention to
classroom assessment topics while specialists spent more time on statistics.
noted findings from previous studies showed that tests made by teachers
tended to be of poor quality, “were too difficult or too easy; … measured
content that had not been taught in class or not specified in the syllabus…
[and] did not show what students had actually achieved” (227). His study
explored the effects of a basic training course in assessment (provided as
part of a graduate course in English Language Teaching) on the quality of
teacher-produced tests.
Although teachers were able to improve their material by editing,
reviewing, and redrafting, even after passing through the recommended
stages of test construction, the resulting tests generally failed to satisfy
basic quality criteria. In addition to insufficient training, Coniam’s (2009)
informants reported that practical constraints had a negative impact on
their work: lack of time, resources, and institutional support restricted the
attention they could give to developing effective assessment and
improving their own knowledge and skills.
7. Changing needs
Adding new urgency to the long-standing concern about the lack of
attention to language testing in teacher training, observers have pointed to
a recent expansion in the role that teachers are expected to play in
assessing language learners – including testing for high-stakes purposes
(those with significant implications for the life chances of test takers such
as access to higher education and employment opportunities). At the same
time, new questions have been raised about the adequacy of what little is
on offer.
In addition to the need to understand and develop tests, which is
covered by Lado (1961) and his successors, commentators have noted a
shift in the focus in educational policy-making away from inputs to
learning (curriculum) towards outcomes from learning (as evidenced by
assessment results) (Cumming 2009). The burden for generating evidence
of outcomes has mainly fallen on teachers, greatly increasing both the
amount of assessment they are asked to carry out and the importance
placed on it.
five most popular text books on the survey were all first published in the
1980s or 1990s.
The tendency to rely on established sources was even more marked in
the Chinese context covered by Jin’s (2010) study. Bachman (1990) and
Heaton (1988) were the two most popular books (mentioned as core text
books by 24 and 19 respondents, respectively). All the titles of the 12 most
popular books listed in the article featured the words “test” or “testing”
rather than “assessment”, and only two were published after 2000.
Genesee and Upshur (1996) did not feature at all.
and the careful analysis of results. The skills learned from making,
sharing, and justifying judgements about performance on a test can also
serve to sensitise teachers to learners’ strengths and weaknesses in the
classroom. Well-crafted test tasks reveal what learners know in ways that
can also make them useful as tools for teaching and learning.
Whether highlighting commonalities or contrasts between the practices
of classroom assessment and testing, commentators seem to agree that
classroom assessment needs a stronger theoretical underpinning and a
more extensive research base. In recent years, attempts have been made to
build coherent theories, establish criteria for quality, and set an explicit
agenda for research in this area (see for example Davison and Leung 2009;
Hill and McNamara 2011). There would seem to be much to learn from
the educational measurement tradition that informs language testing in
these respects.
At the same time, large-scale language testing has been reforming itself
to better match test content with current approaches to teaching and
learning. Test developers have responded to demands for more authentic
tasks that involve realistic performances in keeping with communicative
or task-based approaches to teaching. Taking account of the increasing
interest in test use and consequences, providers of major global language
tests like the Educational Testing Service (US) and Cambridge English
Language Assessment (UK) have been offering increasingly informative
score reports that include some feedback on test-taker performance. They
have built up their support for teachers preparing learners to take their
tests. Resources such as online practice tests that provide instant results
and test preparation guides are now widely available. Both organisations
have implemented major projects to investigate the impact of their tests on
language classrooms (Stoynoff 2012).
Such developments suggest the possibility of a complementary
relationship between more formative and more summative assessment
functions. Boud (2000, 160) suggested that “every act of assessment we
devise or have a role in implementing has more than one purpose”. He
argued that assessments should be developed to perform what he called
“double duty”. This involves focusing both on the performance of the task
at hand and the long-term educational goals: simultaneously indicating
achievement of objectives and providing insights into the learning process.
Assessment tasks that are designed or implemented with only one function
in mind risk undermining the coherence of the educational process.
Through washback, summative assessments may come to supplant
important learning goals. The reason is that passing tests and gaining
certificates seem to be more prized than developing a comprehensive