Criterion-Referenced and Norm-Referenced Assessments: Compatibility and Complementarity
Criterion-Referenced and Norm-Referenced Assessments: Compatibility and Complementarity
To cite this article: Beatrice Lok, Carmel McNaught & Kenneth Young (2015): Criterion-
referenced and norm-referenced assessments: compatibility and complementarity,
Assessment & Evaluation in Higher Education, DOI: 10.1080/02602938.2015.1022136
Download by: [University of California, San Diego] Date: 27 January 2016, At: 02:28
Assessment & Evaluation in Higher Education, 2015
http://dx.doi.org/10.1080/02602938.2015.1022136
Introduction
Assessment of student learning is an essential element in higher education. Raw
scores in assessment are seldom reported as such, but are typically coarse-grained
into broad bands or grades, for example, letter grades. In criterion referencing, each
student is judged against predetermined absolute standards or criteria, without regard
to other students; thus, it is possible for a majority to obtain the top grade, say ‘A’,
or, conversely, for none to do so. In norm referencing, a predetermined percentage
of students (usually with some margin of flexibility) would obtain a certain grade; if
the entire class is inadequate, there would still be the same number of ‘A’s, and,
conversely, if the entire class is outstanding, the same number of ‘D’s must still be
awarded.
It is clear that neither approach, in its pure form, is appropriate in extreme sce-
narios. Most teachers, if pressed, would confess to a pragmatic hybrid, expecting
some level of absolute performance and at the same time respecting the convention
on grade distribution. This approach is coming under increasing stress for a variety
of reasons. The hybrid lacks conceptual clarity, for example, in situations when the
two considerations come into conflict. At a micro-level, such lack of clarity invites
arbitrary and inconsistent practice. At a macro-level, it sits ill with the modern
Purposes of assessment
It is impossible to discuss the relative merits of criterion referencing vs. norm
referencing without considering the purposes of assessment.
Formative assessment
Formative assessment aims to provide feedback to learners and focuses on improve-
ments facilitated by information on what has been mastered and where weaknesses
lie. For this purpose, criterion referencing appropriately reflects the level of achieve-
ment of an individual, whether in self- or diagnostic assessments. The depth of feed-
back can be varied according to the purpose and design of assessment. In the
simplest case, the feedback may be expressed as a single ‘cut-off point’ (i.e. pass or
fail) with no differentiation about degrees of attainments. More commonly, further
diagnostic details are given. As examples, the computer-based Test of English as a
Foreign Language is reported with a score out of 300 and with indicators of profi-
ciency level on specific skills; while an essay can be marked with detailed written
comments against criteria specified in a rubric, with or without an overall grade.
Feedback in formative assessment assists learners to understand the depth of the
criteria and, in forcing the marker to be conscious of different levels of attainment
with respect to the criteria, helps to sharpen the scoring.
Summative assessment
Summative assessment provides scoring information that can be read and easily
understood by a range of stakeholders, including external stakeholders; for this rea-
son, and in contrast to the feedback function in formative assessment, the term
‘feedout’ is sometimes used. Most public examinations at the end of secondary
school are summative in nature.
Because criteria (and their interpretation) are likely to be unfamiliar to external
stakeholders, there is often the need to express summative assessment in norm-refer-
enced terms. Most employers would find a statement such as ‘in the top quartile of
University X’ useful, if they have some familiarity with the university. Many gradu-
ate schools ask referees to state the percentile position of the applicant (e.g. ‘truly
outstanding, top 5%’), incidentally betraying a possible lack of confidence in grades
and grade distributions.
Assessment & Evaluation in Higher Education 3
Learners themselves also want to know where they stand amongst peers because
the answer indicates their competitiveness amongst the group. School students par-
ticularly need this type of information when competing for limited university places.
In these circumstances, summative assessment with a norm-referenced element is
essential because there is truth in the statement that ‘the real world is norm refer-
enced and not criterion referenced’: if there are 20 desirable jobs for 100 applicants,
the top 20% (no matter what their absolute standards may be) will get them; the
same is true for all quota-limited contexts: undergraduate places (especially in highly
sought-after institutions and programmes), graduate school admission, awards of
scholarships, etc.
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
Increasing tension
Whether the assessment is for formative or summative purposes, the tension between
the criterion-referenced and norm-referenced methodologies has been accentuated by
two developments: the trend towards outcomes-based approaches in higher educa-
tion and the phenomenon of grade inflation, leading to the need for instruments to
combat it.
Outcomes-based approaches
In an outcomes-based approach, a clear articulation of desired learning outcomes
takes centre stage in the curriculum, and drives the design of the learning environ-
ment and process. The choice of content, the design of learning activities and the
nature of assessment should be aligned with the stated learning outcomes.
The concept of outcomes-based approaches was influenced by the reform of
education in the USA in the second half of the twentieth century. The complex
political (e.g. the cold war) and ideological (e.g. race relations) reasons that drove
the reform are beyond the scope of this paper. While the main emphasis was on
school education, a flow-on effect into higher education took place – in terms of
both funding and policy.
A number of education problems were tackled in this reform era (Spady 1994):
(1) ‘Clarity of focus’: educators must clearly illustrate the intended learning
outcomes with students and prioritise them in planning, teaching and
assessment;
(2) ‘Designing down’: intermediate outcomes are valuable in scaffolding or
supporting achievement of the broader outcomes;
(3) ‘High expectations’: the level of acceptable performance should be raised;
this entails abandoning a norm-referenced assessment approach and opening
access to all students for high-level learning;
(4) ‘Expanded opportunity’: while the learning outcomes might be fixed, the
time to achieve could vary.
Carroll (1963) also argued that uniform minimum standards should be clearly
stated and understood, and that students should be given the time to achieve these
standards. Bloom (1968) followed Carroll’s thinking and developed the notion of
mastery learning, which has become a core principle of outcomes-based approaches
4 B. Lok et al.
through all levels of education. Parallel processes to those in the USA also took
place in most of the English-speaking world outside Asia.
Much of the early emphasis on specification of reliably measurable outcomes
can be traced to the behaviourist tradition in education with its roots in Skinnerian
psychology (e.g. Tyler 1949). However, particularly since the 1980s, the challenge
of the articulation and measurement of complex and multifaceted educational out-
comes has received increasing attention (Ewell 2009).
Of course, conversations and policy at national and institutional levels often take
time to filter down to departments and individual teachers. Outcomes-based
approaches adopted by a university for quality-assurance (and perhaps even peda-
gogical) reasons may not have actually made much difference to the learning experi-
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
ence of students; the decoupling of policy and rhetoric from practice is not unknown
in other aspects of higher education. The adoption of outcomes-based approaches is
not an on/off switch, and should not be portrayed as easily achievable.
For example, outcomes which are clearly defined and easily measurable can be
quite low level in cognitive demand. However, the essential skills and capacities
that should be fostered in university involve a variety of intellectual, interpersonal
and personal capabilities, such as critical thinking, creative thinking, self-managed
learning, adaptability, problem-solving, communication skills, interpersonal skills
and group work, and computer literacy. Articulating these desired outcomes in
terms of concrete criteria is challenging and possibly self-defeating; moreover,
teachers need to be able to couch discipline-domain outcomes within these
broader capabilities. One solution to this dilemma is to use a checklist/grid/taxon-
omy to monitor the balance of stated outcomes at both programme and course
level. Two main taxonomies are that of Bloom (1956) and its revision (Anderson
and Krathwohl 2001; Krathwohl 2002); and the Structure of the Observed
Learning Outcomes taxonomy (Biggs and Collis 1982) and its simplification (e.g.
McNaught, Cheng, and Lam 2006).
Outcomes-based approaches require clearly stated assessment criteria pegged to
the declared outcomes, and therefore argue for – indeed mandate – the use of crite-
rion referencing in assessment. The strong synergy between outcomes-based
approaches and criterion referencing means that curriculum developers should
address assessment processes explicitly and should not leave the implementation of
outcome statements to the whim of individual teachers. To abandon or minimise
explicit criteria in assessment would undermine the entire curriculum design.
Achieving good curriculum alignment is an iterative process. Thus, outcomes-based
approaches should focus not only on the reliability of educational measurement, but
also on the principle of curriculum alignment (Biggs 1999; Biggs and Tang 2011) in
curriculum development (Shay 2008). Outcomes-based approaches, and the quality
framework of which they are a part, refer not only to assessments, but to the entire
curriculum, with a focus on alignment between clearly articulated outcomes, learn-
ing and pedagogy targeted towards those outcomes, and assessments against the
same outcomes.
Grade inflation
Grade inflation erodes public trust: the public may be inclined to believe that an
‘A’ grade equates to the top 10 or 20% of the class; in practice, if only
Assessment & Evaluation in Higher Education 5
Dilemma
One is therefore caught in an apparent dilemma: the modern quality framework
pushes institutions towards outcomes-based approaches and therefore criterion-refer-
enced assessment, but the abandonment of norm-referenced assessment and espe-
cially grade-distribution guidelines can allow rampant grade inflation. It is therefore
imperative to re-examine the true natures of both criterion referencing and norm
referencing in order to find a way out of the dilemma.
criterion referencing for assessing undergraduate written work. They stressed that
criterion referencing can emphasise higher-order cognitive skills, such as critical
thinking, and clear and incisive writing – though one must caution that such criteria
are difficult to interpret without a context, a point to which we shall return. Wiliam
(1996) also reviewed the respective drawbacks of criterion referencing and norm
referencing, and argued that neither approach was able to offer an adequate and
authentic assessment of performance in higher education.
The differences and similarities between criterion-referenced and
norm-referenced assessments are summarised in Table 1. These distinctions are
mainly conceptual, but blurred in practice.
This present paper does not intend to review the literature on comparing and
contrasting the two approaches in detail; rather, it aims to explore the operational
compatibility of criterion referencing and norm referencing in higher education
assessment.
Many assessment strategies are compatible with both criterion-referenced and
norm-referenced grading methods, and the boundary between the two is fuzzy.
Nightingale et al. (1996, 9) noted that it was difficult to distinguish some assessment
strategies as ‘entirely norm-referenced or criteria-referenced’. This corresponded to
the argument of Rowntree (1987) that criterion referencing and norm referencing
share more commonalities than are generally perceived and any rigid adherence to
either is problematic. In contrast, Frisbie (2005) argued that the different score inter-
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
pretations led to the perception of a dichotomy that requires a binary choice between
criterion referencing and norm referencing, diverting attention away from the
possibility of their synergistic use in one single assessment.
Dual reporting
There are many examples in which both criterion referencing and norm referencing
are present with both dimensions reported in some form. For example, dual report-
ing (e.g. Johnny is able to do X, Y, Z (criteria) and ranks in the top 15% (norm
information) of the class in this subject) is commonly used in many schools in Hong
Kong, with the report card providing for each subject both a grade or numerical
mark (e.g. 60 marks) as well as a position-in-class indicator (e.g. rank 35 out of 60)
for the subject. In this highly competitive education system, the position-in-class
indicator assumes high importance.
Dual reporting can also be found in the California Critical Thinking Skills Test
(CCTST), a popular standardised critical-thinking test in higher education. CCTST
aims to measure an individual’s critical-thinking capability in various skill areas,
and its report provides an overall score and scale scores together with a norm-refer-
enced percentile ranking. In addition, it offers descriptions of the strength of both
scores (Facione and Facione 2009).
The SAT (2015), (formerly Scholastic Aptitude Test, Scholastic Assessment Test
and the SAT Reasoning Test) is generally seen as norm referenced (scores are inter-
preted as percentiles), but one can argue that there are definite criteria associated
with the assessment (Angoff 1974). The national College English Test (CET) in
China provides ‘the dual function’ using criteria related to norms (Zheng and Cheng
2008, 416), with benchmarks adopted for the two levels (i.e. CET-4 and CET-6), but
with scores also equated with predetermined norms (Yang and Jin 2001).
Often in dual reporting, the normative information emerges a posteriori and as a
fact (e.g. a score of 650 happens to equate to the top decile), whereas, strictly
speaking, norm referencing imposes grade distributions a priori and as a policy (e.g.
no more than 10% of the candidates shall be awarded ‘A’s). The terminology
8 B. Lok et al.
Criteria: standards
Determine Grading
for each grade
Distribution of Leads to
Rank or percentile
grades position
(e.g. deciles vs. letter grades) reflects this distinction. However, for a large and stable
candidature, it no longer matters in practice which comes first. Imagine that one
starts with a criterion-referenced approach, so that scores are tied to absolute perfor-
mance. It is found, empirically, that over a number of years the median score in,
say, the SAT is about 500 (National Center for Education Statistics 2013). Then,
because the ability of the candidature is stable, in the next year, the logic can be
reversed: one defines 500 to be the median, which is a norm-referenced prescription.
Betebenner (2009) also suggested the combined use of norm referencing and crite-
rion referencing as a way to profile student growth and development. This state of
affairs can be represented by two free-floating boxes – free floating in the sense that
they are not linked in design, only a posteriori (Figure 1).
ment at a level that is unusual for such a cohort’ – again betraying an implicit norm
reference. Making the norm reference explicit would render it difficult to defend a
large percentage of ‘A’s on the excuse that criterion referencing should not be bound
by percentage distributions. This consideration is especially important in universi-
ties, where higher-order skills such as ‘ability to synthesise a cogent argument’ or
‘critical thinking’ are increasingly emphasised, and these are especially prone to
wide interpretation, in contrast to skill-based criteria such as ‘typing at 60 words per
minute’ or ‘the ability to add’.
There are other reasons why criterion referencing may require the monitoring of
norm-based distributions, for example, when consistency across multiple assessors
in interpreting a set of criteria has to be maintained. In this regard, Orr (2007)
stressed that assessment is co-constructed in communities of practice, implicitly
based on past (normative) knowledge. In large public examinations in which scripts
are randomly assigned to markers, even if there are descriptors attached to each
grade, it is common to ensure that the distributions of scores are broadly similar
across markers. Norm referencing, as a result, becomes a strategy for checking on
decisions made in a criterion-referenced fashion.
Thus, behind the criterion-referenced rubric (that the students see) is an assump-
tion about norms (that the teachers and assessors should be aware of). This state of
affairs can be represented as in Figure 2.
Feedback loop
The third model, extending the one just described, involves three elements: norm
referencing, criterion referencing and actual grading, all in a feedback loop, which
can be iterated many times (Figure 3). First, assumptions about the ability or perfor-
mance of the cohort (norm referencing) are used to derive a set of criteria, which are
then incorporated into assessment rubrics (criterion referencing). Second, these are
applied in assessment to arrive at a set of grades. The markers should simply follow
the rubrics, i.e. act in a criterion-referenced fashion. Third, after the grading is done,
the actual student performance is monitored. Theoretically, the distribution may be
very different from that assumed at the beginning; but with experienced teachers,
good moderation and/or other adjustments, after a few cycles discrepancies would
be small and would appear only slowly over time (as student abilities change), and
the criteria should thus be suitably revised for future assessments. For example, if a
high percentage of students are scoring ‘A’s according to the rubric, this new empiri-
cal evidence should inform the normative assumptions, leading to a revised set of
10 B. Lok et al.
Scrutinized
and used to
adjust Determine
assumed
distribution
Grading
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
criteria, thus ensuring that the rubrics for the top grade remain sufficiently challeng-
ing and the educational experience sufficiently rewarding; the opposite adjustment
should be made if the original criteria were discovered to be too demanding. One
might say this is changing the goal posts, but that is not an appropriate metaphor,
since the procedure is not to change the goal posts during the game, but only for the
next game, and in a way that is announced before the next game starts. The three
steps together constitute a feedback loop, which is consonant with the modern
concept of fitness-for-purpose, or the quality loop sometimes abbreviated as ADRI
(approach–deploy–review–improve) (Woodhouse 2003).
Thinking in terms of such a feedback loop serves to underscore a number of
features:
Taken together, these features constitute a theoretical basis for clarifying the
implicit hybrid model cited at the beginning of this paper.
It may be added that this nuanced model is unnecessary in contexts that require
only certification of skills – in which case results are typically scored as pass/fail,
Assessment & Evaluation in Higher Education 11
multiple attempts are usually allowed, and there is no worry about grade inflation or
an overwhelming passing rate. The example of driving tests comes to mind, and also
other skill-based assessments including those in the use of information technology,
some of which are in fact likened to driving tests, for example, the International
Computer Driving Licence.
Other issues
Norm referencing to which population
There is one further problem: in norm referencing, what is the reference population
for deriving the norm? The obvious answer seems to be the actual class being exam-
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
ined but, very often, reference is made (implicitly) to a larger population. If several
instructors teach and examine different sections of the same course, there is usually
some procedure to homogenise or moderate the assessment. Then it is legitimate to
have a very high percentage of ‘A’s in one section if enrolment patterns for the bet-
ter students are for some reason skewed. In schools, the grading is often referenced
to the estimated distribution in the analogous public examination, which is useful
feedback in preparing for the latter. In higher education, external examiners (a prac-
tice now diluted in many nations) allow reference to a group of peer institutions, or
the students therein. The ideal is that a first-class honours from any university means
the same thing – a statement that few would now take seriously. Thus, the basis of
comparison in norm referencing can be quite complex. In short, there is often an
implicit norm reference to a larger (external) population, whether this is channelled
formally through an external examiner or, more often, informally by the teacher
benchmarking against standards elsewhere (often relying on her/his own experience
in graduate school, as a tutor or teaching assistant).
Anchoring norm referencing to a larger population enhances the reliability of an
assessment; however, the intended learning outcomes central to an outcomes-based
curriculum can only be assessed with any validity by a criterion-referenced
approach. The feedback loop offers flexibility for educators to make iterative adjust-
ments so that claims about both reliability and validity are possible.
broad, a passing grade can usually be achieved by just showing up and trying
conscientiously. In such cases, ipsative referencing can be introduced to measure the
learning gain of individuals throughout the process for different purposes. In
continuing professional development, very often only attendance is required.
Conclusion
There are two main principles of assessment: (1) to grade students with differing
levels of knowledge and skills; and (2) to reflect the growth of individual knowledge
and skills. Teachers have to balance the assessment focus between these principles
and develop an appropriate assessment strategy. Considering the competitiveness of
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
the current job market for graduates, an assessment profile which can reflect not
only absolute competence and learning development, but also the relative position
within a cohort is useful. In addition, students’ confidence about the validity of
assessment measures needs to be maintained. In terms of norm referencing and crite-
rion referencing, there are theoretically four possibilities:
Normative guidelines
First, institution-wide grade-distribution guidelines need to be adopted and publi-
cised for the information of external stakeholders. There must be mechanisms to
ensure that these guidelines are adhered to and rare exceptions are documented.
Such practice should be uniform across an institution (ideally even over an entire
sector), at least for each level of learning, and need make no reference to subject
content.
Subject-specific criteria
Second, for each subject (either an undergraduate degree or a component module,
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
e.g. one course over one term), the intended learning outcomes need to be stated as
part of the curriculum design, and these should be further articulated into grading
rubrics or criteria. The criteria should span different levels of sophistication, includ-
ing some that (in a formative sense) stretch the more able students, or (in a summa-
tive sense) are likely to be met only by a minority who are then to be awarded the
top grades. The range of sophistication promotes depth of learning. The formulation
of these criteria would, of course, be subject-specific, and would rely on assump-
tions about student ability presumably drawn from past cohorts and/or lateral bench-
marking to an external reference.
Quality-assurance process
Educational processes are not well served by either hard rules (‘accountability’) or
the total absence of rules (‘autonomy’). A healthy balance can only be struck if there
is a quality-assurance process which checks that broad normative guidelines exist
and are well understood, that there is broad (but not necessarily literal) adherence to
these guidelines, that exceptions are documented and, most importantly, that con-
flicts are dealt with in the feedback loop. The quality-assurance process is needed at
multiple levels: within a department monitoring its different courses, within the
institution to monitor the different teaching units and even across the sector by a
quality agency advising institutions under its purview – in all cases, not just ticking
check boxes but helping the units to develop robust processes.
It must be said that criterion referencing and norm referencing can be brought
into consistency much more simply, by mandating shifts in the grade boundaries
14 B. Lok et al.
whenever grade distributions fall outside guidelines – a practice that is not uncom-
mon. But such a mechanical adjustment would engender a sense of passive compli-
ance with apparently bureaucratic rules, and will have no effect on learning and
teaching in subsequent years. On the other hand, the feedback loop provides the
opportunity for alignment, leading to changes not only in assessment, but through
the revision of intended learning outcomes, also in learning and teaching. Only
when all components of the curriculum are considered together can there be the
alignment that lies at the heart of outcomes-based approaches.
Acknowledgements
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
Disclosure statement
No potential conflict of interest was reported by the authors.
Notes on contributors
Beatrice Lok is a visiting lecturer of the Centre for the Advancement of Teaching and Learn-
ing (CALT) at University College London. She was a postdoctoral fellow of the Centre for
Learning Enhancement And Research at The Chinese University of Hong Kong. She is inter-
ested in the areas of quality enhancement in higher education, second-language learning
motivation and second-language development of non-native speaking students.
Carmel McNaught is Emeritus Professor of Learning Enhancement and former director of the
Centre for Learning Enhancement And Research (CLEAR) at The Chinese University of
Hong Kong. Since the early 1970s, Carmel has worked in higher education in Australasia,
southern Africa and the UK in the fields of chemistry, science education, second-language
learning, eLearning, and higher education curriculum and policy matters. She is actively
involved in several professional organizations and is a fellow of the Association for the
Advancement of Computers in Education; is a university quality-assurance consultant and
auditor in Australia and Hong Kong; has served on the editorial boards of 18 international
journals; and is a prolific author with well over 300 academic publications. Her recent pub-
lications and activities can be viewed at http://www.cuhk.edu.hk/clear/people/Carmel.html.
She is currently a consultant, working mostly in Australia, Hong Kong, New Zealand,
Singapore, the UAE and the UK.
Kenneth Young is master of CW Chu College and also professor of Physics at The Chinese
University of Hong Kong. For a number of years until 2011, he was Pro-Vice-Chancellor of
the University with responsibility for education. Currently, he is a member of the Quality
Assurance Council under the Hong Kong University Grants Committee. He also chairs the
Curriculum Development Council in Hong Kong, and is a member of the Education
Commission. He is a fellow of the American Physical Society.
Assessment & Evaluation in Higher Education 15
References
Anderson, L. W., and D. R. Krathwohl. 2001. A Taxonomy for Learning, Teaching, and
Assessing: A Revision of Bloom’s Taxonomy of Educational Objectives. Boston, MA:
Allyn & Bacon.
Angoff, W. H. 1974. Criterion-referencing, Norm-referencing and the SAT. Research
Memorandum. Princeton, NJ: Educational Testing Service.
Betebenner, D. W. 2009. “Norm- and Criterion-referenced Student Growth.” Educational
Measurement: Issues and Practice 28 (4): 42–51.
Biggs, J. 1999. “What the Student Does: Teaching for Enhanced Learning.” Higher
Education Research & Development 18 (1): 57–75.
Biggs, J. B., and K. F. Collis. 1982. Evaluating the Quality of Learning: The SOLO
Taxonomy (Structure of the Observed Learning Outcome). New York: Academic Press.
Biggs, J., and C. Tang. 2011. Teaching for Quality Learning at University. 4th ed.
Downloaded by [University of California, San Diego] at 02:28 27 January 2016
Orr, S. 2007. “Assessment Moderation: Constructing the Marks and Constructing the
Students.” Assessment & Evaluation in Higher Education 32 (6): 645–656.
Rojstaczer, S., and C. Healy. 2012. “Where A is Ordinary: The Evolution of American
College and University Grading, 1940–2009.” Teachers College Record 114 (7): 1–23.
Rowntree, D. 1987. Assessing Students: How Shall We Know Them? Revised ed. London:
Kogan Page.
Rust, C., B. O’Donovan, and M. Price. 2005. “A Social Constructivist Assessment Process
Model: How the Research Literature Shows Us This Could Be Best Practice.” Assessment
and Evaluation in Higher Education 30 (3): 231–240.
Sadler, D. R. 2005. “Interpretations of Criteria-based Assessment and Grading in Higher
Education.” Assessment & Evaluation in Higher Education 30 (2): 175–194.
SAT. 2015. Understanding Your Scores. Accessed January 20. http://sat.collegeboard.org/
scores/understanding-sat-scores
Shay, S. 2008. “Beyond Social Constructivist Perspectives on Assessment: The Centering of
Downloaded by [University of California, San Diego] at 02:28 27 January 2016