0% found this document useful (0 votes)
28 views

Reliability

The document discusses the concept of reliability in educational assessment. It defines reliability as the consistency, stability, or dependability of measurement results over time and conditions. The document then describes five types of reliability: test-retest, parallel forms, internal consistency, inter-rater, and split-half. It provides details on each type, such as how they are measured and examples. Internal consistency is further broken down into split-half reliability and Cronbach's alpha.

Uploaded by

Fareena Baladi
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
28 views

Reliability

The document discusses the concept of reliability in educational assessment. It defines reliability as the consistency, stability, or dependability of measurement results over time and conditions. The document then describes five types of reliability: test-retest, parallel forms, internal consistency, inter-rater, and split-half. It provides details on each type, such as how they are measured and examples. Internal consistency is further broken down into split-half reliability and Cronbach's alpha.

Uploaded by

Fareena Baladi
Copyright
© © All Rights Reserved
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 9

Understanding

Reliability in
Educational
Assessment

Muhammad Roshan
Understanding Reliability
• Reliability in assessment refers to the consistency, stability, or
dependability of the measurement process.
• It is the extent to which a test or assessment tool produces consistent and
reproducible results over time and across different situations.
• In other words, a reliable assessment should yield similar scores for the
same individual or group under consistent conditions.
Types of Reliability
1. Test-Retest Reliability:
2. Parallel Forms Reliability (Alternate Forms Reliability)
3. Internal Consistency Reliability
4. Inter-Rater Reliability
5. Split-Half Reliability
1. Test-Retest Reliability

• This type of reliability assesses the consistency of scores over time


when the same test is administered to the same group of individuals on
two separate occasions. Then looking at the correlation between the two
sets of scores.
• If a group of students takes a math test, and then the same test is
administered to them again a few weeks later, high test-retest reliability
would mean that the scores are very similar on both occasions.
• This does not apply to the volatile subjects like opinions and
perceptions.
2. Parallel Forms Reliability (Alternate Forms Reliability)

• This type of reliability involves using different but equivalent forms of the
same test to assess consistency in measurement.
• It is a measure of reliability obtained by administering different versions of an
assessment tool (both versions must contain items that probe the same
construct, skill, knowledge base, etc.) to the same group of individuals. The
scores from the two versions can then be correlated in order to evaluate the
consistency of results across alternate versions.
• If two versions of a biology test are created, with each version containing
similar questions of equal difficulty, the scores obtained by the same group of
students on both versions should be highly correlated.
3. Inter-Rater Reliability (Observer Reliability)

• This type of reliability assesses the consistency of scores when different


raters or observers assess the same behavior or performance.
• In a performance assessment where multiple raters evaluate a student's
presentation skills, high inter-rater reliability would mean that the raters
assign similar scores for the same presentation.
4. Internal Consistency Reliability

• This type of reliability assesses the consistency of results across items


within the same test. It measures the extent to which different items on
the test measure the same underlying construct.
• It is a measure of reliability used to evaluate the degree to which
different test items that probe the same construct produce similar
results.
• Two Types of Internal Consistency Reliability: Split Half and Cronbach
Alpha
A. Split-Half Reliability:
• Split-half reliability is a method used to assess the internal consistency of a test by splitting it into
two halves and examining the correlation between the scores obtained on each half.
• This method involves splitting a test into two halves and comparing the scores on each half to assess
consistency through correlation.
• The goal is to determine whether the items within a test are measuring the same underlying construct
consistently. This reliability method assumes that the two halves of the test are equivalent, and any
differences in scores are due to measurement error.
• If a language proficiency test is divided into two halves, and the scores on both halves are highly
correlated, it indicates good split-half reliability. If it indicates different results, it shows low
internal consistency.
B. Cronbach's alpha

• Cronbach's alpha is a common measure of internal consistency


reliability. For instance, in a personality test, if questions related to
extraversion consistently correlate with each other, it indicates high
internal consistency.
• If the value of alpha is ≥ 0.7 then it shows the higher level of internal
consistency.

You might also like