0% found this document useful (0 votes)
46 views28 pages

Reliability Test by Group 2

The document discusses the concept of reliability testing and different types of reliability such as test-retest reliability, parallel form reliability, inter-rater reliability, and internal consistency reliability. It also discusses factors that can impact reliability like student factors and how reliability is an important part of ensuring validity.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
46 views28 pages

Reliability Test by Group 2

The document discusses the concept of reliability testing and different types of reliability such as test-retest reliability, parallel form reliability, inter-rater reliability, and internal consistency reliability. It also discusses factors that can impact reliability like student factors and how reliability is an important part of ensuring validity.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 28

Reliability Test

By Group 2

Andre Pratama (1920300011)


Inayati Putri Sitompul (1920300015)
Lutfiah Sa'adah Rangkuti (1920300014)
Nursa'adah Pasaribu (1920300016)
Sri Utami Endang Retno Ningsih (1920300012)
Wildan Wahidi Situmorang (1920300013)
RELIABILITY
A Reliability test is consistent and dependable. If you
give the same test to the student or matched students
on two diffirent occasions, the test should yield similar
of factors that may contribute to the unreliability of a
test. Consider the following possibilities: fluctuations
in the student, in scoring, in test administration, and in
the test itself .
Reliability refers to how dependably or
consistently a test measures a
characteristic. If a person takes the test
again, will he or she get a similar test score,
or a much different score? A test that yields
similar scores for a person who repeats the
test is said to measure a characteristic
reliably.
The reliability of an assessment tool is the
extent to which it consistently and
accurately measures learning. When the
results of an assessment are reliable, we can
be confident that repeated or equivalent
assessments will provide consistent results.
How do we account for an individual
who does not get exactly the same
test score every time he or she takes
the test?
Some possible reasons are the following :
• Test taker's
• Multiple raters.
temporary
psychological
or physical
state. • Test form.
• Environmental
factors.
Student
Related
Reliability
The most Common learner
related issue in reliability is
caused by temporary
illness,fatigue, a ‘’bad day’’,
anxiety, and other physical or
psychological factors, which
may make an observed score
deviate from one’s true score.
Also included in this category
are such factor as a test
taker’s test wiseness or
Types of
Reliability
Test-retest Inter-rater
Reliability reliability

Parallel form or Internal


Alternate- consistency
Forms reliability
Reliability
Test-retest Reliability

Test-retest reliability is
the degree to which scores
are consistent over time. It
indicates score variation
that occurs from testing
session to testing session
as a result of errors of
measurement
Parallel form or Alternate-Forms
Reliability

Parallel forms reliability is a


measure of reliability obtained by
administering different versions of an
assessment tool (both versions must
contain items that probe the same
construct, skill, knowledge base,
etc.) to the same group of
Inter-rater reliability

Inter-rater reliability is a
measure of reliability used
to assess the degree to
which different judges or
raters agree in their
assessment decisions.
Internal consistency
reliability

Internal consistency
reliability is a measure of
reliability used to evaluate
the degree to which
different test items that
probe the same construct
produce similar results.
Test manuals and reviews
report several kinds of
internal consistency
reliability estimates. Each
type of estimate is
appropriate under certain
circumstances. The test
manual should explain why
a particular estimate is
reported.
a. Average inter-item correlation is a
subtype of internal consistency
reliability.

b.Split-half reliability is another


subtype of internal consistency
reliability. The process of obtaining
split-half reliability is begun by
“splitting in half” all items of a test
that are intended to probe the same
area of knowledge (e.g., World War
Split-Half
Reliability:
Requires only one administration. Especially
appropriate when the test is very long. The most
commonly used method to split the test into two is
using the odd-even strategy. Since longer tests tend to
be more reliable, and since split-half reliability
represents the reliability of a test only half as long as
the actual test, a correction formula must be applied to
the coefficient. Spearman-Brown prophecy formula.
Split-half reliability is a form of internal consistency
reliability.
Standard
error of
measurement
It gives the margin of error that you should expect in
an individual test score because of imperfect reliability
of the test. The SEM represents the degree of
confidence that a person's "true" score lies within a
particular range of scores. For example, an SEM of "2"
indicates that a test taker's "true" score probably lies
within 2 points in either direction of the score he or
she receives on the test.
When evaluating the reliability coefficients of a test, it
is important to review the explanations provided in the
manual for the following:

1. Types of reliability used.


2. How reliability studies were conducted.
3. The characteristics of the sample group.
Estimating
the Validity of
a Measure:
1. A good measure must not only be reliable, but
also valid
2. A valid measure measures what it is intended
to measure
3. Validity is not a property of a measure, but an
indication of the extent to which an assessment
measures a particular construct in a particular
context—thus a measure may be valid for one
purpose but not another
4. A measure cannot be valid unless it is reliable,
CONCLUSION
Reliability test, ensures that our measuring
instrument (questionnaire) really has the
ability to be consistent / steady in
measuring something even though it is
repeated many times on the same object.
A test that produces the same score for
someone repeating the test is said to
measure the characteristic reliably. Like If
someone took the test again, would he or
she get the same test score, or a much
THANK YOU

You might also like