_Chapter 6 (1)w
_Chapter 6 (1)w
EVALUATING SELECTION
TECHNIQUES AND
DECISIONS
Test-Retest Reliability
A method, where each one of the several people takes the
same test twice. The scores from the first administration of
the test are correlated with scores from the second to
determine whether they are similar.
Temporal stability
Alternate-Forms
Reliability
The extent to which two forms of the same test
are similar. People are administered both forms
of the test; half of the sample first receive Form
A and the other half receive Form B.
Counterbalancing
Form stability
The extent to which the scores on two forms of a test
are similar.
Internal Reliability
Item stability
The extent to which responses to the same test
items are consistent.
Item homogeneity
Split-half method
A form of internal reliability in which the
consistency of item responses is determined by
comparing scores on half of the items with scores
on the other half of the items.
Internal Reliability
Spearman-Brown prophecy formula
Used to correct reliability coefficients resulting
from the split-half method.
Coefficient alpha
A statistic used to determine internal reliability of
tests that use interval or ratio scales.
Scorer reliability
The extent to which two people scoring a test agree on the
test score, or the extent to which a test is scored correctly.
Validity
Validity
Criterion validity
The extent to which a test score is related to some
measure of job performance.
Criterion
A measure of job performance, such as attendance,
productivity, or a supervisor rating.
Validity
Concurrent validity
A form of criterion validity that correlates test
scores with measures of job performance for
employees currently working for an organization.
Predictive validity
A form of criterion validity in which test scores of
applicants are compared at a later date with a
measure of job performance.
Validity
Restricted range
A narrow range of performance scores that makes it
difficult to obtain a significant validity coefficient
Synthetic validity
A form of validity generalization in which validity is inferred
on the basis of a match between job components and tests
previously found valid for those job components.
Validity
Construct validity
The extent to which a test actually measures the
construct that it purports to measure.
Known-group validity
Barnum statements
Statements, such as those used in astrological
forecasts, that are so general that they can be
true of almost anyone.
Finding Reliability and Validity Information
Computer-adaptive
testing (CAT)
A type of test taken on a computer in which the
computer adapts the difficulty level of questions
asked to the test taker’s success in answering
previous questions.
ESTABLISHING THE
USEFULNESS OF A
SELECTION DEVICE
Cost-efficiency
Taylor-Russell tables
A series of tables based on the selection ratio, base
rate, and test validity that yield information about the
percentage of future employees who will be
successful if a particular test is used.
Cost-efficiency
Selection ratio
A series of tables based on the selection ratio, base rate, and test
validity that yield information about the percentage of future
employees who will be successful if a particular test is used.
number hired
Selection ratio number=
number of applicants
Cost-efficiency
Base rate
Percentage of current employees who are considered successful.
1. Number of employees hired per year (n) – The total number of employees hired
for a job in one year.
2. Average tenure (t) – The average number of years employees stay in a job,
calculated by adding all employees' years of service and dividing by the total
number of employees.
3. Test validity (r) – A measure of how well a test predicts job performance, based on
research or past studies.
4. Standard deviation of performance in dollars (SDy) – The difference in
performance value between an average and a good worker, estimated as 40% of
the employee’s yearly salary.
5. Mean standardized predictor score (m) – The difference between the average
test scores of hired and non-hired applicants, divided by the overall test score
variation.
Proportion of correct decisions
Determining the Fairness of a Test
Once a test has been determined to be reliable and valid and to have
utility for an organization, the next step is to ensure that the test is fair
and unbiased.
Measurement bias
Group differences in test scores that are unrelated
to the construct being measured.
Adverse impact
An employment practice that results in members of a
protected class being negatively affected at a higher
rate than members of the majority class. Adverse impact
is usually determined by the fourfifths rule.
Predictive bias
A situation in which the predicted level of job
success falsely favors one group over another.
Multiple-hurdle approach