developing a hiring system reliability of measurement

24
Developing a Hiring System Reliability of Measurement

Post on 20-Dec-2015

221 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Developing a Hiring System Reliability of Measurement

Developing a Hiring System

Reliability of Measurement

Page 2: Developing a Hiring System Reliability of Measurement

Key Measurement Issues

• Measurement is imperfect

• Reliability--how accurately do our measurements reflect the underlying attributes?

• Validity --how accurate are the inferences we draw from our measurements?– refers to the uses we make of the measurements

Page 3: Developing a Hiring System Reliability of Measurement

What is Reliability?

• The extent to which a measure is free of measurement error

• Obtained score =– True Score +– Random Error +– Constant Error

Page 4: Developing a Hiring System Reliability of Measurement

What is Reliability?

Reliability coefficient = % of obtained score due to true score– e.g., Performance measure with ryy = .60 is 60%

“accurate” in measuring differences in true performance

Different “types” of reliability reflect different sources of measurement error

Page 5: Developing a Hiring System Reliability of Measurement

Types of Reliability

• Test-retest Reliability– Assesses stability (over time/situations)

• Internal Consistency Reliability– Assesses consistency of content of measure

• Parallel Forms Reliability– Assesses equivalence of measures– Inter-rater reliability is special case

Page 6: Developing a Hiring System Reliability of Measurement

Developing a Hiring System

Validity of Measurement

Page 7: Developing a Hiring System Reliability of Measurement

What is Validity?

The accuracy of inferences drawn from scores on a measure

• Example: An employer uses an honesty test to hire employees. – The inference is that high scorers will be less

likely to steal. – Validation confirms this inference.

Page 8: Developing a Hiring System Reliability of Measurement

Validity vs. Reliability

• Reliability is a characteristic of the measure– Error in measurement– A measure either is or isn’t reliable

• Validity refers to the uses of the measures– Error in inferences drawn– May be valid for one purpose but not for

another

Page 9: Developing a Hiring System Reliability of Measurement

Validity and Job Relatedness

• Federal regulations require employer to document job-relatedness of selection procedures that have adverse impact

• Good practice also dictates that selection decisions should be job-related

• Validation is the typical way of documenting job relatedness

Page 10: Developing a Hiring System Reliability of Measurement

Methods of Validation

• Empirical: showing a statistical relationship between predictor scores and criterion scores– showing that high-scoring applicants are better

employees

• Content: showing a logical relationship between predictor content and job content– showing that the predictor measures the same

knowledge or skills that are required on the job

Page 11: Developing a Hiring System Reliability of Measurement

Methods of Validation

• Construct: developing a “theory” of why a predictor is job-relevant

• Validity Generalization: “Borrowing” the the results of empirical validation studies done on the same job in other organizations

Page 12: Developing a Hiring System Reliability of Measurement

Empirical Validation

• Concurrent Criterion-Related Validation–

• Predictive Criterion-Related Validation–

Page 13: Developing a Hiring System Reliability of Measurement

Concurrent Validation DesignTime Period 1

Test currentemployees

Measure employeeperformance

Validity?

Page 14: Developing a Hiring System Reliability of Measurement

Predictive Validation DesignTime Period 1 Time Period 2

Test applicantsHire

applicantsObtain criterion

measures

Validity?

Page 15: Developing a Hiring System Reliability of Measurement

Empirical Validation: Limitations

Page 16: Developing a Hiring System Reliability of Measurement

Content Validation

• Inference being tested is that the predictor samples actual job skills and knowledge– not that predictor scores predict job

performance

• Avoids the problems of empirical validation because no statistical relationship is tested– potentially useful for smaller employers

Page 17: Developing a Hiring System Reliability of Measurement

Content Validation: Limitations

Page 18: Developing a Hiring System Reliability of Measurement

Construct Validation

Making a persuasive argument that hiring tool is job-relevant

1. Why attribute is necessary– job & organizational analysis

2. Tool measures the attribute– existing data usually provided by developer of tool

Page 19: Developing a Hiring System Reliability of Measurement

Construct Validation Example

Validating FOCUS as measure of attention to detail (AD) for QC inspectors

• Develop rationale for importance of AD• Defend FOCUS as measure of AD

– Comparison of FOCUS scores with other AD tests

– Comparison of FOCUS and related tests

– Comparison of scores for people in jobs requiring high or low levels of AD

– Evidence of validity in similar jobs

Page 20: Developing a Hiring System Reliability of Measurement

Construct Validation Example

Validating an integrity (honesty) test• Develop rationale for importance of honesty• Defend test as measure of honesty

– Comparison of test scores with other honesty measures• Reference checks, polygraphs, other honesty tests

– Comparison of test scores with related tests– Comparison of scores for “honest” and “dishonest”

people– Evidence of validity in similar jobs

Page 21: Developing a Hiring System Reliability of Measurement

Validity Generalization

• Logic: A test that is valid in one situation should be valid in equivalent situations

• Fact: Validities differ across situations

• Why?

Page 22: Developing a Hiring System Reliability of Measurement

Validity Generalization

1. Situations require different attributes vs.

2. “Statistical artifacts”; differences in:• Sample sizes• Reliability of predictor and criterion measures• Criterion contamination/deficiency• Restriction of range

Two possible explanations why validities differ across situations:

Page 23: Developing a Hiring System Reliability of Measurement

VG Implications

• Validities are larger and more consistent

• Validities are generalizable to comparable situations

• Tests that are valid for majority are usually valid for minority groups

• There is at least one valid test for all jobs

• It’s hard to show validity with small Ns

Page 24: Developing a Hiring System Reliability of Measurement

Validation: Summary

• Criterion-Related– Predictive– Concurrent

• Content

• Construct

• Validity Generalization

• “Face Validity”