ch 6 validity of instrument. outline of validity definition types of validity 1. construct validity...

30
Ch 6 Validity of Instrument

Upload: reginald-goodwin

Post on 26-Dec-2015

270 views

Category:

Documents


3 download

TRANSCRIPT

Page 1: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Ch 6 Validity of Instrument

Page 2: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Outline of Validity

• Definition

• Types of Validity

1. Construct Validity

2. Criterion-Related Validity

3. Content Validity

Page 3: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Validity Determinants

Content

ConstructCriterion

Page 4: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

• There are three major forms of validity: content, construct, and criterion. Students need to know that they are in control of the level of validity of their own measurement. • Content validity refers to the extent to which measurement scales provide

adequate coverage of the investigative questions. If the instrument contains a representative sample of the universe of subject matter of interest, then content validity is good. To evaluate content validity, one must first agree on what elements constitute adequate coverage. To determine content validity, one may use one’s own judgment and the judgment of a panel of experts.

• Criterion-related validity reflects the success of measures used for prediction or estimation. There are two types of criterion-related validity: concurrent and predictive. These differ only on the time perspective. An attitude scale that correctly forecasts the outcome of a purchase decision has predictive validity. An observational method that correctly categorizes families by current income class has concurrent validity. Criterion validity is discussed further on the following slide.

Page 5: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Construct validity is a measurement scale that demonstrates both convergent validity and discriminant validity. In attempting to evaluate construct validity, one considers both the theory and measurement instrument being used. For instance, suppose we wanted to measure the effect of trust in relationship marketing. We would begin by correlating results obtained from our measure with those obtained from an established measure of trust. To the extent that the results were correlated, we would have indications of convergent validity. We could then correlate our results with the results of known measures of similar, but different measures such as empathy and reciprocity. To the extent that the results are not correlated, we can say we have shown discriminant validity.

Page 6: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Definition of Validity

• The appropriateness of the inferences or interpretations of an assessment's results

Page 7: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

How should one select a criterion?

Page 8: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Validity Determinants

Content

Construct

Page 9: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

• Construct validity is a measurement scale that demonstrates both convergent validity and discriminant validity.

• In attempting to evaluate construct validity, one considers both the theory and measurement instrument being used.

• For instance, suppose we wanted to measure the effect of trust in relationship marketing. We would begin by correlating results obtained from our measure with those obtained from an established measure of trust. To the extent that the results were correlated, we would have indications of convergent validity. We could then correlate our results with the results of known measures of similar, but different measures such as empathy and reciprocity. To the extent that the results are not correlated, we can say we have shown discriminant validity.

• This example is expanded upon in the following slide.

Page 10: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Construct Related Validity

• Assessment scores are valid for making inferences regarding a psychological characteristic or theory of human behavior.

• Example: Matrix

Page 11: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Example: Matrix

Content

Journal

Content 1 Content 2 Content 3

Article 1

Article 2

Page 12: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Case example: Marketing

Contents

Journal

Price Product Promotion

Kotler, P. (1997). Marketing management: Analysis, planning and control (9th ed. ). Englewood Cliff, NJ: Prentice-Hall.

P 12 P 23 P 32

Louis, E. B., & David, L. K. (1974). Contemporary marketing. Hinsdale, IL: Dryden Press.

P 61

Page 13: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Construct Validity is determined through:

• Intervention Studies

--- Pre-test / post-test changes based on treatment

• Differential-Population Studies

--- Determine how different populations perform on measure

• Related-Measures Studies

--- Compares scores to other measures valid for measuring the same construct

• Literature Review

Page 14: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Increasing Construct Validity

New measure of trustNew measure of trust

Known measure of trustKnown measure of trust

EmpathyEmpathy

CredibilityCredibility

Page 15: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

• Again, for a measure to illustrate construct validity, it must have both convergent validity and discriminant validity.

• Continuing with the trust in relationship marketing example, to show convergent validity, we must show that our measure correlates with a known and trusted measure.

• To show discriminant validity, our measure must be able to discriminate from other similar, but different measures.

• This means we would begin by conducting a pilot test to gather data using both the new and known measures of trust and the similar variables such as empathy and credibility.

• We would then run a correlation analysis to determine if the measures are correlated.

• To the extent that the new and known measures of trust are correlated, we have demonstrated convergent validity.

• To the extent that the new trust measure and the measures of empathy and credibility are not correlated, we have shown discriminant validity.

Page 16: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

What is a content domain?

Page 17: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Content Validity

• Does the instrument items represent the defined domain of content?

1. Samples the content domain

2. Elicits student behavior that demonstrates content domain

Page 18: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Determining Content Validity

• During assessment development

1. Define the content domain

2. Define the components of the content domain

3. Write questions that reflect the content domain

4. Determine the weight of the components of the content domain

5. Stratified sample from content domain

Page 19: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Determining Content Validity (Continued)

• After instrument construction

1. Panel of experts examinee assessment to determine if content domain is adequately represented

2. Have panel members individually review each test item

3. Calculate the percentage of the experts who agree to the appropriateness of each question for the content domain (content and relevance)

4. Compute an overall index that reflects the test’s content coverage

Page 20: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Increasing Content Validity

ContentLiterature Search

Literature Search

Expert Interviews

Expert Interviews

Group Interviews

Group Interviews

Question DatabaseQuestion Database

Etc.Etc.

Page 21: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

• Content validity refers to the extent to which measurement scales provide adequate coverage of the investigative questions.

• If the instrument contains a representative sample of the universe of subject matter of interest, then content validity is good.

• To evaluate content validity, one must first agree on what elements constitute adequate coverage.

• To determine content validity, one may use one’s own judgment and the judgment of a panel of experts.

• Using the example of trust in relationship marketing, what would need to be included as measures of trust? Ask the students for their own ideas. To extend the questions included and to check for representativeness, students could check the literature on trust, conduct interviews with experts, conduct group interviews, check a database of questions, and so on.

Page 22: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Validity Determinants

Content

ConstructCriterion

Page 23: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

• Criterion-related validity reflects the success of measures used for prediction or estimation.

• There are two types of criterion-related validity: concurrent and predictive. These differ only on the time perspective.

• An attitude scale that correctly forecasts the outcome of a purchase decision has predictive validity.

• An observational method that correctly categorizes families by current income class has concurrent validity.

• Criterion validity is discussed further on the following slide.

Page 24: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Judging Criterion Validity

RelevanceRelevance

Freedom from biasFreedom from bias

ReliabilityReliability

AvailabilityAvailability

Criterion

Page 25: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

• The researcher must ensure that the validity criterion used is itself valid. Any criterion measure must be judged in terms of the four qualities named in the slide.

• A criterion is relevant if it is defined and scored in the terms the researchers judge to be the proper measures.

• Freedom from bias is attained when the criterion gives each unit of interest an opportunity to score well.

• A reliable criterion is stable or reproducible.

• Finally, the information specified by the criterion must be available.

Page 26: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Criterion-Related Validity

• Demonstrates that the test scores relate to a specific criteria of performance

• Two types:

--- Predictive Validity

--- Concurrent Validity

Page 27: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Predictive Criterion-Related Validity

• Assessment’s scores are valid for the prediction of future behavior.

--- Subjects take assessment

--- Researcher waits a significant amount of time, then collects data on the subjects performance on the specified criteria

Page 28: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Concurrent Criterion Related Validity

• Assessment’s scores are valid for indicating current behavior

--- Subjects take assessment

--- Subjects scores are compared to data reflecting their current proficiency of the specified criteria

--- The scores from both measures are correlated to determine if the assessment agrees with the

other measure.

Page 29: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

Understanding Validity and Reliability

Page 30: Ch 6 Validity of Instrument. Outline of Validity Definition Types of Validity 1. Construct Validity 2. Criterion-Related Validity 3. Content Validity

• Exhibit 11-6

• Exhibit 11-6 illustrates reliability and validity by using an archer’s bow and target as an analogy.

• High reliability means that repeated arrows shot from the same bow would hit the target in essentially the same place.

• If we had a bow with high validity as well, then every arrow would hit the bull’s eye.

• If reliability is low, arrows would be more scattered.

• High validity means that the bow would shoot true every time. It would not pull right or send an arrow careening into the woods.

• Arrows shot from a high-validity bow will be clustered around a central point even when they are dispersed by reduced reliability.