questionnaire design issues: asking the right questions

33
1 Questionnaire design issues: asking the right questions Jettie Hoonhout Philips Research, Eindhoven 28 July 2014 Tallinn Summerschool 2014

Upload: others

Post on 12-Jun-2022

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Questionnaire design issues: asking the right questions

1

Questionnaire design issues:

asking the right questions

Jettie HoonhoutPhilips Research, Eindhoven

28 July 2014

Tallinn Summerschool 2014

Page 2: Questionnaire design issues: asking the right questions

2

• Background in Human Factors Psychology

• University of Utrecht, University of Maastricht

• Philips Research Eindhoven since 2000

Page 3: Questionnaire design issues: asking the right questions

3

Overview

• Issues in the design of your own questionnaire / survey

• Issues in the use of existing scales / questionnaires

• Many aspects discussed also very relevant for other types of

conducting research (e.g. experiments / tests, interviewing)

• Slash upgrade existing questionnaire

Page 4: Questionnaire design issues: asking the right questions

4

Page 5: Questionnaire design issues: asking the right questions

5

Subjective assessment:

indirect “observation” techniques �

collecting information from users about their experiences and opinions �

the users of a system that is being studied act as measuring instruments

(but instruments might have imprecisions and give inaccurate readings…. ;

and users are no passive respondents but actively thinking and interpreting)

Page 6: Questionnaire design issues: asking the right questions

6

Subjective assessment formats

• Face-to-face interview (unstructured, semi-structured,

structured)

• Telephone interview

• Self-administered questionnaire (mail, website, ..)

• Questionnaires versus scales:

– Questionnaires to address specific research questions in the

context of a particular project (Use and appreciation of

device X in the context of project Y)

– Scales to conduct repeated measurements across different

studies to determine some psychological construct: e.g.

satisfaction, pleasure, engagement, presence

Page 7: Questionnaire design issues: asking the right questions

7

When (not) to use….

several performance indicators are better determined through observation (e.g. task completion, time on task, errors, use of support tools (manuals), effort…)

other performance indicators can not so easily be determined through observation (perceived ease of use, perceived comfort, attitudes, opinions, motivations,…)

Do not use a questionnaire if you do not know anything (yet) about the topic….

Page 8: Questionnaire design issues: asking the right questions

8

Questionnaire design requirements:

• Precision/sensitivity/discriminate: can tell meaningful and

important differences between different cases, situations,

persons

• Reliability : results are stable across multiple administrations of

the instrument

• Validity : The questionnaire measures what it says it is measuring

• Usability to respondent and researcher (designing a

questionnaire is like designing a product!)

Page 9: Questionnaire design issues: asking the right questions

9

Writing questions should not be the first thing on

your mind!

• Carefully consider research questions and objectives

• Carefully consider how to conduct the analysis

• And even before that: carefully consider if a questionnaire is the

right instrument, given project questions and objectives!

Page 10: Questionnaire design issues: asking the right questions

10

Questionnaire design considerations:

1. questionnaire purpose �� research aim

2. respondent selection (non-response)

3. question content

4. question wording

5. response format

6. question sequence

7. questionnaire length

8. questionnaire layout

9. piloting (= testing a questionnaire, e.g. talk aloud, cognitive interviewing)

Page 11: Questionnaire design issues: asking the right questions

11

Completing a questionnaire has both social and

cognitive components:

• social encounter, comparable to conversation, following rules for

conversation

• cognitive task for respondent, tapping into language

understanding, memory, reasoning, etc.

Page 12: Questionnaire design issues: asking the right questions

12

Question-answer process

question � 1. interpretation of question and

response options

2. retrieval of information from

memory

3. formulating the response, and

checking the response against

internal standard

4. matching response with answer

categories, editing and reporting

the answer

� answer

e.g., how do you rate the breakfast in our hotel?

Page 13: Questionnaire design issues: asking the right questions

13

Some potential pitfalls…

• selection of respondents

• failing memory of respondents

• effect of fatigue, feeling bored, time

pressure

• More anonymity compared to

interviewing (+ and -)

• social desirable replies

• always an answer: respondent does not

want to appear stupid and “please the

researcher” effect

Page 14: Questionnaire design issues: asking the right questions

14

Closed-ended questions:

advantages disadvantages

easy and quick to answer

answers across respondents easy to compare

analysis of answers easier

response choices make question clearer

easy to replicate study

can put ideas in respondent’s head

respondents with no opinion might answer anyway

respondents can feel constrained and frustrated

many choices can be confusing

cannot tell if respondent misinterpreted the question

fine distinctions may be lost

Page 15: Questionnaire design issues: asking the right questions

15

Open-ended questions:

advantages disadvantages

permit unlimited number of answers

respondents can qualify and clarify responses

can find the unanticipated

respondents give answers with different level of detail

answers can be irrelevant

inarticulate or forgetful respondents are at disadvantage

coding responses is subjective and tedious

requires more time and effort from the respondent

Page 16: Questionnaire design issues: asking the right questions

16

Some tips

• a questionnaire with a title is generally perceived to be more

credible than one without

• begin with a couple of non-threatening and interesting items

• leave ample white space; for comments, but it also makes the

questionnaire appear easier

• test online questionnaires in different browsers, different screen

sizes

• the affiliation of the researcher can make a world of difference

• resist the temptation to add more and more questions….

Page 17: Questionnaire design issues: asking the right questions

17

Finally…. Writing questions…

Page 18: Questionnaire design issues: asking the right questions

18

Question wording

• Use simple wording (but not too simple…)

• Be brief (although questions should not be too short – reduces

credibility)

• Be specific

• Avoid:

– Being condescending or talking down to respondent

– Using biased wording

– Using abbreviations or scientific jargon

Page 19: Questionnaire design issues: asking the right questions

19

• Questions should ask for only 1 piece of information, so avoid:

• Asking two questions at once

– "Do you buy frozen meat and frozen fish?" A "Yes" answer

can mean the respondent buys meat or fish or both.

• Asking questions that contain assumptions

– Did you like product X? (assumption that respondent has

bought at least once this product)

Page 20: Questionnaire design issues: asking the right questions

20

• A questionnaire gets people to express their feelings,

perceptions, behaviors, and experiences, both past & present;

but do not overask, and avoid:

– Asking hypothetical questions• If we are going to build in feature x, are you going to use it?

– Asking for solutions to complex issues, or issues about which

a respondent most likely never thought of before

Page 21: Questionnaire design issues: asking the right questions

21

Page 22: Questionnaire design issues: asking the right questions

22

• Question wording should ensure that every respondent will be answering the same thing, so avoid:

• Ambiguous wording or wording that might mean different things

to different respondents

• Using terms for which the definition can vary (If it is

unavoidable, provide the respondent with a definition.)

• Being ambiguous about the time period the respondent should

consider

• Asking complex questions (double-barreled)

Page 23: Questionnaire design issues: asking the right questions

23

On scale points (not at all 0---0---0---0 very much)

• A good question is also one that will elicit a range of responses.

• Two or three (realistic) options may be appropriate, but four, five

or six will usually produce a more interesting result, especially as

respondents may avoid the extremes.

• Better label the extreme responses in a 'mild' way for this reason.

Use 'poor' rather than ‘horrendous' and 'very good' rather than

'excellent'.

• Even / non-even number of points on a scale?

• 5, 7, more, …, ?

• And a nice little fact: Rensis Likert, the developer of the scale, pronounced his name

'lick-urt' with a short "i" sound. It has been claimed that Likert's name "is among the

most mispronounced in [the] field", as many people pronounce it with a diphthong "i"

sound ('lie-kurt').

Page 24: Questionnaire design issues: asking the right questions

24

On (re)usage of existing scales

Self assessment manikin, with 3 subscales, pleasure, arousal, dominance

Page 25: Questionnaire design issues: asking the right questions

25

Scales used in hair removal studies

• Focus on pain, with large parts of the scale referring to

severe levels of pain � not really the business we are in...

• Not full use of scale points

– Resulting in psychometric issues

• Pain is a psychological construct

– Not all pain is the same, depends on various factors,

such as, ‘purpose of application inflicting the pain’,

test conditions

A number of

examples of pain

scales, from well-

known validated

ones (e.g. VAS, top

right corner), to

more funny ones

(bottom ones)

Page 26: Questionnaire design issues: asking the right questions

26

Before adopting a scale, check a range of

properties (Green, Dunn & Hoonhout, 2008)

2) Scale validity

Aim of scale validity

The extent to which a scale measures what it intend to

measure

Pertinent questions regarding scale validity for practitioners when adopting a scale

2a) Construct validity 2b) Context 2c) Sample

3) Scale reliability

Aim of scale reliability

Ensure that the measure consistently reflects the

construct, not only internally (cronbach’s

alpha) but also over time (test-retest)

Pertinent questions regarding scale reliability for practitioners when adopting a scale

3a) Inter-item reliability

3b) Test-retest reliability

1) Construct definition

Aim of construct definition

This is arguably the most important element of scale development: good scale

items cannot be formulated without it. Determine what

you are and are not intending to measure.

Pertinent questions regarding construct definition for practitioners when adopting a scale

1a) Is the construct grounded in

theory?

1b) Clarity: need to know what you measure is not ambiguous and

confused, when it is you are unsure of

what you are measuring.

1c) Discriminating: you must know that

you are not measuring

something that you are not intending to

measure (confound).

Page 27: Questionnaire design issues: asking the right questions

27

Scale design requirements:

• Precision/sensitivity/discriminate

• Reliability : results are stable across multiple administrations of

the instrument

• Validity : The scale measures what it says it is measuring

• Usability to respondent and researcher (a scale is like a product!)

Page 28: Questionnaire design issues: asking the right questions

28

Validity of scales

• Refers to measuring the construct of interest, the whole

construct, and nothing but the construct

• But constructs often ill defined! So chicken-egg problem…

• Dutch Psychologist Piet Vroon: “IQ is what is measured by the

test”

Page 29: Questionnaire design issues: asking the right questions

29

Validity

Construct validity: “refers to measuring the construct of interest, the whole construct, and nothing but the construct”

– (Face validity – does it appear to do what it promises to do?)

– Content validity – representative set of items for domain of interest; does it cover all aspects of the construct

– Criterion validity – relation between scores and other, independent criterion (variables, e.g. particular behavior)

• Predictive validity

• Concurrent validity

– Convergent validity – correlation between this measure and e.g. other, alternative scales

Page 30: Questionnaire design issues: asking the right questions

30

Discussion

• Scales developed for certain settings might not be applicable in others (e.g. mood scales such as PANAS, POMS)

• Underlying construct quite often not fully clear, underlying theory still under development

• Published scales often claim to be valid, but might turn out not to be adequately tested…

• If internal consistency is high (cronbach’s alpha) certainly something is measured, but not yet clear what!!

• Apart from validity: norms….

Page 31: Questionnaire design issues: asking the right questions

31Philips Research, January 2014

Issues I often encounter

• Using a questionnaire when one should not

– E.g. because one is afraid to use interviews instead, or other

means

• Not resisting to ask many, too many questions

• Thinking that respondents are as much interested in your project

as you are

• Using lots and lots of words that are familiar to the researcher

• Using scales in a test with say just <20 participants

• Using scales in a test without anything to compare the results

with

• Not pilot testing a questionnaire

• No proper pre-investigation before developing the questionnaire

Page 32: Questionnaire design issues: asking the right questions

32

references:

• a classic: Oppenheim, A.N., 1996, reprinted 2008, Questionnaire design, Interviewing and Attitude

measurement, London: Continuum.

• food for thought on question wording: Schwarz, N., 1999, Self-reports. How the questions shape the answers,

American Psychologist, 54, 93-105

• Bradburn, N.M., Sudman, S., & Wansink, B. 2004. Asking questions: the definitive guide to questionnaire

design. San Fransisco, CA: Jossey-Bass.

• not a manual, but reports on studies into questionnaire design issues: Sudman, S., Bradburn, N.M., & Schwarz,

N. 1996. Thinking about answers. The application of cognitive processes to survey methodology. San Francisco:

Jossey-Bass Publisers

• Clark, L.A., Watson, D. (1995). Constructing validity: basic issues in objective scale development. Psychological

assessment, 7(3), 309-319.

• Good, solid, readable and compact (approx 90 pages): Spector, P.E. (1992). Summated rating scale

construction. An Introduction. Newbury Park: Sage.

• Compact, practical guide: Wilson, C., 2013, Credible Checklists and quality questionnaires. A user centered

Design Method. Waltham, WA: Morgan Kaufmann.

• no bedtime reading, but essential to understand the limitations of asking questions about what motivates,

drives people to engage in behavior: Nisbett, R.E. and Wilson, T., Telling more than we can know: verbal

reports on mental processes. Psychological Review, 1977, 84(3), 231-259.

Page 33: Questionnaire design issues: asking the right questions

33