how in-house usability testing increases product development1368133/fulltext01.pdf · how in-house...

15
IN DEGREE PROJECT COMPUTER SCIENCE AND ENGINEERING, SECOND CYCLE, 30 CREDITS , STOCKHOLM SWEDEN 2019 How In-House Usability Testing Increases Product Development JONATAN SEGER KTH ROYAL INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE

Upload: others

Post on 24-Jul-2020

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

IN DEGREE PROJECT COMPUTER SCIENCE AND ENGINEERING,SECOND CYCLE, 30 CREDITS

, STOCKHOLM SWEDEN 2019

How In-House Usability Testing Increases Product Development

JONATAN SEGER

KTH ROYAL INSTITUTE OF TECHNOLOGYSCHOOL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCE

Page 2: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

ABSTRACT Can the Product Development Process at Tcom be improved by moving usability testing in-house?

The interest in this question lies in the agile transformation Tcom initiated in early 2018. Once agile,

the company’s interests are efficiency and increasing quality. One way of increasing the quality of

their digital products is to improve the process behind it, where involving users in usability testing

should be central in an iterative development process. A qualitatively driven mixed methods research

approach facilitated the acquisition of the empirical data necessary to successfully answer the research

question. The culture of Tcom’s usability testing was collected via a survey. The survey was

complemented with the current testing process through four interviews and two observations of

usability tests, one internal and one external. Parameters between the two tests were compared and

weighted against each other. The main findings reveal that the internal process shows promising

characteristics. It is, up to, 1/10 as expensive per participant, has a quicker call-to-test-time, facilitates

transparency, and allows refinement of the process. With the main findings in hand, the conclusion

points to the fact that usability testing should be composed and conducted by Tcom itself. The greatest

challenge to such a change is how a usability perspective can be integrated into the Scrum

methodology.

Page 3: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

SAMMANFATTNING Kan produktutvecklingsprocessen på Tcom förbättras genom att flytta användbarhetstester internt?

Intresset i denna fråga ligger i den agila transformation som företaget påbörjade tidigt 2018.

Effektivitet och kvalitet är centrala förbättringspunkter med det agila arbetssättet. Ett sätt att öka

kvaliteten på deras digitala produkter är att förbättra processen bakom, där involveringen av användare

i användbarhetstester bör vara centralt i ett iterativt arbetsförlopp. En kvalitativt driven mixed methods

research metod tillämpades för att förvärva den empiriska data som krävdes för att framgångsrikt

besvara frågeställningen. Tcoms användbarhetstestkultur undersöktes via en enkät. Denna metod

kompletterades med 4 intervjuer och 2 observationer av användbarhetstest, ett externt och ett internt.

Parametrar från de två observationerna ställdes mot varandra. De centrala upptäckterna avslöjar att den

interna processen uppvisar lovande egenskaper. Den är upp till en tiondel så kostsam per deltagare, har

snabbare beställning-till-test-tid, tillåter transparens, och möjliggör förbättring av själva testprocessen.

Med hänsyn till huvudfynden kan vi konkludera att användbarhetstest bör utformas och utföras av

Tcom själva. Den största utmaningen till en sådan förändring ligger i hur ett sådant

användbarhetsperspektiv ska implementeras i det agila ramverket Scrum.

Page 4: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

How In-House Usability Testing Increases Product Development

Jonatan Seger

KTH, Royal Institute of Technology

Stockholm, Sweden

[email protected]

ABSTRACT

Can the Product Development Process at Tcom be

improved by moving usability testing in-house? The

interest in this question lies in the agile transformation

Tcom initiated in early 2018. Once agile, the company’s

interests are efficiency and increasing quality. One way of

increasing the quality of their digital products is to improve

the process behind it, where involving users in usability

testing should be central in an iterative development

process. A qualitatively driven mixed methods research

approach facilitated the acquisition of the empirical data

necessary to successfully answer the research question. The

culture of Tcom’s usability testing was collected via a

survey. The survey was complemented with the current

testing process through four interviews and two

observations of usability tests, one internal and one

external. Parameters between the two tests were compared

and weighted against each other. The main findings reveal

that the internal process shows promising characteristics. It

is, up to, 1/10 as expensive per participant, has a quicker

call-to-test-time, facilitates transparency, and allows

refinement of the process. With the main findings in hand,

the conclusion points to the fact that usability testing should

be composed and conducted by Tcom itself. The greatest

challenge to such a change is how a usability perspective

can be integrated into the Scrum methodology.

Author Keywords

Product Development Process; User-Centered Design;

Usability Testing; Agile Framework.

ACM Classification Keywords

H.5.m. Information interfaces and presentation (Human-

Computer Interaction).

INTRODUCTION Product validation at any company is an important

component [20] in the Product Development Process

(PDP). Uncertainties and irregularities affecting this

process have a tendency of decreasing consistency and

usability throughout a product, most notably in cases where

a product is continuously iterated upon. There are various

types of methods to assess quality in products.

A shift towards an agile working process is under

implementation at the company. Although many aspects of

an agile environment are in place, there are areas that need

further attention. Such as the lack of usability attention

when developing products. Where an insufficient product

validation, in other words testing, is what this report aims to

investigate. Testing with users, which according to User-

Centered Design (UCD) methodology [26] concludes an

iteration as well as providing insights for the following

iteration. This perspective puts the users in center by

involving them in the process. UCD methodology with a

focus on usability is applied in this report since the author

perceives it can enhance product quality and due to a trend

of usability as sales arguments [4]. UCD is not yet fully

established in the PDP at Tcom and we asked ourselves

how difficult it would be to integrate usability testing with

their existing processes and whether the end result would be

better if the testing was done in-house as compared to being

done by others afterwards. In particular, we wanted to hear

the attitudes of the employees towards such a change. Thus,

part of the study presented here aims at investigating the

company culture and responses towards integrating such a

testing phase. Interviews were conducted with project

managers in order to understand the current testing process.

This method was complimented by a survey, where Tcom

employees’ perception of testing and the testing process

were in focus. The results indicated the prerequisites to

scale up the testing process, which is the next step for Tcom

in their pursuit of improving their PDP.

Two usability tests were studied. The former being a first-

ever in-house usability test executed at Tcom. The purpose

of this test was to investigate both a newly developed

purchase flow and the effectiveness of having this form of

testing in-house. Both tests will be described in greater

detail later in the report. The second usability test was

carried out by an external agency called Usbil, which is a

market research company that operates business-to-

business. They conducted a usability test ordered by an

employer organization called Mploy, where the purpose

was to test a set of scenarios in their new website. Usbil is a

Tcom partner, where they perform various types of testing

for the company. I took part in both respective testing

processes, acting as an active observer. As an active

observer I focused mainly on the process itself, while also

contributing with comments and insights to the test results.

Paste the appropriate copyright/license statement here. ACM now supports

three different publication options:

• ACM copyright: ACM holds the copyright on the work. This is the

historical approach.

• License: The author(s) retain copyright, but ACM receives an

exclusive publication license.

• Open Access: The author(s) wish to pay for the work to be open

access. The additional fee must be paid to ACM.

This text field is large enough to hold the appropriate release statement

assuming it is single-spaced in Times New Roman 8-point font. Please do

not change or modify the size of this text box.

Each submission will be assigned a DOI string to be included here.

Page 5: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

Qualitative and quantitative data [13] was obtained through

the three earlier mentioned methods: online survey,

interview, and active observation. The questionnaire was

distributed in the company through two channels, for the

sole reason to increase the number of responses. The

interviews were held with four employees which resulted in

qualitative data regarding the current PDP. The two active

observations were included in order to procure hands-on

insight into the execution and post-process of usability

testing. In combination, these provided adequate data to

answer the research question:

- Can the PDP be improved by moving usability

testing in-house?

BACKGROUND

The background section presents the basics of UCD as well

as how UCD methods are applied in industry. Benefits and

different usability testing methods are furthermore

introduced. The section ends with further attention to the

two agile frameworks Scrum and U-Scrum.

UCD & frameworks The UCD process has a strong focus on involving the

human perspective in all its stages during its iterative

development cycle [26]. Which is essential for a computer-

based system to be used sufficiently by its users [10]. It puts

a central focus on insights guiding iterations until a desired

solution has been established. Insights are most often

generated from the likes of testing. UCD principles revolve

around: early focus on users and their tasks, evaluation and

measurement of product usage, and iterated design [26].

They can be decomposed further, but for the sake of this

report the key takeaway is that UCD is an ongoing process

with respect to users. Different methodologies are applied

by companies working with software development.

Whether it is Scrum [17], Kanban [12], RABBIT [26], U-

Scrum [22], or any other agile PDP the iterative design

process is to a great extent similar. All of them advocates

testing with users. But the testing method and execution

time of which differs. A plethora of testing standards can be

applied to accomplish the last stage of each iteration.

However, not all are suitable for every occasion and some

of the agile frameworks have not successfully integrated

HCD into the process [25]. Various parameters must be

considered in order to execute the most appropriate form of

testing.

Usability testing methods

When collecting input from (representative) users there is a

great number of methodologies to be used, to name a few:

A/B-testing [21], guerilla testing [16], usability testing [14],

and various statistical tools to map user behavior. Each of

these methods is acknowledged for its strengths and how

they suit different occasions, all will not be covered since

this work revolves around the process of usability testing.

Usability assessment, as defined by Murillo and colleagues

[14], is another way to distinguish usability testing from

other forms of testing. The assessment consists of two types

of methods: inspection methods and test methods. The

executor of the assessment decides which of the methods is

used. Inspection methods are performed by inspectors, most

often experts in the subject. Test methods, on the other

hand, engage representative users of the targeted customer

segment in a controlled environment, where usability

testing is a subcategory to these test methods. Depending on

which form of usability testing the observation differs, both

in terms of technology as well as the applied strategy of the

observers. Online or remote, usability testing utilizes the

opportunity of recording screen, video, and/or audio while

the user performs a test [1]. Face-to-face usability testing

has the same possibilities as well as the advantage of

interviewing the participant. A think-aloud method can

complement both types of usability testing [19].

Benefits of usability testing What makes the testing phase so important is the

opportunity to discover and map out errors and problems

perceived by the test participants [3]. Their input results in

insights which can pivot the following iteration, advocating

improvement. All errors cannot be acted upon, rather it

mainly depends on the magnitude and worth of fixing the

error, as Cockton argues in his research on Worth Centered

Design [2]. Nonetheless, usability testing provides

procedures to unveil problems that could harm products.

Usability testing is advised to execute in different stages

with different focus during the PDP [27] to ensure a high-

quality product. Most of the studied companies in [5]

perform usability testing in the stage before launch, to make

sure there are no vital problems. This approach may seem

as cost-efficient since the company only carries out one test.

However, since changes late in the PDP are costly, the

strategy is flawed. Last-minute fixes are also not a

guarantee that the product did in fact improve. What has

been suggested by the Nielsen Norman Group in their paper

[16], is how light-versions of usability testing can be

applied and generate important feedback with time- and

cost-efficient orientation. The main argument is that

usability testing can provide valuable insights without it

being conducted as a full-scale procedure.

Scrum The agile working method, mentioned earlier, is

furthermore implemented at Tcom utilizing the Scrum

methodology [6]. This framework is currently popular in

the software development world, where it has its

advantages of establishing a flexible and time-efficient

development process. The latter aspects of Scrum have

however not been fulfilled by a set of companies studied in

[5]. The problematics occur when the testing phase is

executed in the next coming sprint, and not in the current

one. This can result in insights from the test becoming

expired, and thus, are not relevant. The little attention to

usability in Scrum methodology is not surprising due to its

software development foci. The organization of Scrum

defines three roles: Product Owner (PO), Scrum team, and

Scrum master [4]. The PO is responsible for the product

Page 6: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

backlog and deliverables. The development team can

include workers from many disciplines, they produce

deliverables through working in sprints. The Scrum master

makes sure that the development team can complete its

deliverables. This is a brief general view on Scrum

organization.

U-Scrum

Traditional Scrum fails to adequately address the usability

needs of the user [23]. Mona Singh applies a usability

perspective on Scrum, where she argues how Scrum lacks

the focus and resources to develop usable software

products. U-Scrum is suggested to enhance usability as it

applies such expertise on a high level in the Scrum team

organization through having two POs. The traditional,

single PO is complemented with another PO focused on

usability and User Experience (UX). This extension of

Scrum has not been adequately studied in current research,

but its takeaways can yield value for future agile

development teams.

METHOD Mixed Methods Research (MMR) approach has evolved

among researchers in recent years [9]. This way of

combining multiple forms of data in one research project

provides a deep understanding of context, experience, and

complexity [7]. A qualitatively driven approach was applied

for this report in order to cope with the complex nature of

usability testing at Tcom. Where qualitative data helped

shape the online survey as well as complementing the

quantitative data for a deeper understanding. This research

method enables comprehensive data collection of both

qualitative and quantitative characteristics [8]. Whereas

using only one method would have either supported a

detailed understanding of a rather niched aspect, or a wide

perception on a general level [8]. Therefore, this method

illustrates how accountability, budget, and opinions

influence the company culture in terms of in-house usability

testing. The empirical data was furthermore used to detect

any deviation of an objective reality. Each method has its

strengths and weaknesses, which is why results of the

survey and the interviews were put in relation to one

another, i.e. two methods validated each other’s

truthfulness. To deal with systematic errors during face-to-

face interviews [22] for example, the result of the

interviews is compared with results from other methods to

ensure its veracity. The three following methods were

utilized: online survey, observation, and interview. The

headquarter was the appointed office for this study since it

runs Tcom’s digital products.

Survey An online survey allows for fast distribution to a wide

audience, which made it the best possible tool for the

quantitative data collection since the headquarters of Tcom

holds 2200 employees and consultants ranging across

multiple departments distributed throughout an 8-story

building. The survey had a two-sided purpose in this data

collection: provide insight into employees’ attitudes

towards usability testing and ensuring interviews are in line

with the objective reality. Attitudes serve as a fundamental

argument for changing the current usability testing process

since employees are the ones utilizing processes and tools

in the work their opinion matter, which is similar to the

fundamentals of UCD.

However, the questionnaire designed for this thesis had to

take a wide set of respondents into account which affected

the technical terms and structure used. Survey participants’

answers will constantly be influenced by the phrasing of

questions and associated answer alternatives [22]. This

influence is possible to avoid up to a certain extent by

carefully structuring the order of questions. Where the most

important questions are advised to ask in the middle of the

survey, since respondents need priming before encountering

more cognitive demanding questions. While having

important questions late in the structure are likely to be

influenced by respondent mood, such as tiredness. Unlike

an interview, which can be slightly modified toward the

respondent in question [18], a survey must be the same for

every response.

Abstract concepts and detailed niche questions have a

tendency of confusing the respondent making her unable to

comprehend what is asked for [18]. This, in turn, results in

answers that do not reflect the question and increasing the

measurement error. Non-response is also a consequence due

to complex questionnaires. That is why an easily grasped

design and user-friendly approach was applied. Complex

subjects should be addressed by decomposing questions

into multiple elements, this has been explored in the

MultiTrait-MultiMethod (MTMM) [22]. This methodology

was appointed in the survey used for this report. Concept-

by-postulation is defined, by Willem E. Saris and Irmtraud

N. Gallhoferas, as the complex subject being investigated

[22]. This is achieved through exploring concept-by-

intuition, which is the subject in question decomposed to a

number of sub-subjects. These sub-subjects are additionally

converted into multiple measurable questions, defined as

request-for-answer. Request-for-answer will furthermore be

the terminology used throughout this report.

The survey generated both qualitative and quantitative data,

with more focus on the quantitative part. It included 31

requests-for-an-answer with the following structure: 3

straight open-ended requests, 4 multiple choice closed-

ended requests with an open-ended ‘other’ option, 12 single

choice closed-ended requests with an open-ended ‘other’

option, 2 multiple choice closed-ended requests, 10

multiple choice closed-ended requests. Both nominal and

ordinal [22] response scales were utilized. Where 4- and 6-

point symmetrical ordinal scales were applied when

requesting ratings. Saris et. al. suggests odd-number scales

to provide a neutral option for the respondents. However,

this was not implemented because of a will to encourage

respondents to take a stand and thus activating the

participants’ system 2 [11]. The nominal scales allowed

Page 7: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

respondents to provide neutral answers, which were used in

the rest of the closed-ended requests.

The survey consisted of two separate flows: usability

testing and user testing. The requests were distributed over

6 sections, in the following order: a) respondent

background, b) experience with user/usability testing, c)

benefits of user/usability testing, d) user/usability testing

and Tcom, e) user/usability testing statements, f) Tcom in-

house user/usability testing. Section b) separated

respondents by knowledge to assign them into their

appropriate flow. Respondents familiar with usability

testing faced the term usability and were assigned to group

1, whereas the rest met the term user and were assigned to

group 2 or group 3. Group 2 had knowledge of user testing

but not of usability testing, and group 3 had knowledge of

neither. The data from the respective groups will be

analyzed and interpreted with respect to their size and

knowledge, i.e. group 1 is perceived as more significant

than group 3.

Together the 6 concepts-by-intuition generated company

attitudes towards testing, i.e. concept-by-postulation. The

online survey was distributed in two channels: Tcom

intranet and 90 posters put around the headquarter. The

posters were integrated with a link and a QR-code. The

intranet article generated 242 views whereas the Bitly-link

via the posters had 17 clicks. The posters around the

headquarters were not used for the response rate since one

could not identify how many of the roughly 2200 people in

the building noticed it. The QR-code had no tracker and

thus could not be used in calculations. Together the intranet

article and the Bitly-link served as the denominator for the

response rate of 29,3%. Although this response rate is

viewed as representative for the audience that accessed it.

An alternative response rate with responses / total number

of employees yields a less representative rate of 3,1%. Since

the survey targets opinions, the results can serve as a

guideline for Tcom employee’s attitudes, but not as

decisive.

Active observation of usability tests The two observations enabled an understanding of how

usability tests ordered by Tcom functions and how Tcom

could execute tests themselves. The two usability tests were

observed with a focus on the process. The following aspects

were studied: agenda, test environment, staff behavior,

participants, post-test effort, time, and expenses. Agenda

included features such as welcoming participants,

transcript, and test scenarios. Technology, furnishing, and

room layout were studied to understand which facilities

were available. The staff’s behavior was studied to

understand how the personnel acted throughout each test.

Whereas the participant gathering process revolved around

how attendants were chosen. The post-test effort category

focused on the efforts put in after the usability tests had

been conducted, where report compilation was of major

interest. Expenses for parts of the process and the whole

process formed the economical perspective. A time

perspective was used in three ways: time per test, time to

execute a usability test, and time until the report is

finalized.

Unstructured interviews Four unstructured interviews were conducted at the

headquarters with managers from different departments to

map the organization and the Tcom PDP. The first of which

was held with Product in order to understand how the PDP

works in practice, from idea to launch. Secondly, an

interview was conducted with Human Resources with the

aim of comprehending how the departments are distributed

and function alongside one another. The two described

interviews were conducted before the survey since that

qualitative data was a prerequisite for section a) and d) in

the survey. The third interview was held with an employee

from a department responsible for ordering usability tests

among other responsibilities. This interview provided an

understanding of how the testing process functions. Lastly,

the head of UX was interviewed to interpret how the testing

process functions. Additional interviews were not

conducted due to adequate data from the four interviews as

well as the restricted timeframe.

Except for having a main topic, unstructured interviews do

not follow a certain pattern [28]. Where the purpose is to

get as much qualitative data as possible through a

conversation with a user or stakeholder. Despite the number

of flaws with an unstructured interview [28] they serve as

an honest and investigative tool that results in an

understanding of opinions and experiences. A semi-

structured interview format would have resulted in more

structured in-depth data [15]. However, since it requires a

lot more effort to analyze the long transcripts it was not

assessed to fit the timeframe of the project.

RESULTS & ANALYSIS Data gathered with the three methods are organized

throughout the following section. Where it initiates with

presenting the survey respondent distribution followed by

the current usability testing process at Tcom. The latter is

complemented with relevant findings from the Usbil

observation. Whereas the Tcom in-house test presents how

a usability test could be executed. The section ends with

opinions towards usability testing and preferences regarding

a desired testing process.

Respondents 76 Tcom employees responded to the survey out of which

69 were located at the office of interest. The product

development is performed at this office and that is why the

7 employees from other offices were excluded. 47, or

68,1% of the respondents had previous experience from

usability testing and were assigned to group 1. Group 2

made up 13,0% of the respondents who had experience of

user testing but not of usability testing. Group 3 (18,8%)

had no experience with any testing methodology. The most

represented departments: IT Development (14%), IT User

Page 8: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

Experience (14%), Network (13%), Product (12%), and IT

System Management (9%). Group and department

distribution are visualized in figure 1. Abbreviations used

for the departments are listed in table 1.

Figure 1. Department distribution of respondents.

Department Abbreviation Information Technology IT IT - Development IT D IT - QA IT QA IT - System Management IT SM IT - Architecture IT A IT - UX IT UX Customer Service CS Quality and Processes for CS CS Q Customer Relationship Management CRM Sales Sa Product Pr Communication Comm Business Intelligence BI Insight In Commercial Cml Marketing Ma Finance Fi Legal/Security L/S Business Transformation and Integration BT&I Human Resources HR Internet of Things IoT Technology Te Network Ne Customer Implementation CI Project Management PM

Table 1. Abbreviations of Tcom departments.

Current process

Interview The Tcom department structure is complex since it utilizes

the agile framework Scrum in parallel with a traditional

structure. The UX department is a good example of a

‘floating’ department since it functions dynamically in the

sense that its resources are applied in teams when

necessary. It is difficult to categorize all 2200 employee’s

positions in the building because not all departments work

agile. It also depends on how one defines the departments.

Which is further complicated by a recent company

reorganization. Scrum explains how the development teams

are organized but not the more traditional departments.

Although there is not one coherent way to successfully

divide all departments, the attempt used showed positive

results in the analysis of the quantitative data.

Currently, Tcom orders between two and three usability

tests each year from Usbil. The order-procedure is hugely

dependent on which employee(s) are involved in the

project. Hence, there is no established process for how to

request a usability test. The agile transformation which

Tcom initiated in early 2018 has been carried out in less

than one and a half years. The transformation has been

distressed by a reorganization in the company. The agile

environment is not completely fulfilled, which is

highlighted by gaps in the PDP. How to order a usability

test is one process that has been affected by this. The

process is described as ‘unclear’ and serves a lot of

competition from other processes in terms of budget and

attention. The result of this flaw is that usability testing

occurs rarely and late in the PDP.

When Tcom orders a usability test from Usbil, which is

their only partner in terms of testing, they regularly need

two weeks of preparation until the test can be conducted.

However, their fastest order-to-test saw a usability test

executed within a week from the date of the request. The

post-test effort from Usbil consists of a report that usually is

completed a week after the test. The total cost for these tests

is between 60-80 KSEK, most commonly 80 KSEK. The

expenses are bound to number of participants, participant

characteristics, post-test effort inter alia.

The absence of in-house usability testing at Tcom is mainly

due to budget restrictions according to the head of UX. The

resources are not enough at the moment to conduct more

usability testing. Tcom utilizes the following procedures of

UCD: personas (epics in Scrum), user journeys, expert

review, sporadic guerilla testing, and rare usability testing.

Guerilla testing and expert review are perceived as more

time- and cost-efficient than of usability testing.

Survey An average of 3,07 (out of 6) reveal that employees partly

disagrees with Tcom performing testing properly (Fig. 2).

26% of group 1 and 22% of the total respondents disagrees

(2) with the testing process at Tcom being performed

properly. Only 26% agrees (4 to 6) to some extent the

testing process is properly executed, whereas 74% rates the

process on the disagreeing part of the scale (1 to 3).

Page 9: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

Figure 2. Tcom performs testing properly.

25% of the total respondents believe Tcom performs their

own testing. As mentioned above, this is not the case as

current external partner Usbil conducts testing for Tcom.

16% got the correct answer “no”, whereas 43% had no

opinion and 16% answered “I don't know”.

54% assess prototyping as the most important step in the

PDP (Fig. 3). 10 respondents had the opinion that testing

should be performed during each step in the PDP.

Figure 3. Which stage Tcom should test during the PDP.

Usbil observation The usability test, ordered by Mploy and held by Usbil, was

split into two days. Day one focused on the testing itself

whereas the second day worked through the notes during a

workshop. Usbil aimed for 5 tests having 7 participants

scheduled. One participant per hour was the predefined

testing rate, where each test was scheduled for the duration

of 45 minutes. 15 minutes in between the tests was reserved

for discussion within the crew regarding the participant, the

results, and any other comments.

The test room in which the moderator and test subject were

located in were equipped with microphones, video cameras,

a 4*2-meter glass-mirror, and an iPhone 6 used to run the

prototype. The smartphone-screen was live-streamed to a

monitor in the room behind the mirror where the rest of the

crew were located. The crew consisted of 4 Mploy

employees, me, and an Usbil employee. The moderator was

furthermore also employed by Usbil.

Each participant was welcomed, and a small chit-chat took

place. The participants were encouraged to present

themselves with name, age, work, and work role. Their

general knowledge of the topic in question was discussed

followed by specific Mploy questions. Each participant was

instructed to perform 4 scenarios. 5 minutes were reserved

at the end of the test for any further questions from the

crew. The test ended with the participants receiving their

compensation.

The crew had responsibility to catch everything of interest

on post-it-notes. They were furthermore put under either of

the four following categories: pain points, gain points,

about the participant, and other. The post-it notes were re-

categorized the next morning by the Usbil crew before the

workshop officially started. Sub-categories filled with

unique post-it notes were the major changes made.

The 3-hour workshop was structured as follows: refine

problem definition, construct a parking lot (thoughts that

might be paid attention), list quick-fixes, discuss actions

(main part of the workshop), and prioritize the actions. The

remaining effort of assembling a report is made by Usbil.

The average time for the report to arrive is 5 business days.

Mploy had stated certain requirements for the participants

that were recruited by Usbil. The more preferences, the

longer time Usbil need to acquire a minimum number of

participants.

Total cost for the 2-day usability testing session was 80

KSEK. Divided on 5 participants, the cost ended up being

16 KSEK per participant.

Tcom in-house usability test observation The first ever in-house usability test at Tcom was scheduled

during a full day, where a maximum of 10 participants

would attend, 9 showed up. Each test had a duration of 30

minutes with an additional 15 minutes reserved in between

tests. When recruiting participants, Tcom reasoned that they

aimed for at least 6 tests in total and thus booking more

than needed.

The testing took place at the company headquarters. A

laptop with the desktop adapted prototype was located on a

table, with a Tobii eye tracker mounted below the screen.

Video, audio, and the eye tracking movement were all live

streamed to another room, where the crew was located to

observe and analyze.

The moderator initiated the interview with a quick session

of friendly conversation before presenting the agenda of

their upcoming 30 minutes. Once the participant had agreed

to be recorded, the moderator lead the participant into the

subject by asking a few test-oriented questions about the

participants’ situation. The moderator put emphasis on the

fact that the participant is not being tested, it is the

prototype. The test consisted of three scenarios.

The crew in the observer-room thoroughly studied the eye-

movements, comments, and body language of the

Page 10: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

participants on a 42-inch screen. The crew consisted of a

floating number of employees, at least three employees

were present for each of the tests. Notes were generated

based on anything which the crew felt appropriate. Which

was shared and discussed in the room after each test, with

the moderator present. Thoughts, tips, and manuscript

comments were continuously shared to iterate and improve

the test.

Once the 9 tests were completed, the following procedure

took place: combine and summarize notes, draw

conclusions, create a report. Dropbox Paper was utilized

throughout this process, where we were 4 persons in total

that contributed to the post-test effort. The report was

compiled after three business days.

Tcom appointed Rcru to recruit participants for the usability

test. Specific requirements can be accomplished by the

recruiting company. In this case, Tcom desired participants

based on two requirements. The more requirements, the

longer time Rcru need.

Total expenses for the Tcom usability test was 15,5 KSEK.

The expenses include recruitment costs and compensation.

Cost per participant ends in 1’722 SEK.

Tcom employees’ opinions towards testing 98,6% of the respondents believe testing is beneficial/partly

beneficial on a scale from 1 (not beneficial) to 4

(beneficial). Where 65,2% rated it as beneficial. The three

groups agree to a large extent about testing being beneficial,

where the exception is group 3 in which 46% consider

testing to be partly beneficial and 46% beneficial.

“Doing the right things” was a common response in the

open-ended feedback. Additional comments: “finding

errors”, “mitigate risk”, “to see if you are on track”,

“understand how the product is used”, “gives insights that

are impossible to predict”.

The primary benefit of testing stated by group 1 and group

2 is to study customer behavior, whereas group 3 ranks

error discovery as the primary asset of testing. Validating

assumptions is considered as the second most beneficial

attribute from the respondents familiar with usability testing

(group 1). This is however not shared by the other two

groups, where study customer behavior and error discovery

are those group’s top choices.

88,4% believes testing contributes to the quality of products

when ranked from 1 (nothing) to 6 (essential contribution),

shown in figure 4. Group 3 is slightly skeptical towards

rating the contribution higher than 4 (moderate

contribution). In average, group 2 were most optimistic

about testing contributing to product quality with an

average score of 5,1. Followed by group 1 with an average

of 4,9, whereas group 3 indicated more skepticism in

relation to the other groups with an average of 4,5.

Figure 4. Testing contributes to product quality.

Request-for-answers about continuous testing during PDP

resolved in an average score of 5,39. The scale in figure 5

was as follows: 1 (strongly disagree) to 6 (strongly agree).

Where 59% of the respondents strongly agreed continuous

testing should be part of the PDP. 5 (agree) and 4 (partly

agree) were the second and third most common answers

with 25% and 12% respectively. Group 3 returned the most

pessimistic result with an average score of 4,69.

Figure 5. Testing should be continuous during PDP.

Two-thirds of the respondents strongly agreed with testing

helps increasing product quality. An average of 5,51

revealed almost no negative opinions towards the relation

of testing and product quality. Group 1 indicated the highest

average of 5,62, whereas group 3 showed a more

discouraging relation with an average of 5,08.

The importance of testing saw respondents across the three

groups agree, indicated by an average of 5,41. 62% strongly

agreed (6), 20% agreed (5), and 13% partly agreed (4) with

testing being important.

Tcom employees testing process preferences There is nothing stopping Tcom from introducing a test-day

every other week. Where a strong will amongst the

employees is to move usability testing in-house. There are

however a few unresolved aspects that obstruct in-house

testing: ownership, budget, time, participant gathering,

compensation, and administration. Many of which are

admittedly plausible to resolve, where the most significant

facet, ownership, is where the many discussions have come

to a halt. Expertise is not considered as an issue since UX

and Insight both obtain the skills needed.

Page 11: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

There are aspirations of having an in-house testing process

at Tcom where an employee responsible for the whole

testing process is suggested. Consistency is a concern if an

in-house process would be realized.

Tcom was rated as the number one company to perform

testing for its products followed by Tcom’s current external

partner Usbil. Which received nearly half the votes in

relation to Tcom. Other external partner showed a similar

result as Usbil. 62% of the respondents belonging to group

3 opted no opinion, which is higher than for the other two

groups.

One day is considered as 8 hours. The average testing time

per week lies somewhere between ½ day and 1 day (Fig. 6).

4 respondents stated that the testing time per week

“depends on the product” and 46% had no opinion in the

subject. Where no opinion distribution was as follows: 36%

of group 1, 78% of group 2, and 62% of group 3. A

majority stated that the time needed depends on the

complexity and the magnitude of the product in question.

Some of the respondents indicated that the time needed

should make up 20-25% in relation to the total time spent

on the product. “More tests - better products” was also

mentioned.

Figure 6. Average testing time per week.

61% of the respondents believe Tcom should perform their

own testing. Whereas 16% are against that idea and 23%

had no opinion. 5 answers were excluded due to

inconclusive format “Both”. 70% of the most

knowledgeable respondents (group 1) are of the opinion

that Tcom should perform testing themselves. Respondents

who answered yes (39) could answer the two follow-up

questions regarding who should perform and who should be

responsible for in-house testing at Tcom.

UX (31%) and Product (13%) are thought to be the two

departments that should be conducting in-house testing.

10% of the respondents had no opinion in the matter.

Figure 7. Which Tcom department to conduct in-house testing.

36% of Tcom employees believe Product should have the

responsibility of in-house testing. Insight (15%) and UX

(13%) are the two following options stated by the

employees.

Figure 8. Which Tcom department to be responsible for in-

house testing.

The primary benefit of in-house testing is not mutual

between the three groups. Opportunity to refine the process

was ranked as group 1’s first choice of benefit with 30%.

Whereas group 2 assessed more frequent testing and control

in the process as the two most important benefits with 33%

of the votes respectively. Group 3 considered cost-

efficiency as the greatest impact of in-house testing. 38% of

group 3 had no opinion in the subject. Three benefits stand

out when the primary benefit was requested: opportunity to

refine the process (1), control in the process (2), and cost-

efficiency (3) (Fig. 9).

Page 12: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

Figure 9. Top 5 benefits with in-house testing at Tcom.

All groups agreed bias may influence the results is the main

drawback of in-house testing (42%) if we ignore the 46% of

group 3 which had no opinion. “Demands too much time”

is the second rated primary drawback with a total of 28%.

Only a few (6%) assessed no drawbacks of in-house testing.

Figure 10. Top 5 drawbacks with in-house testing at Tcom.

DISCUSSION

The discussion revolves around the findings in result &

analysis and is structured in a similar way. It initiates with

the current testing process, followed by Tcom employees’

opinions towards usability testing. Then follows a

comparison between the two observed usability tests. The

section ends with employees’ preferences on the process of

usability testing.

Current testing process Ownership, budget, time, and administration are the main

factors restricting an in-house usability testing process.

Although most of the employees believe an in-house

process is a more efficient solution, the process falls in

between departments as it is structured now. Insight lacks

the resources to be responsible and to carry out the testing.

Although UX employees have the mandate to request

usability tests, the current go-to for executing a usability

test demands a lot from the requester as they are required to

lead the whole process which results in few tests.

A clear majority of the respondents (74%) believes Tcom’s

testing is inadequately performed. The few numbers of

annual usability tests, as well as a vague order-process,

have raised these opinions around the agile departments.

The prototyping stage is considered as the most vital stage

to test. When testing occurs at Tcom, it is during this stage.

It can be dangerous to stall tests until the prototyping stage

since decisions on the fundamental aspects of the project

have been made without solid validation. Hence, expensive

adjustments may have to be carried out in order to make

sure products are of the desired quality. Scrum

methodology is implemented to avoid these incidents, but

since Scrum mostly concerns the software development

rather than usability these incidents can still occur.

Tcom employees’ opinions towards usability testing Respondents in all three groups consider testing as

important, beneficial, and that it increases product quality.

Testing should be conducted between four and eight hours

per week according to employees. The answers indicate

differences in their perception of testing showing a pattern

descending with experience of user- and usability testing,

i.e. employees with much knowledge rates testing as more

beneficial and vice versa. Respondents in group 3 consider

a transparent process as the most beneficial aspect with in-

house testing, which could originate from a lack of insight

into Scrum teams progress.

Bias and expertise in an in-house process is a concern

among respondents. The expertise exists as observed in the

active observation of Tcom’s in-house usability test and

stated during two of the interviews. Skewed results from in-

house testing are however not believed to be a concern

according to an Insight manager. Some of the respondents

furthermore considered “the ones closest to the project” as

best suited to conduct testing. Insight can assist with

moderators for tests with no insight into the PDP and thus

providing an objective approach.

Comparison between in-house and Usbil test Tcom in-house tests show promising costs of 1/10 per

participant in relation to Usbil. Although this ratio does not

include expenses for Tcom employees time and efforts, the

price tag is still favorable. The execution time is not

advantageous compared to Usbil unless the process is

changed. Where a standing test day every other week could

be the solution if time-efficiency is to be enhanced in terms

of call-to-test-time. There is another aspect of time-

efficiency which revolves around the post-test effort and

go-to steps based on insights during the tests. Usbil requires

one week to compile a report, which is inefficient compared

to Tcom’s three business days. The report is furthermore

not as important as one might expect, the go-to steps are the

reason for usability tests to be conducted in the first place.

Which can be extracted directly from the tests and avoiding

a report, and thus reducing the time until changes in the

product are applied. Usbil’s usability test procedure is

however more organized than Tcom’s, as it follows a

certain framework, which is likely to generate more on-

point insights.

Tcom employees testing preferences Employees have expressed their willingness to have a

greater insight into the advancement of projects. Which can

Page 13: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

be realized via inviting squad members to observe tests as

well as through the possibility to stream in-house tests.

The benefits and gains that can be achieved when testing

early in the process cannot be pursued with the current

testing process due to scarce recognition of usability. An in-

house testing process is believed to proclaim more testing

during different stages in the PDP. Expenses can

furthermore be reduced if the project validates its

advancement from the get-go. Although a range of

limitations are holding in-house usability testing back,

employees have expressed its potential of improving the

PDP. Continuous usability testing has been requested for a

few years but has served resistance from the current

framework Scrum.

Nature of usability testing

The usability testing methodology studied in this work is

one of the many formats’ customer input can be obtained.

Various methodologies and technologies can be applied for

the sake of validating and studying prototypes and products.

The in-house testing process shows the most prominent

characteristics of the two, but there are further alternatives

who can bear an equal or perhaps better process. A

compromise between an in-house and external process is

one example, where the test is conducted through a near

collaboration with external and internal experts at an off-

site lab. Another procedure could be how a company

utilizes external experts in an in-house process. The main

argument here is that there are several ways in which

usability testing can be organized and that this work is just

one of the many ways’ usability testing can be conducted.

Like the data collection method used in this work (MMR),

companies do not solely rely on one product validation tool.

Usability tests are usually complemented with quantitative

statistics such as heat maps or drop-out rates. They can be

further complemented with A/B-testing, where two versions

of a product/prototype are released and compared against

each other. Customer feedback during the customer journey

is a voluntary instrument which can help companies in

understanding their products. Multiple tools will strengthen

the empirical data and thus increasing the accuracy of the

analysis.

U-Scrum

The usability focus of U-Scrum is thought to provide

fundamental aspects in the product lifecycle, foremost

during the early stages in order to understand the users and

their needs. Once the business case is mapped the need for

two POs could thus decrease and the organization is back to

having solely one PO. One PO would resolve in distinct

leadership whereas two of them could perhaps trigger a

stalemate situation where focus would turn to who is the

“true” leader instead of the methodology’s main purpose.

Such a situation is likely to harm the PDP. In contrast, two

POs have the opportunity to focus on different aspects

throughout the process resulting in efficient management.

The main concern against two POs is when they both are

involved in the same aspect with different opinions, who

will obey? This question is left unanswered in today's

research and will need further attention to be dissolved.

CONCLUSION Tcom initiated their implementation of Scrum in early

2018, where the agile working method is applied but has an

acknowledged flaw in terms of usability attention.

Employees perceive usability testing as scarcely performed.

Budget motives and ownership restrains testing due to other

processes being prioritized. As a result, usability testing

occurs twice a year on average. Tcom has little control in

the current testing process since Usbil takes care of the

process once the order is made, which is expressed as

opaque by some employees in the Scrum teams.

- Can the PDP be improved by moving usability

testing in-house?

In-house usability testing could conduce to more testing

since it is far cheaper. The expense per participant from the

in-house pilot test was roughly 1,7 KSEK, that is a close to

1/10 of Usbil’s cost per participant. Tcom will also have

total control in the process which enables iterative

improvement of the process itself. Where time-efficiency

shows great potential, both call-to-test-time and time spent

on post-test effort can serve enhancement. The ability to

provide a more transparent process is another advantage

with usability testing conducted in-house. Although bias

and expertise were stated as concerns of in-house testing,

Tcom currently holds the expertise to manage those

concerns. As a result, improving the usability testing

process will give greater insight into products and thus

advocating improvement of the PDP.

Another finding during this study was how the agile

framework Scrum deals with the concept of UCD and

usability. Personas, user journeys, and expert reviews are

utilized in Tcom’s PDP. Conclusions from those methods

are not validated through testing with users, where one must

question if Scrum is the most appropriate framework from a

usability perspective.

FUTURE WORK

In order to introduce an in-house usability testing at Tcom

there must be an executive decision made regarding which

department(s) should be involved in the process as well as

setting aside a budget. Exploring U-Scrum methodology is

possibly one solution for this issue. Once the responsibility

is resolved Tcom can initiate structuring the process and

investigating which partners to work with. The integration

of a usability testing process will presumably see a trial

period before it is fully implemented.

ACKNOWLEDGMENTS

I would like to thank Tcom for taking me in and giving me

the opportunity to be a part of their team and allowing me

to explore their PDP. It has been a pleasure to have worked

side-by-side with such an enthusiastic and welcoming set of

Page 14: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

individuals. A special thanks go to my supervisor and my

stand-in-supervisor at Tcom.

REFERENCES

1. Albert, B., Tullis, T., & Tedesco, D. 2010. Beyond the

usability lab : conducting large-scale user experience

studies. Beyond the Usability Lab.

2. Cockton, G. 2006. Designing worth is worth designing.

In Proceedings of the 4th Nordic conference on

Human-computer interaction: changing roles (pp. 165-

174). ACM.

3. Davis, K., & Horst, D. 2012. The note-taker’s

perspective during usability testing: A hands-on

approach to recognizing what’s important, what isn’t.

In Proceedings of the Human Factors and Ergonomics

Society. pp. 2412–2416.

4. Düchting M., Zimmermann D., Nebe K., 2007.

Incorporating User Centered Requirement Engineering

into Agile Software Development. In: Jacko J.A. (eds)

Human-Computer Interaction. Interaction Design and

Usability. HCI 2007. Lecture Notes in Computer

Science, vol 4550. Springer, Berlin, Heidelberg.

5. Eloranta, V.-P. et al. 2013. Scrum Anti-Patterns -- An

Empirical Study. 2013 20th Asia-Pacific Software

Engineering Conference (APSEC), 1, pp.503–510.

6. Gonçalves, L. 2018. Scrum. Controlling &

Management Review, 62(4), pp.40–42.

7. Halcomb, E. & Hickman, L. 2015. Mixed methods

research. Nursing standard (Royal College of Nursing

(Great Britain): 1987), 29(32), pp.41–47.

8. Hall, Jori N. & Ryan, Katherine E. 2011. Educational

Accountability: A Qualitatively Driven Mixed-

Methods Approach. Qualitative Inquiry, 17(1), pp.105–

115.

9. Hunter, B. 2009. Advances in Mixed Methods

Research: Theories and Applications. Drug and

Alcohol Review, 28(5), p.575.

10. Issa T., Isaias P. 2015. Usability and Human Computer

Interaction (HCI). In: Sustainable Design. Springer,

London.

11. Kahneman, D. 2011. Thinking, fast and slow 1. ed.,

New York: London: Farrar, Straus and Giroux; Allen

Lane.

12. Leopold, K., & Kaltenecker, S. 2015. Kanban Change

Leadership: Creating a Culture of Continuous

Improvement. Somerset: Wiley.

13. McLeod, S. A. Qualitative vs. quantitative research.

2017. Retrieved May 21, 2019 from

https://www.simplypsychology.org/qualitative-

quantitative.html

14. Murillo B., Vargas S., Moquillaza A., Fernández L.,

Paz F. 2017. Usability Testing as a Complement of

Heuristic Evaluation: A Case Study. In: Marcus A.,

Wang W. (eds) Design, User Experience, and

Usability: Theory, Methodology, and Management.

DUXU 2017. Lecture Notes in Computer Science, vol

10288. Springer, Cham.

15. Newcomer, K.E. 2015. Conducting Semi‐Structured

Interviews. In Handbook of Practical Program

Evaluation. Hoboken, NJ, USA: John Wiley & Sons,

Inc., pp. 492–505.

16. Nielsen, J. 1994. Guerrilla HCI: using discount

usability engineering to penetrate the intimidation

barrier. In Cost-justifying usability, Randolph G. Bias

and Deborah J. Mayhew (Eds.). Academic Press, Inc.,

Orlando, FL, USA 245-272.

17. Opelt, A. 2013. Agile contracts creating and managing

successful projects with Scrum, Hoboken, N.J.: Wiley.

18. Ornstein, M. 2013. A Companion to Survey Research,

London: SAGE Publications Ltd.

19. Paz, F. & Pow-Sang, J.A. 2015. Usability Evaluation

Methods for Software Development: A Systematic

Mapping Review. 2015 8th International Conference

on Advanced Software Engineering & Its Applications

(ASEA), pp.1–4.

20. Quiñonesa, D., Rusua, D. & Rusub, V. 2018. A

methodology to develop usability/user experience

heuristics.

21. Rubin, J., Chisnell, D., & Dawsonera. 2008. Handbook

of usability testing how to plan, design, and conduct

effective tests (2nd ed.). Indianapolis, IN: Wiley Pub.

22. Saris, W.E. & Gallhofer, I.N. 2014. Design,

Evaluation, and Analysis of Questionnaires for Survey

Research, 2nd Edition, John Wiley & Sons.

23. Singh, M. 2008. U-SCRUM: An agile methodology for

promoting usability. In Agile 2008 Conference (pp.

555-560). IEEE.

24. Siroker, D., Koomen, P., & Harshman, C. 2013. A/B

testing: The most powerful way to turn clicks into

customers.

25. Sohaib, O., & Khan, K. 2010. Integrating usability

engineering and agile software development: A

literature review. In 2010 international conference on

Computer design and applications (Vol. 2, pp. V2-32).

IEEE.

26. Still, B., & Crane, K. 2017. Fundamentals of User‐

Centered Design: A Practical Approach. Boca Raton,

FL: CRC Press.

27. Sy, D. 2007. Adapting usability investigations for agile

user-centered design. Journal of usability Studies, 2(3),

112-132.

28. Wilson, C. 2014. Interview techniques for UX

practitioners: A user-centered design method.

Page 15: How In-House Usability Testing Increases Product Development1368133/FULLTEXT01.pdf · How In-House Usability Testing Increases Product Development Jonatan Seger KTH, Royal Institute

www.kth.se

TRITA -EECS-EX-2019:561