0% found this document useful (0 votes)
9 views14 pages

Chp43 A Reference To UsabilityInspection Methods

The document discusses various usability inspection methods, highlighting their costs and effectiveness for indie developers. It categorizes these methods into formative and summative approaches, with formative methods being more cost-efficient but reliant on expert evaluators. The paper also identifies gaps in current research and emphasizes the need for further exploration into usability evaluation techniques.

Uploaded by

malikaqeel aqeel
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
9 views14 pages

Chp43 A Reference To UsabilityInspection Methods

The document discusses various usability inspection methods, highlighting their costs and effectiveness for indie developers. It categorizes these methods into formative and summative approaches, with formative methods being more cost-efficient but reliant on expert evaluators. The paper also identifies gaps in current research and emphasizes the need for further exploration into usability evaluation techniques.

Uploaded by

malikaqeel aqeel
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 14

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/314091950

A Reference to Usability Inspection Methods

Chapter · January 2015


DOI: 10.1007/978-981-287-332-3_43

CITATIONS READS

13 916

2 authors:

Lin Chou Cheng Muhizam Mustafa


Inland Revenue Authority of Singapore University of Science Malaysia
5 PUBLICATIONS 31 CITATIONS 33 PUBLICATIONS 135 CITATIONS

SEE PROFILE SEE PROFILE

All content following this page was uploaded by Lin Chou Cheng on 13 December 2020.

The user has requested enhancement of the downloaded file.


Chapter 43
A Reference to Usability Inspection Methods

Lin Chou Cheng and Muhizam Mustafa

Abstract Engineering usability in software application is never cheap so as the


methods to evaluate it. Although there already exist several inspection methods that
are touted to be cost-discounting, they are still too hefty to be borne and applied by
most indie developers. This paper will discuss the various methods of usability
inspections and will provide an overview on the key variables that affect the costing
of usability evaluation while discussing the mechanics of the formative inspection
methods in depth. Lastly, the paper will highlight the gaps and opportunities for
more research into the usability inspection methods.

Keywords Usability evaluation methods • Cost • Task analysis • Cognitive


walk-through • Heuristic evaluation

43.1 Introduction

Usability evaluation is a form of user context analysis, which mainly draws from the
direct observations of users’ task performance when they are interacting with an
application. The evaluation can be a costly quantitative experiment or a formative
qualitative study which comprises both large and small sample sizes [1, 2]. The
methods to evaluate usability are usually a costly construct which would need a
considerable amount of resources to administer [3]. The referring costs in usability
are the total cost that has been spent on an evaluation cycle, where the definition of
costs is based on man-hours (time) and the value of money that has been used dur-
ing the evaluation activities. According to Nielsen, on average, the costing of usabil-
ity would siphon an additional 8–13 % of a project’s total budget [4]. Although the
cost for usability testing does not increase linearly with project size, Nielsen advo-
cated that it would be best to devote additional 10 % of a project’s total budget for

L.C. Cheng (*)


Game Design & Development, School of Informatics & IT, Temasek Polytechnic,
21 Tampines Avenue 1, 529757 Singapore, Singapore
e-mail: [email protected]
M. Mustafa
School of the Arts, Universiti Sains Malaysia, Gelugor, Penang 11800, Malaysia

© Springer Science+Business Media Singapore 2015 407


O.H. Hassan et al. (eds.), International Colloquium of Art and Design
Education Research (i-CADER 2014), DOI 10.1007/978-981-287-332-3_43
408 L.C. Cheng and M. Mustafa

usability testing, as building usability is an iterative process that required recurring


spending.
The spending on usability largely goes to a series of evaluation activities, which
include planning of evaluation process, creating test tasks, recruiting participants
and evaluators, and analyzing the data before preparing the recommendation report
to revise the application. The entire evaluation activities would involve great amount
of man-hour. For instance, based on a documented experiment by the Technical
University of Denmark, the average time spent for evaluating the usability of a web-
site was 39 h [5], with fifty (50) teams of user interface design students as the evalu-
ators. Prior to the experiment, the students underwent 15 h of additional training in
user test methodology. This would equate to a total of 6.75 workdays if both the
evaluation time and training hours were to be combined and calculated. These man-
hours are an upper estimate of the required time for a first run of usability test, and
the investment of time can be reduced to two (2) workdays if the evaluation uses
discounted approach with experienced evaluators [5]. According to Nielsen [6], a
usability testing staff with 5 years of experience would cost as much as USD 84,000
per annum to hire. Although the salaries of usability testing practitioners are lower
outside the United States, it would not be much lower than other regional standards
of an IT professional’s earning. To conduct usability evaluation, one can either
choose to self-manage the evaluation or outsource it to a usability consulting firm.
However, managing an internal usability team on short-term basis can be very
costly, as building a two-room usability lab furnished with one-way mirror and test-
ing equipment can cost as much as USD 100,000 in the early 1994, and the compa-
rable quotes to build usability lab today are in the USD 25,000 range [7]. This is yet
to include the cost of hiring usability personnel. On the other hand, the outsourcing
solution for usability evaluation is not any cheaper. For instance, one of the leading
usability consulting firms – the Nielsen Norman Group – has their services priced
between USD 10,000 and USD 150,000 per project (see Fig. 43.1) [8]. With refer-
ence to the services offered by [8], pricing is not the benchmark to determine the
effectiveness of a method. Usability evaluation methods (UEMs) are just a collec-
tive set of evaluation techniques designed to audit the usability of a user interface.
In a tight development framework, the best applicable method to evaluate usability
would probably be the quickest and least expensive method that meets the devel-
oper’s timeline and budget for their product.

Type of Usability Evaluations Price (USD)


Qualitative Usability Tests $20,000- 40,000
Formative
Iterative Design Usability Tests $40,000- 70,000
Competitive Benchmarking $50,000
Quantitative Tests $70,000
Summative Remote Usability Testing $10,000- 70,000
International Tests $50,000- 150,000

Fig 43.1 Pricing for usability evaluations


43 A Reference to Usability Inspection Methods 409

From the price list above, the offered five evaluation services can be broadly clas-
sified into two distinct categories: the formative and summative methods. These cat-
egorical terms of formative and summative originated from assessment design in
education where they are used to classify the approach that set to evaluate students’
performance [9]. In education, formative methods use interim results to inform one’s
learning with immediate feedback for self-improvement, whereas summative meth-
ods use a standardized test to summarize the performance of a population. The
nature of assessment design in education is similar to usability evaluation. The only
difference is that the test subject in usability is an application and not students or the
users. All the expenditures in usability evaluation are basically a floating operational
cost that scales according to the selected type of evaluation methods. By observa-
tion, the costing in usability evaluation is largely subjected to the mechanics of the
method itself. For instance, based on the sample price list in [10] (see Fig. 43.1), the
median cost for both the summative and formative UEMs is at the distinct price mark
of USD 60,000 and USD 40,000, respectively. All summative testing is more costly
compared to the formative one, as it required a large quantity of samples to conclude
its finding through discrete statistical distributions [10]. A typical summative usabil-
ity test would be a one-to-one session that involves two (2) hired individuals: a
moderator and a recruited tester (user). Each of the test sessions will have different
testers and the same pool of moderators to moderate the test for at least 30–50 ses-
sions. In its formal procedure, the usability test would continue to be run even the
findings are about the same after several rounds of initial testing. As such, the spend-
ing on such mode of recurring testing would have set a total cost of $30,000 if each
of the test sessions is worth $1,000 and is being rendered for thirty (30) times.
The formative UEMs on the other hand have much lower overhead cost as com-
pared to the summative UEMs. This is because the findings of a standard formative
evaluation can be concluded among three to five different experienced evaluators.
Hence, the formative UEMs are also best known as discounted inspection methods
[11]. Although formative UEMs seem to be more cost-efficient as compared to the
summative test methods, the challenge lies in hiring usability experts, as there are
no standard ways to qualify such expertise and it would not be cheap to engage one
[7, 12, 13]. The objective of usability evaluation is to discover what is usable and
not usable in an application. The gathered insights aimed to aid the developers to
improve their design with better informed decisions, which would benefit their end
users. However, usability evaluation is still costly, be it summative or formative.

43.2 Taxonomy of Usability Evaluation Methods

Besides the formative and summative approaches, methods of usability evaluation


can be classified into the following streams of inspection and the test approach as
shown in Fig. 43.2 [14]. UEMs have long been used to examine the usability of vari-
ous types of user interface which range from electronic products to application soft-
ware that are still under development or are about to be released [15–17]. Being part
410 L.C. Cheng and M. Mustafa

Usability Evaluation Methods (UEMs)


Inspection Approach Test Approach
(Formative) (Summative)
Task Analysis Heuristic Thinking Questionnaire
Evaluation Aloud
Cognitive Walkthrough Field Observation

Fig. 43.2 Taxonomy of usability evaluation methods

of the formative techniques, the usability inspection approach is actually similar to


the process of quality control (QC), but instead of inspecting a product’s defects,
usability inspection is set to identify any potential user interface problems that
would hinder its users’ task performance.
The inspection approach can be divided into techniques like heuristic evaluation
(HE), cognitive walk-throughs (CW), and task analysis (TA). All these techniques
can be combined or applied separately to inspect the usability of an application. In
practice, the techniques of usability inspection are usually being carried out by two
to three evaluators who have the expertise in usability. As such, a typical inspection
process would not involve real users, as often the professional judgment by the
evaluators is regarded to be sufficient.
On the other hand, the test approach is a set of summative evaluation techniques
which depend on the real users to detect usability problems for an interface. The
basis of the test approach is that real users’ feedback is more contextualized and
specific, as compared to inspection approach that is based on the general inputs of a
selected few. However, the test approach would require high number of real users
for ensuring the reliability of the findings, as not all the collated users’ opinions or
complaints about an interface are reliable. Most UEMs were created back in the
early 1990s, which aimed to improve the usable qualities of desktop computing
systems at the time. Nevertheless, they are still being used by today’s usability pro-
fessionals. For instance, based on the biannual survey led by the Usability
Professionals’ Association (UPA) in 2009, a classic inspection technique such as
heuristic evaluations is still a favorable technique among many usability practitio-
ners [18]. In usability evaluation, costing is still a prevalent issue, and this could
explain why most summative testing methods have constantly lower popularity
[18]. By far, formative inspection techniques are the preferred option when there is
constraint in costing. Thus, in the following sections, this paper will focus on dis-
cussing usability inspection techniques (Table 43.1).

43.3 Usability Inspection Techniques

The inspection approach was first conceived from the informal auditing procedure
in software development for debugging a program’s source code during the mid-
1980s [19, 20]. The practice was then brought over into the development of
43 A Reference to Usability Inspection Methods 411

Table 43.1 Change in usability evaluation techniques used 2007–2009


Technique 2007 (%) 2009 (%)
Heuristic evaluation (expert review) 77 74
Task analysis 60 58
Usability testing (in a lab) 54 54
Usability testing (remote, moderated) 42 42
Usability testing (remote, un-moderated) – 18
Creating prototypes (wireframes or low fidelity) 73 69
Contextual inquiry/ethnography 46 42
Creating prototypes (high fidelity) 42 40
Benchmarking or competitive studies 44 40
Satisfaction surveys 35 34
Eye tracking 13 13

graphical user interface (GUI) and subsequently became an alternate solution for
usability testing, as real users are often expensive and difficult to recruit in sufficient
number for testing all aspects of an evolving design [21]. For instance, a typical
usability inspection process would only require three to five (3–5) evaluators,
whereas the procedure of usability testing would require at least 30 of real users and
a number of moderators in each of the test sessions. Prior to 1990, most usability
inspections relied on the general skill and experience of an evaluator without any
formalized procedures or guided inspection criteria [22]. Such mode of inspection
is usually questionable and plagued with validity issues, as every time a new evalu-
ator who evaluated the same interface would deliver a different scope of findings
from the previous one. Over the time, discussions on formalizing usability inspec-
tion approach were called to attention, and both the academics and industrial
researchers have discovered that the traditional scientific inquisition methods can
actually be fruitfully employed to inspect the usability of an interactive system with
much better validity [23]. Hence, from 1990 to 1993, there was an explosion of
interest in developing new evaluation methods which are built on the existing ones
[24]. For instance, UEMs like heuristic evaluation (HE) were an extension of the
previously informal inspection technique that is armed with a standardized usability
heuristics and checklist. In the following subsection, this paper will look into the
inspection techniques of task analysis, cognitive walk-through, and heuristic evalu-
ation, all of which have stood the test of time relatively well.

Task Analysis

Task analysis (TA) is a generic umbrella term of techniques used to investigate


users’ cognitive processes and physical actions when they are acting on a task [25].
TA is also known as action analysis, where the very definition of task is seen as a
412 L.C. Cheng and M. Mustafa

sequence of laborious actions with no internal control structure [14]. The focus of
TA as usability inspection technique is to identify and analyze users’ mental and
physical efforts in details, when they are interacting with a user interface. TA was
originally developed to improve occupational task performance of the labor force
[26], and it was only later being adapted into the field of human-computer interac-
tion when the industrial landscape has changed from analogue to digital technology.
The use of TA is to comprehend users’ task requirements, by breaking down an
observed task into its lowest level of acts, and then re-cluster them into plausible
scenarios where the users would perform in an actual course. For instance, consider-
ing the task of buying an e-book via online store [27], the digital task was observed
and deconstructed into five different sequential acts as follows:
1. Locate the book.
2. Add a book to shopping cart.
3. Enter payment details.
4. Complete billing and mailing address.
5. Confirm order.
Based on the preliminary task flow, the online book buying action can be orga-
nized into two plausible scenarios based on the predefined users’ experience profile
(see Fig. 43.3). By drawing from these two scenarios, an inspection process would
then follow suite to assess the usability of the e-commerce website, whether the
website’s interface can effectively and efficiently support the course of book buying
actions for both the new and experienced users.
The process of TA can be extended by using questionnaires or open-ended user
interviews to gather more detailed information about how people would actually
perform a specific type of task [28]. The questionnaires can be administered as a
face-to-face interview or an online survey for identifying users’ task requirements
[29]. The gathered requirements enable the application designers to have a deeper
insight into users’ task needs before actual development [25]. Overall, TA is a user-
centered inspection technique, where only one experienced practitioner is needed to
do the inspection. However, such user-centered inspection technique has its disad-
vantages, as it is too time consuming and skill dependent. In addition, TA has lim-
ited application, as it cannot model after complex users’ task [25]. Hence, this
explained why the usage of task analysis has fifty-eight percent (58 %) declining
popularity in the 2009 UPA’s survey; a time-consuming method is not favorable in
a tight developing environment.

New Users (Customers) Experience Users (Regular


Do: [1,2,3,4,5] Customers) Do: [1,2,5]
1. Locate the book;
2. Add a book to shopping cart; 1. Locate the book;
3. Enter payment details; 2. Add a book to shopping cart;
4. Complete billing and mailing address; 5. Confirm order.
5. Confirm order.

Fig. 43.3 Users’ task analysis for online book buying


43 A Reference to Usability Inspection Methods 413

Cognitive Walk-Through

Cognitive walk-through (CW) is another usability inspection technique that used to


predict usability problems without engaging any real users. The idea of CW origi-
nated from exploratory learning principles, where the method is used to study users’
mental processes, for how well they learn to solve a given directive by exploring the
limited yet salient options plotted by the designer [10, 30]. The referred directive is
simulated user instructions set in the format of human-computer dialogue, which
was later displayed for its evaluator to solve within the available system’s function-
alities [11]. CW is commonly used to assess whether an application’s interface has
enough instructional cues to support users’ mental state when they are attempting a
task, whereas TA focuses on whether an application has the right interfaces to
accommodate the required interaction to complete a task. For instance, the follow-
ing is a snippet of CW from [25], which is based on another similar online book
buying scenario (Fig. 43.4).
The process of CW usually involves a team of evaluators, comprises of applica-
tion designer and developers who would walk through the features set of their prod-
uct, and notes any usability issue that they may encounter, through a series of
reflective questions. The process of CW requires a fully functioning prototype for
the evaluators to effectively explore their own design from the end users’ perspec-
tive. The use of CW was originally developed for inspecting desktop systems, but it
can be adapted to evaluate web-based applications and other tangible handheld
devices [25]. The working mechanism for a successful CW is to document the tes-
ters’ account in each attempted task sequence, depicting what works and what does
not work. All documented usability issues would then be further analyzed before
classifying the severity of the problem. The strength of CW lies in its micro level of
inspection, by examining every assumptive step that the users might take when they
are interacting with the prototype. Such deep-level mode of inspection approach is
found to be invaluable when evaluating systems that are safety critical [25]. CW is
considered to be cost-effective as the development team can role-play themselves to
identify usability problems without hiring the real users. However, the

Task: Buy a copy of this eBook from this


Simulated
publisher. Target users: Students who connect the
User Task
internet regularly.

Directive Step 1. Selecting the correct category of goods


on the homepage.
Q: Will users know what to do?
Answer: Yes, they know that they must find books.
Reflective
Questions
Q: Will users see how to do it?
Answer: Yes, they have seen menus before and will
know to

Fig. 43.4 Cognitive walk-through test task


414 L.C. Cheng and M. Mustafa

noninvolvement of the real users or any other independent evaluators would create
the danger of inherent bias during the walk-through [14]. For instance, the applica-
tion’s designers who acted as the evaluators would tend to be more defensive when
potential usability problems with their design are being highlighted to them; this
would often evoke a long argument to justify their designs at the expenses of any
necessary fixes [31]. Like TA, CW is also found to be too time consuming and
tedious as the method needs to query and answer at every level of actions. The effec-
tiveness of CW is tied to its deep-level mode of slow data gathering process; hence,
the walk-through would need to be executed in a slower pace.

Heuristic Evaluation

Heuristic evaluation (HE) is a popular and widely used inspection technique pio-
neered by Jakob Nielsen and Rolf Molich during the early 1990s [32]. The inspec-
tion process of HE involves having usability specialists to evaluate a user interface
with a set of usability principles known as heuristics [14, 21]. HE is also known as
expert review, as the evaluators who review or inspect the user interface are usually
the product domain experts who know about usability requirement [7]. By far, HE
is the only inspection technique that uses usability experts to audit an interface’s
navigation structure, dialogue box, menu, etc., through a set of empirical or vali-
dated heuristics. The original mechanic of HE is to have a single expert evaluator
performing the inspection alone. However, the method was later revised by Nielsen
[33] to include a few more evaluators, to widen the scope of the inspection. The
rationale behind the refinement came from other Nielsen’s experiments when
inspecting a voice response system. In the experiment, Nielsen had asked nineteen
(19) expert evaluators to identify sixteen (16) usability problems, which were all
purportedly sowed into the voice response system before the actual experiment.
Interestingly, all the returned discoveries by the nineteen (19) expert evaluators
were more varied and more diverse than the predicted findings. Then, there came
the question about how many expert evaluators are needed for hosting a reliable
HE. The common connotation back then was that the more expert evaluators are on
the job, the higher the usability problem discovery rate would be. But to include
more evaluators, the purpose of HE for being a cost-effective inspection technique
would be defeated, as it would not be any different from summative UEMs. In a real
development, all usability studies should use as few evaluators as possible when
resources are tight.
As a result, both Nielsen and Landauer come up with a predictive model and set
the optimum cost-benefit ratio of HE, which is not to have more than five evaluators
per testing [34]. This is because the findings after the fifth evaluator would be
repeatedly the same, as 85 % of the usability problems would have been identified
by the previous five evaluators. To explain their substantial discovery, both Nielsen
and Landauer used the following nomograph to address the cost-effectiveness by
having only five evaluators (Fig. 43.5).
43 A Reference to Usability Inspection Methods 415

Fig. 43.5 Proportion of usability problems with numbers of evaluators

For instance, the nomograph showed there will be zero usability problems if
there is zero evaluator in action. As soon as there are two evaluators entering the
evaluation, the usability problems’ discovery rate would jump astoundingly to fifty
percent (50 %). Subsequently, there will be lesser and lesser new usability problems
to be uncovered after the fifth evaluator as most of the obvious problems would have
been pointed out by the first five evaluators. This influential finding in [34] by
Nielsen and Landauer has also indirectly asserted one of the earlier hypothetical
claims by Virzi in [35], where the act of “observing additional participants will
reveal fewer and fewer new usability problems.” The findings by these researchers
were also repeatedly confirmed in the published findings by Lewis [36] who stated
that the law of diminishing returns applies to HE, as discovery rate of new usability
problems would begin to stay stagnant and incrementally diminish after the eighth
evaluator. In other words, it is not economical to repeat the inspection process for
the same interface for more than eight times, as the subsequent findings would deem
to be redundant. In [36], Lewis also pointed out that it would be highly unrealistic
for most usability evaluation studies to uncover ninety-nine percent (99 %) of
usability problems within an application, as the evaluation would require the sample
inputs from four hundred and eighteen (418) expert evaluators. The cornerstone to
a successful HE is the selection of usability heuristics that are contextually
relevant.
Usability heuristics are general rules of thumb that set to guide the design of an
interface. In their first founding method of HE, both Nielsen and Molich proposed
a smaller set of heuristics that are made up of nine empirical usability principles
[32]. These were empirically tested principles derived from the analysis of 249
usability problems [31]. Later, Nielsen revised the heuristics to the ten with better
explanatory power [37]. Besides the 10 Usability Heuristics by Nielsen, there are
other usability guidelines such as Gerhardt-Powals Heuristics [38] which were
based on cognitive design principles. By far, the heuristics by Gerhardt-Powals is
the only set of heuristics that has been validated by human factors research [39].
416 L.C. Cheng and M. Mustafa

Gerhardt-Powals Heuristics
1. Visibility of system status 1. Automate unwanted workload
2. Match between system and the real 2. Reduce uncertainty
world 3. Fuse data
3. User control and freedom 4. Present new information with
4. Consistency and standards meaningful aids to interpretation
5. Error prevention 5. Use names that are
6. Recognition rather than recall conceptually related to function
7. Flexibility and efficiency of use 6. Limit data-driven tasks
8. Aesthetic and minimalist design 7. Include in the displays only
9. Help users recognize, diagnose, and that information needed by the
recover from errors user at a given time.
10. Help and documentation 8. Provide multiple coding of
data when appropriate.
9. Practice judicious redundancy.

Fig. 43.6 Comparison of usability heuristics

However, Nielsen’s 10 Usability Heuristics are the most widely adopted guidelines
compared to others, as it is more concise and easily understood (Fig. 43.6).
Most usability heuristics were developed way before the emergence of apps cul-
ture. Having said so, Nielsen’s heuristics is still relevant for evaluating mobile tech-
nologies [40]. For instance, Wright et al. have applied the heuristics in [37] to
evaluate a mobile fax application known as MoFax [41]. MoFax was an application
created to support industry representatives who often send faxes of plans to conven-
tional fax machines while out in the construction field. Initially, Wright et al. have
planned to perform field testing with real users. Prior to the summative testing,
Wright et al. have discovered that the MoFax interface was so unusable that they
decided to conduct an HE instead of having a costly user testing when the problems
are obvious. With three (3) expert evaluators, fifty-six (56) usability problems were
then identified for MoFax, and the developers mitigated all the problems by rede-
signing the application. The advantages of HE are in its effectiveness to identify all
major and minor usability problems of an application in any given stage [14]. With
the inputs from three to five experienced evaluators, all identified problems could be
prioritized and analyzed along with proposed solutions to improve usability. The
benefit to having experienced evaluators for usability inspection is that time can be
saved during the process of problem analysis as the evaluators can act as the ana-
lysts themselves to provide sample solutions. Testing with end users would require
the presence of moderators and external analysts to analyze the identified usability
problems before generating any solutions. HE can be performed by less-experienced
people as they will be guided with the heuristics [14]. For instance, in a comparative
study of web usability for the blinds, Mankoff et al. [42] have discovered that the
developers who are not the usability expert were capable to find 50 % of known
usability problems with HE. Mankoff et al. have concluded that HE, as an inspec-
tion method, is more cost-effective in their experiment if they were to perform user
testing with the blind users.
43 A Reference to Usability Inspection Methods 417

However, HE is not perfect. Several studies have reported that HE is not always
reliable in identifying usability problems, even though the evaluators have been
guided with a set of heuristics [15, 43–47]. The issues come from inconsistent
reporting among different evaluators who evaluated the same interface, as there are
no standard ways of documenting the findings with a common lingo [48]. Thus,
there are two notable revised HE methods, namely, HE-Plus and HE++ which are
developed by Chattractichart and Lindgaard which aimed to address the reliability
issues [49–51]. The method of HE-Plus used a catalogue of node-based problem
descriptors that was found in [48] and overlapped it with Nielsen’s heuristics.
However, such approach is quite time consuming and complex, as it requires the
evaluators to keep checking with an exhaustive list of catalogue, and there will be
situations that the problem description nodes were too rigid to describe a usability
problem. To validate the effectiveness of these new methods, a comparative study
had been conducted between the methods of HE, HE-Plus, and HE++ [12]. From
the study, both HE++ and HE-Plus have demonstrated that they have better reliabil-
ity and effectiveness compare to HE. However, there is no way to ascertain the find-
ings in [12], as the profile of the recruited expert evaluators was being misguided as
part of the sample population were students.

43.4 Overview

Usability evaluation is essential for creating a usable application that is acceptably


easy and sometimes safe to use within a specific environment. However, testing
with real users is not cost-effective. This paper had revisited the inspection tech-
niques of TA, CW, and HE. All these methods usually need three to five usability
experts to role-play the interaction of real users and provide constructive insights
whether a prototype is fit to be used. By far, HE is one of the most widely adopted
inspection methods that is seemingly cost-effective. However, HE has some reli-
ability issue that needs to be addressed. In view of that, this paper has highlighted
the gaps and research opportunities for developing a more streamlined usability
inspection method. There is a need to further advance the existing usability inspec-
tion methods for meeting today’s shorter time to market development model.

References

1. Nielsen, J. (1992). Finding usability problems through heuristic evaluation. In Proceedings of


CHI’92 (pp. 373–800). New York: ACM.
2. Van Greunen, D. & Wesson, J. L. (2002). Formal usability testing of interactive educational
software: a case study. In J. Hammond, T. Gross & J. Wesson (eds). Usability: creating a com-
petitive edge. In Proceedings of the world computer congress, Montréal, Québec, Canada.
3. Kock, E. D., Biljon, J. V., & Pretorius, M. (2009). Usability evaluation methods: Mind the
gaps. In SAICSIT conference 2009 (122–131).
418 L.C. Cheng and M. Mustafa

4. Nielsen, J. (2003). Return on investment for usability. Nielsen Norman Group. Retrieved
September 16, 2013 from http://www.nngroup.com/articles/return-on-investment-for-usability/
5. Nielsen, J. (1998). Cost of user testing a website. Jakob Nielsen’s Alertbox. Retrieved June 16,
2013 from http://www.nngroup.com/articles/cost-of-user-testing-a-website/
6. Nielsen, J. (2013). Salary trends for usability professionals. Jakob Nielsen’s Alertbox: 8
May 2012. Retrieved June 16, 2013 from http://www.nngroup.com/articles/
salary-trends-usability-professionals/
7. Barnum, C. M. (2011). Usability testing essentials: ready, set…test! Burlington: Morgan Kaufmann.
8. Nielsen Norman Group. Usability evaluation, (n.d.). Retrieved June 16, 2013 from http://www.
nngroup.com/consulting/usability-evaluations/
9. Scriven, M. (1967). The methodology of evaluation. In R. E. Stake (Ed.), Curriculum evalua-
tion. Chicago: Rand McNally. American Educational Research Association.
10. Hofman, R. (2011). Range statistics and the exact modeling of discrete non-gaussian distribu-
tions on learnability data. In A. Marcus (Ed.), Design, user experience, and usability. Theory,
methods, tools and practice (Vol. 6670, pp. 421–430). Berlin/Heidelberg: Springer.
11. Nielsen, J., & Mack, R. L. (Eds.). (1994). Usability inspection methods. New York: Wiley.
12. Chattratichart, J., & Lindgaard, G. (2008). A comparative evaluation of heuristic-based usabil-
ity inspection methods. In CHI’08 Extended abstracts on human factors in computing systems
(pp. 2213–2220). New York: ACM.
13. Nielsen, J. (2007). High-cost usability sometimes makes sense. Jakob Nielsen’s Alertbox. Retrieved
June 16, 2013 http://www.nngroup.com/articles/when-high-cost-usability-makes-sense/
14. Holzinger, A. (2005). Usability engineering methods for software developers. Communication
of the ACM, 48(1), 71–74.
15. Hornbaek, K. (2005). Current practice in measuring usability: challenges to usability studies
and research. International Journal Human-Computer Studies, 64(2006), 79–102.
16. Desurvire, H., Kondziela, J., & Atwood, M. (1992). What is gained and lost when using evalu-
ation methods other than empirical testing. In Proceedings of human-computer interaction
(HCI’92), University of York, Heslington, York, UK.
17. Coursaris, C. K., & Kim, D. J. (2011). A meta-analytical review of empirical mobile usability
studies. Journal Of Usability Studies, 6(13), 117–171.
18. Usability Professionals’ Association (UPA). (2009). UPA 2009 salary survey (public version).
Bloomingdale: UPA.
19. Ackerman, A. F., Buchwald, L. S., & Lewski, F. H. (1989). Software inspections: an effective
verification process. IEEE Software, 6(3), 31–36.
20. Eagan, M. E. (1986). Advances in software inspections. IEEE Transaction Software
Engineering, 12(7), 744–751.
21. Nielsen, J. (1994). Usability inspection methods. In Conference companion on human factors
in computing systems (CHI’94), Catherine Plaisant (Ed.) (pp. 413–414). New York: ACM.
22. Nielsen, J. (1994). Enhancing the explanatory power of usability heuristics. In Proceedings of
CHI’94 (pp. 152–158). New York: ACM.
23. Shneiderman, B. (1987). Designing the user interface: strategies for effective human computer
interaction. Reading: Addison-Wesley.
24. Dumas, J. (2007). The great leap forward: The birth of the usability profession (1988–1993).
Journal of Usability Studies, 2(2), 54–60.
25. Rogers, Y., Sharp, H., & Preece, J. (2011). Interaction design: beyond human-computer inter-
action (3rd edition). West Sussex: Wiley.
26. Annett, J., & Duncan, K. D. (1967). Task analysis and training design. Occupational
Psychology, 41, 211–221.
27. Hornsby, P. (2012). Hierarchical task analysis. UX Matters. Retrieved November 16, 2012,
from http://www.uxmatters.com/mt/archives/2010/02/hierarchical-task-analysis.php
28. Cooper, A., Reimann, R., & Cronin, D. (2007). About face 3: the essentials of interaction
design. Indianapolis: Wiley.
29. Jonassen, D. H., Tessmer, M., & Hannum, W. H. (1999). Task analysis methods for instruc-
tional design. Mahwah: Lawrence Erlbaum Associate.
43 A Reference to Usability Inspection Methods 419

30. Lewis, C., & Wharton, C. (1997). Cognitive walkthroughs. In M. G. Helander, T. K. Landauer,
& P. V. Prabhu (Eds.), Handbook of human-computer interaction (pp. 717–732). Amsterdam:
Holland.
31. Spencer, R. (2000). The streamlined cognitive walkthrough method, working around social con-
straints encountered in a software development company. In Proceedings of the SIGCHI confer-
ence on Human Factors in Computing Systems (CHI’00) (pp. 353–359). New York: ACM.
32. Nielsen, J., & Molich, R. (1990). Heuristic evaluation of user interfaces. In Proceedings of
ACM CHI’90 conference, (pp. 249–256). Seattle, 1–5 April 1990.
33. Nielsen, J. (1992). Finding usability problems through heuristic evaluation. In Proceedings of
CHI’92. New York: ACM.
34. Nielsen, J., & Landauer, T. K. (1993). A mathematical model of the finding of usability prob-
lems. In Proceedings of ACM/IFIP INTERCHI’93 conference, Amsterdam, Netherlands
(pp. 206–213).
35. Virzi, R. A. (1992). Refining the test phase of usability evaluation: how many subjects is
enough? Human Factors, 34(4), 457–468.
36. Lewis, J. R. (1994). Sample sizes for usability studies: additional considerations. Human
Factors, 36, 368–378.
37. Nielsen, J. (1995). 10 usability heuristics for user interface design. Jakob Nielsen’s Alertbox, 1995.
Retrieved November 16, 2012 from http://www.nngroup.com/articles/ten-usability-heuristics/
38. Gerhardt-Powals, J. (1996). Cognitive engineering principles for enhancing human – computer
performance. International Journal of Human-Computer Interaction, 8(2), 189–211.
39. U.S. Department of Health and Human Services. (2012). The research-based web design &
usability guidelines. Washington: U.S. Government Printing Office. Retrieved September 16,
2012, from http://www.usability.gov/basics/index.html
40. Brewster, S. A., & Dunlop, M. D. (2004). Mobile human-computer interaction – mobile HCI
2004. Lecture notes in computer science. Springer. Note: Vol. 3160.
41. Wright, T., Yoong, P., Noble, J., Cliffe, R., Hoda, R., Gordon, D., & Andreae. C. (2005).
Usability methods and mobile devices: an evaluation of MoFax. In Proceedings of the 4th
international conference on mobile and ubiquitous multimedia (pp. 29–33), New York: ACM.
42. Mankoff, J., Fait, H., & Tran, T. (2005). Is your web page accessible?: A comparative study of
methods for assessing web page accessibility for the blind. In Proceedings of the SIGCHI
conference on human factors in computing systems, Portland, Oregon, USA (pp. 41–50).
43. Jacobsen, N. E., Hertzum, M., & John, B. E. (1998). The evaluator effect in usability tests. In
CHI’98 conference summary (pp. 255–256). Reading: Addison-Wesley.
44. Hertzum, M., & Jacobsen, N. E. (2001). The evaluator effect: A chilling fact about usability
evaluation methods. International Journal of Human-Computer Interaction, 13(4), 421–443.
45. Kjeldskov, J., & Graham, C. (2003). A review of mobile HCI research methods. In Proceedings
of 5th international mobile HCI 2003 conference. Italy: Udine.
46. Cockton, G., & Woolrych, A. (2002). Sale must end: Should discount methods be cleared off
HCI’s shelves? Interactions 9, 5, (pp. 13–18). September 2002.
47. Kock, E. D., Biljon, J. V., & Pretorius, M. (2009). Usability evaluation methods: Mind the
gaps. In SAICSIT conference 2009, Vaal River, South Africa (pp. 122–131).
48. Andre, T. S., Hartson, H. R., Belz, S. M., & McCreary, F. A. (2001). The user action frame-
work: A reliable foundation for usability engineering support tools. International Journal of
Human – Computer Studies, 54(1), 107–136.
49. Chattratichart, J., & Brodie, J. (2002). Extending the heuristic evaluation method through con-
textualisation. In Proceedings of 46th annual meeting of the human factors & ergonomics
society, Baltimore, Maryland, USA. (pp. 641–645).
50. Chattratichart, J., & Brodie, J. (2003). HE-plus – towards usage-centered expert review for
website design. In Proceedings of for use 2003, (pp. 155–169). MA: Ampersand Press.
51. Chattratichart, J. & Brodie, J. (2004). Applying user testing data to UEM performance metrics.
In Proceedings of CHI 2004 (pp. 1119–1122). Vienna: ACM Press.

View publication stats

You might also like