Site icon UX Mastery

How to run an heuristic evaluation

Heuristic analysis

Being a user experience designer often requires juggling tensions, whether they be juggling creative tension, managing stakeholder tension, or constantly living in the tension between design expertise and user-centred design methods. In the early 1990’s this third tension sparked a lot of high-level thinking by people wanting to define, once and for all, what the principles and best practices for interface design should be. But they didn’t quite crack it.

The process we use for heuristic evaluations today can be traced directly back to that period. Rolf Molich and Jakob Nielsen published their foundational ‘Improving a human-computer dialogue’ in 1990, building on work by people like David Cheriton (1976), Neilsen’s future consulting partner Donald Norman (1983), and Ben Schneiderman (1987), amongst others.

At the same conference where Molich and Nielsen presented their heuristic evaluation method, Clayton Lewis, Peter G Polson and their colleagues at the Institute of Cognitive Science presented a walkthrough methodology tackling the same problems. Usability practitioners recognised the benefits of combining these two methods, and although the naming rights are a bit grey, this combination is what most UXers consider an heuristic evaluation today.

The good and the bad

There are some good reasons why you might use this technique:

But there are also some fundamental problems with the technique, so it is  critical that you understand its limitations and dangers too:

The argument goes: “If you’re such an expert and have the experience you claim, how come you can’t just give us the answers for the right design?” Well, if you have sufficient experience and expertise in usability, you’ll also be well aware that users are notoriously unpredictable. It’s common to experience ‘aha!’ moments in usability tests that show something we never expected, even when the design seems to conform to key heuristics perfectly.

Heuristic evaluations are no substitute for usability testing, but they can help improve the potential of usability tests if they’re used in conjunction; running a heuristic evaluation before beginning a round of usability tests will reduce the number and severity of design errors discovered by users, helping minimise problems and distractions during the testing.

Getting your hands on some suitable heuristics

Although Nielsen is probably the best known, there are many other sets of heuristics, and some of them are considered better. Here’s a list of all the ones I know of. Any that I’ve missed?

If you really want to get stuck in, check out Sidney Smith & Jane Mosier’s nine hundred and forty four guidelines for designing user-interface software (from 1986).

Susan Weinschenk and Dean Barker decided to amalgamate the usability guidelines from multiple sources (including Nielsen’s, Apple and Microsoft) and did a massive card sort resulting in twenty of their own heuristics.

I use Jill Gerhardt-Powals’ heuristics as they take a more holistic approach to evaluating the system, which I find easier to get my head around.

Running your own heuristic evaluation

Ingredients:

Before you start:

Method:

  1. Decide on a set of heuristics to use. If you don’t have a good reason not to, I would recommend Neilsen’s or Gerhardt-Powals.
  2. Select your team of 3-5 evaluators. Ideally they both usability experience as well as domain knowledge related to your project, and might be found in your design team or list of professional contacts. You should give them all the same training on the principles and process, and ensure they’re interpreting the heuristics properly.
  3. Set up a system of severity codes (critical issue, serious issue, minor issue, good practice), or traffic light scheme (red, orange, yellow, green), and make sure the evaluators understand them and can use them consistently.
  4. Conduct the walkthrough itself. Step by step through each agreed task, wearing the users shoes (terminology, priorities, likely choices).
  5. Look for and identify problems based on the set of heuristics.
  6. Note where the problem is (the page/screen, location on the page), how bad it is (rating scale), and push on to the next issue.
  7. I like to go back and note the options for fixing the issues seprately afterwards, as switching between ‘seek mode’ and ‘fix mode’ is pretty distracting.
  8. After you’re done, collate and analyse results from the multiple experts. This is where you realise the consistency between evaluators was important. Otherwise you’re trying to reconcile too many levels and interpretations.

Writing up a report

The whole point is to surface the key issues and start a discussion about how to improve. There’s enough detail that you can’t avoid a certain amount of documentation, but it’s also pretty important that you deliver the findings in person. You’ve systematically deconstructed and criticised something that your team and stakeholders will likely be protective of, so being there to show your loyalty, to clarify and explain gently, and to justify your findings is pretty key.

Presenting the report itself is often done as a deck of slides, or as a written report. Some really slick presentations have captured video of the  screens to include mouse movement, interactions, audio of the issues being described, and to highlight particular issues.

However you choose to put it together, make sure you cover the following:

What comes next?

Heuristic evaluations only give you an indication of where likely issues are. They suggest some options for heading towards fixes, but you still haven’t put the design in front of real users, so you won’t know how much you’re missing out on. Your next step is to take a stab at solving the biggest problems, and then arranging your next set of usability tests. You’ll find that you have both some assumptions that you can base design hypotheses on, and some ideas of where to focus your testing. From there, you can be guided by the users themselves. Just promise me you won’t stop with only a heuristic evaluation!