The concept of heuristics has a long history, spanning the fields of philosophy, law, psychology, and human-computer interaction among others. Wikipedia's discussion of the topic includes links to numerous well-known psychological heuristics such as anchoring and adjustment, availability, and peak-end rule.
Heuristics in Human Computer Interaction
In the field of human-computer interaction (HCI), one of the most popular inspection-based methods for evaluating usability is the Heuristic Evaluation (HE) as described originally by Nielsen and Molich and later refined by Nielsen. Promoted for its cost efficiency and ease of implementation, the HE method consists of one or more experienced evaluators (3-5 recommended) applying an established set of guidelines (or heuristics) as they review a given system. Also known as an 'inspection' method or 'discount" method' of usability evaluation, the Heuristic Evaluation is seen as an economical alternative to empirical usability tests involving actual users.
In a “true” HE, the group of evaluators apply Nielsen's ten heuristics during their independent review of a system. The evaluators systematically explore the product or system, considering each heuristic in turn in an effort to identify potential usability issues.
The Value of Heuristic Evaluations
Heuristic evaluations are arguably most valuable as an early evaluation technique, performed on an existing system or first prototype in an effort to identify the major usability problems. By conducting a heuristic evaluation early on, two things are effectively accomplished:
- The major 'showstopper' problems are identified without incurring the expense, time, and certain frustration on the part of the users that would be accompany formal tests with users
- Anticipated problem areas are identified for later evaluation when empirical sessions with users are conducted
When the circumstances and constraints of a project warrant, heuristic evaluations have been shown by many researchers to offer designers and evaluators a valuable alternative to user testing.
The Challenges of Heuristic Evaluations
In my experience, true Heuristic Evaluations are rarely performed in the manner that Neilsen and Molich originally intended. Rather, they often amount to a single evaluator's 'expert review' of a system, and are subsequently open to criticism about the reliability of their findings and recommendations.
Even when multiple evaluators are used in a Heuristic Evaluation, it often proves difficult for evaluators to reliably interpret the guidelines and identify the same usability issues. Another common criticism of HE's is that they are capable of generating a lot of 'false positives' - usability problems that would likely go unobserved in an empirical usability study, but that end up being identified by evaluators whose awareness is heightened by the guidelines to identify relatively minor problems.
Modifying Heuristics to Meet New Needs
It has now been over 15 years since the Heuristic Evaluation method was first developed, and while it continues to be considered somewhat of a standard in the HCI industry, many evaluators have found that Nielsen’s original list does not always meet their specific needs. They discover that they frequently require alternative guidelines or some re-interpretation of Nielsen’s original descriptions in order for each heuristic to make sense. As a result, several modifications of Nielsen's heuristics have been developed over the years in an effort to help improve their interpretation, reliability, and 'goodness of fit' for new technologies and various industries.
So what's next for the Heuristic Evaluation method? Here are my predictions:
- The HE method will continue to be regarded as a core method in the HCI toolbox in future years.
- It will continue to be practiced largely in an ad-hoc, customized manner that only resembles its original form, namely in an informal, individual evaluator format. This will hamper the method's further development, and reduce both its perceived and actual value.
- As new technologies, products, and interactions are developed, new heuristics will be developed in concert.
If you are interested in reading more about Heuristic Evaluation, you may find this reading list useful.
- Original paper on heuristic evaluation
- Nielsen, J., & Molich, R. (1990). Heuristic evaluation of user interfaces. Paper presented at the ACM CHI'90 Conference on Human Factors in Computing Systems, Seattle, WA.
- Refinements on heuristic method
- Nielsen, J. (1994). Heuristic evaluation. In J. Nielsen & R. L. Mack (Eds.), Usability Inspection Methods. New York: John Wiley & Sons, Inc.
- Neilsen's heuristics
- Neilsen, J. Ten usability heuristics.
- Pierotti, D. Heuristic evaluation - A system checklist.
- Support for inspection-based methods and their comparison results
- Fu, L., Salvendy, G., & Turley, L. (2002). "Effectiveness of user testing and heuristic evaluation as a function of performance classification". Behaviour & Information Technology, 21(2), 137-143.
- Kantner, L., & Rosenbaum, S. (1997) "Usability Studies of WWW Sites: Heuristic Evaluation vs. Laboratory Testing". In the Proceedings of the 1997 Association for Computing Machinery's Special Interest Group on Documentation (ACM SIGDOC).
- Nielsen, J. (1992). "Finding usability problems through heuristic evaluation". Paper presented at the ACM CHI'92 Conference on Human Factors in Computing Systems, Monterey, CA.
- Nielsen, J., & Phillips, V. L. (1993). "Estimating the relative usability of two interfaces: Heuristic, formal, and empirical methods compared". Paper presented at the ACM INTERCHI'93, New York.
- Sawyer, P., Flanders, A., & Wixon, D. (1996). "Making a difference: The impact of inspections". Paper presented at the ACM CHI 1996 Conference on Human Factors in Computing Systems, Vancouver, BC, Canada.
- Challenges to inspection-based methods and their comparisons
- Cockton, G., & Woolrych, A. (2002). "Sale must end: Should discount methods be cleared off HCI's shelves?" Interactions, 9.5 (September & October), 13-18.
- Gray, W. D., & Salzman, M. C. (1998). "Damaged merchandise? A review of experiments that compare usability evaluation methods". Human-Computer Interaction, 13, 203-261.
- Jeffries, R., Miller, J. R., Wharton, C., & Uyeda, K. M. (1991). "User interface evaluation in the real world: A comparison of four techniques". Paper presented at the ACM CHI'91 Conference on Human Factors in Computing Systems, New Orleans, LA.
- Karat, C. M. (1994). "A comparison of user interface evaluation methods". In J. Neilsen & R. L. Mack (Eds.), Usability Inspection Methods. New York: John Wiley & Sons, Inc.
- Alternatives to Neilsen's Heuristics
- Bastien, J.M.C. & Scapin, D.L. (1995) "Evaluating a user interface with ergonomic criteria". International Journal of Human-Computer Interaction, 7(2), 105-121.
- Borges, J. A., Morales, I., & Rodriguez, N. J. (1996). "Guidelines for designing usable World Wide Web pages". Paper presented at the ACM CHI 1996 Conference on Human Factors in Computing Systems, Vancouver, BC, Canada.
- Emmus (European Multimedia Usability Services) (1999). CELLO: Evaluation by inspection.
- IBM (2003). Design basics.
- Instone, K. (1997). Site usability heuristics for the Web.
- ISO 9241-12 (1998). ISO 9241-12:1998 Ergonomic requirements for office work with visual display terminals (VDTs) -- Part 12: Presentation of information
- ISO 9241-110 (1996). ISO 9241-110:2006 Ergonomics of human-system interaction -- Part 110: Dialogue principles.
- Kamper, R. J. (2002). "Extending the usability of heuristics for design and evaluation: Lead, follow, and get out the way". International Journal of Human Computer Interaction, 14(3&4),447-462.
- Kantner, L., Shroyer, R., & Rosenbaum, S. (2002). Structured Heuristic Evaluation of Online Documentation. IPCC 2002 Proceedings. IEEE Professional Communication Society.
- Mayhew, D. (1992). General principles of user interface design. In Principles and guidelines in software user interface design. Englewood Cliffs, NJ: Prentice Hall.
- Miller, R. H. (2000). Web site evaluation criteria
- Muller, M. J., Matheson, L., Page, C., & Gallup, R. (1998). "Participatory heuristic evaluation". Interactions, 5 (September & October), 13-18.
- Purho, V. (2000). Heuristic inspections for documentation - 10 recommended documentation heuristics. STC Usability SIG Newsletter, 6(4).
- Rosenfeld, L. (1998). Information architecture heuristics. LouisRosenfeld.com, Aug 17, 2004.
- Rosenfeld, L. (1998). Heuristics for Search Systems. LouisRosenfeld.com, Sept 02, 2004.
- Sears, A. (1997). "Heuristic walkthroughs: Finding problems without the noise". International Journal of Human-Computer Interaction, 9(3), 213-234.
- Shneiderman, B. (1998). Eight golden rules of interface design. In "Designing the User Interface". Addison Wesley, 3rd Edition.
- Skinner, Grant. (2003). Usability heuristics for rich internet applications. boxesandarrows.com.
- Tognazzini, B. (2001). First principles.
- Usability.gov (2003). Research-Based Web Design & Usability Guidelines.
About the author
Dr. Todd Zazelenchuk (@ToddZazelenchuk on Twitter) holds a BSc in Geography, a BEd, an MSc in Educational Technology and a PhD in Instructional Design. Todd is an associate of Userfocus and works in product design at Plantronics in Santa Cruz, CA where he designs integrated mobile, web, and client-based software applications that enhance the user experience with Plantronics' hardware devices.
Love it? Hate it? Join the discussioncomments powered by Disqus
Foundation Certificate in UX
Gain hands-on practice in all the key areas of UX while you prepare for the BCS Foundation Certificate in User Experience. More details
Every month, we share an in-depth article on user experience with over 10,000 newsletter readers. Want in? Sign up now and download a free guide to usability test moderation.
If you liked this, try
Most people that carry out usability expert reviews use Jakob Nielsen's ten usability 'heuristics'. Many of these guidelines are common sense but they are not based on substantive research. The International usability standard, BS EN-ISO 9241-110 proposes an alternative set of seven guidelines. These guidelines have the benefit of international consensus and they can be applied to any interactive system. Usability Expert Reviews: Beyond Heuristic Evaluation.
User Experience Articles
Our most recent articles
- Apr 3: Is Usability a Science?
- Mar 6: Why iterative design isn’t enough to create innovative products
- Feb 6: The Beginners' Guide to Contextual Interviewing
- Jan 9: The 8 competencies of user experience: a tool for assessing and developing UX Practitioners
- Dec 5: Non-UX books that every UX practitioner should read
Our most commented articles
Search for articles by keyword
- 7 articles tagged accessibility
- 4 articles tagged axure
- 5 articles tagged benefits
- 16 articles tagged careers
- 8 articles tagged case study
- 1 article tagged css
- 8 articles tagged discount usability
- 2 articles tagged ecommerce
- 14 articles tagged ethnography
- 14 articles tagged expert review
- 1 article tagged fitts law
- 4 articles tagged focus groups
- 1 article tagged forms
- 6 articles tagged guidelines
- 10 articles tagged heuristic evaluation
- 7 articles tagged ia
- 14 articles tagged iso 9241
- 11 articles tagged iterative design
- 3 articles tagged layout
- 2 articles tagged legal
- 11 articles tagged metrics
- 3 articles tagged mobile
- 7 articles tagged moderating
- 3 articles tagged morae
- 2 articles tagged navigation
- 9 articles tagged personas
- 15 articles tagged prototyping
- 7 articles tagged questionnaires
- 1 article tagged quotations
- 4 articles tagged roi
- 16 articles tagged selling usability
- 12 articles tagged standards
- 44 articles tagged strategy
- 2 articles tagged style guide
- 4 articles tagged survey design
- 5 articles tagged task scenarios
- 2 articles tagged templates
- 21 articles tagged tools
- 53 articles tagged usability testing
- 3 articles tagged user manual