710 likes | 1.55k Views
Heuristic Evaluation. ☃. HCC 729, 2/13/14. We’ll follow up next time. Inspirations, reading feedback Your HTAs and personas. How to conduct a Heuristic Evaluation. Read this: http :// www.nngroup.com /articles/how-to-conduct-a-heuristic-evaluation/.
E N D
Heuristic Evaluation ☃ HCC 729, 2/13/14
We’ll follow up next time • Inspirations, reading feedback • Your HTAs and personas
How to conduct a Heuristic Evaluation Read this: http://www.nngroup.com/articles/how-to-conduct-a-heuristic-evaluation/ Adapted from slides by Karen Tang and Ryan Baker
What is an evaluation? • Gather data about the usability of a product or design by a particular group of users for a particular activity or task within a particular environment or context • Evaluation goals: • Assess extent of system’s functionality • Assess effect of interface on user • Identify specific problems with system
HE vs. user testing • When we can, we want to test with real users • HE is a “discount” usability technique • When it’s useful: • When real users are unavailable • Very early in the design • As a sanity check (but not a replacement for user testing)
Why HE is great • Cheap • Doesn’t “spend” users • Fast • 1-2 days (instead of 1 week) • Good • Proven effective: the more careful you are, the better it gets • Easy to use • Relatively easy to learn, can be taught
Heuristic Evaluation • A type of discount usability testing • A rational method – an expert applies “heuristics” • mentally apply a theory or rule to the design and see if that theory/rule’s advice is being followed • Key Idea: Multiple expert evaluators independently apply a set of heuristics to an interface, produce Usability Action Reports (UARs), combine & prioritize their findings
What can you evaluate with a HE? • Any interface that has been “developed” • Pre-existing webpage • Sketch of a future interface (can be fully implemented or only exist as a sketch) • This method can be applied on your own interface, or a competitor’s • You will evaluate the interface according to a standard set of 10 heuristics
How many evaluators are needed? • Nielsen recommends at least 3, but go for 5!
Who should do the HE? • Anyone who knows the appropriate heuristics can do a HE • But, heuristic evaluation experts find almost twice as many problems as novices • Heuristic evaluation experts who are also domain experts find almost three times as many problems as novices
Phases of Heuristic Evaluation 0) Pre-evaluation training (optional) Give evaluators needed domain knowledge & information on the scenario • Evaluate the interfaceto find usability problems • Record the problem • Aggregate problems • Assign severity rating 5) Find a solution complexity rating
Which heuristics to use? • Many possible heuristic sets • Some standard sets (e.g. Nielsen’s usability heuristics) • You might create your own heuristics, e.g. for specific applications • We’ll focus on Nielsen’s, which cover a range of general usability issues
Find which heuristic is violated • Simple & Natural Dialog • Speak User’s Language • Minimize User’s Memory Load • Consistency • Feedback • Clearly Marked Exits • Shortcuts • Good Error Messages • Prevent Errors • Help & Documentation Nielsen’s 10 Heuristics http://www.nngroup.com/articles/ten-usability-heuristics/
Examples of applying the heuristics • http://www.slideshare.net/sacsprasath/ten-usability-heuristics-with-example
Record the problem • Each evaluator writes a Usability Action Report (UAR) describing each usability problem they encounter • HEs are typically used to report problems • However, UARs can be used to report both the goodand bad qualities of an interface in other usability evaluations… • I have posted a template UAR on the blog with the assignment
Sample UAR • EVALUATOR: XXXXX • ID NUMBER: XXX • NAME: Descriptive name for the problem • EVIDENCE: Describe the violation, and why you wrote this report. what heuristic was violated, • EXPLANATION: Your interpretation: and why. • Severity: Write up at the end of the evaluation • Fixability: Write up at the end of the evaluation • Possible Fix: Write up at the end of the evaluation
Keep looking for problems! • Usually takes a few hours • A shorter time may not find important problems • A longer time will exhaust the evaluator, and they may become less productive • For very large interfaces, it is good to break heuristic evaluation into several sessions
What about multiple problems? • This happens a lot, record them separately. • This is not busywork…. • It may be possible to fix some of the problems, but not all of them • The problems might not always be linked to each other – one may show up in other situations too
You are not done yet… • You still need to address the bottom half of the UAR: • Severity • Solution Complexity • Possible Fix • You may want to take a break before finishing these UARs…
Aggregate Problems • Wait until all UARs are in • You are aggregating across all evaluators • Aggregating usability problems: • Combine problems by consensus • Gain a sense of relative importance after you’ve seen a few problems • At this point, decide which entries are and aren’t problems (but keep original version of report somewhere)
Assign Severity Rating to UARs • Severity Ratings help project leads determine what problems should be given more developer time • Not all problems can be fixed • Some problems will have more severe consequences • Each evaluator should assign severity separately
Assign Severity Rating to UARs Based on a combination of: • Frequency • How common or rare is the problem? • Impact • How easy is it to overcome the problem? • How disastrous might the problem be? • Persistence • How repeatedly will users experience the problem? • Are workarounds learnable?
Assign an Overall Severity Rating • It is helpful to developers in allocating resources to have one severity rating for the problem. • Therefore, evaluators need to combine their opinion of a problem’s Frequency, Impact, & Persistence ratings into one Severity evaluation
Nielsen’s Severity Ratings • Usability Blemish. Mild annoyance or cosmetic problem. Easily avoidable. • Minor usability problem. Annoying ,misleading, unclear, confusing. Can be avoided or easily learned. May occur only once. • Major usability problem. Prevents users from completing tasks. Highly confusing or unclear. Difficult to avoid. Likely to occur more than once. • Critical usability problem. Users won’t be able to accomplish their goals, and may quit using system.
False positives • There’s no virtue in finding 6,233 problems, if very few of them actually cause problems for a user • Every problem reported in a heuristic evaluation takes time for the developers to consider • Some interface aspects that seem like problems at first might not be problems at all
5: Solution Complexity Rating • Some problems take more time to fix than others, so it’s important to allocate developers’ time well • Ideally this could be made by either a developer, or someone who is familiar with development in the target platform
Solution Complexity Rating • Trivial to fix. Textual changes and cosmetic changes. Minor code tweaking. • Easy to fix. Minimal redesign and straightforward code changes. Solution known and understood. • Difficult to fix. Redesign and re-engineering required. Significant code changes. Solution identifiable but details not fully understood. • Nearly impossible to fix. Requires massive re- engineering or use of new technology. Solution not known or understood at all.
Record Possible Fixes • While evaluating solution complexity evaluator may have thought about how the problem could be fixed • Record these possible fixes as suggestions to developers • Don’t focus on feasibility of solutions (that is their job) • Your suggestions may be thought-provoking
Phases of Heuristic Evaluation 0) Pre-evaluation training (optional) Give evaluators needed domain knowledge & information on the scenario • Evaluate the interfaceto find usability problems • Record the problem • Aggregate problems • Severity rating 5) Solution complexity rating
Why HE? • They find a reasonably large set of problems • They are one of the easiest, quickest, and cheapest methods available
HE vs. User Testing • User tests are more effective at revealing when a system’s manifest model or metaphor is confusing • User tests are less effective at finding obscure problems • User tests are also much more expensive • Advice: use HE first, to find the obvious problems, then user test.
For next week Assignment Readings
In-class assignment Perform HE on UMBC class search with PeopleSoft Use template form from the blog
HW: Perform HE on your sites • Go through the 5 stages of HE for your website • If on a team, each go through HE individually, then combine later • Turn in 1 completed form for each incident • Come up with at least 8 UARs for each webpage • Aggregate, finish filling out the template! • Everyone writes 100-200 words describing what they learned
HW Extra credit • Perform a HE on another student’s web site (same as previous slide)
Reading • Required • Usability Engineering Chapter 6 (focus on 6.1-6.6 and 6.8) • Optional • Think aloud method