Resource: Usability Test Quality Heuristics
The following indicators can help determine if a usability test will produce useful results.
This is not an exhaustive list, but it should be helpful in planning and assessing usability tests. Additions are welcome!
Study design
Indicator | Good sign | Bad sign |
---|---|---|
Study purpose | Researcher clearly articulates the purpose of the usability test (for example, as a specific question to answer, an area of inquiry, etc.). | Researcher does not specify a purpose for the test, or the purpose specified is very broad, like “testing the app” or “finding problems.” |
Number of participants | Tests include a sufficient number of participants necessary to see patterns, and there’s only one participant per session. | Study includes only one participant, or includes more that two participants testing simultaneously (like a focus group). |
Sampling bias | Participants are a diverse set of people who actually use the application (or do the task the application is supposed to support). | Participants are experts (vs. average users), friends, family, or close colleagues of the product team. |
Moderator style
Indicator | Good sign | Bad sign |
---|---|---|
Question style | Researcher asks follow up questions that are open-ended and non-leading. | Questions are leading or subtly suggest potential solutions or hypotheses. |
Task variability | Researcher asks participants to complete similar tasks related to the study’s purpose. | Researcher asks participants to complete different tasks, or tasks unrelated to the study’s purpose. |
Framing | Researcher asks participants to complete tasks with the product or service that align with their work-related goals. | Researcher asks participants to complete tasks unrelated to their work-related goals. For example, asking a participant how they might send a fax when their job doesn’t call for that. |
Priming | Moderator asks participants to complete tasks without indicating how that might be done. For example, “How would you view the status of your application?” | Researcher guides participants in completing tasks. For example, “Which of the links in the header would you click to login?” |
Team participation
Indicator | Good sign | Bad sign |
---|---|---|
Number of teammates | The team designates a single moderator for the test, and at least one member of the product team observes the usability test. | A single person from the product team participates in and leads the test. |
Observer role | Observers do not moderate. They are generally quiet, and ask open-ended questions after the test has concluded. | Observers interrupt the participant, or attempt to sell or explain the product. Observers debate the participant’s actions in their presence. |
Sensemaking
Indicator | Good sign | Bad sign |
---|---|---|
Notetaking | Tests are recorded or notes are taken for sharing with absent stakeholders. | Tests are not recorded, or test results are not documented. |
Synthesis | Moderator debriefs with teammates immediately after each interview. Researcher looks for patterns across multiple participants and surfaces problems that affected several people. | Moderator reports the most memorable problems without conducting affinity mapping or some other analysis technique. |
Presentation of findings | Researcher reports findings to team and stakeholders in an easy to follow, well prioritized way. | Researcher presents the team a “basket of issues” or an unprioritized laundry list of potential changes. |
Incorporation of findings | Product team translates findings into future user stories or product refinements. | Researcher reports do not affect the product backlog or ongoing development work. |