Block 4 - Unit 2: Usability testing and field studies Flashcards
Usability testing. (4)
Involves users.
Emphasises ‘usable’ property - product is being tested rather than user.
(Often) controlled environment; performance of users on pre-planned tasks is repeatedly measured.
Goal - test if product is usable by intended user population to achieve the tasks designed for.
Key components of usability testing. (2)
User test.
User satisfaction questionnaire.
User testing. (Include examples)
Measures human performance on specific tasks.
Eg. reading different typefaces, navigation of menu types, info searching.
Logging of mouse / keyboard and video used to record performance.
User satisfaction questionnaire.
How do users feel about using the product - rating along different scales, after interaction.
Measures used in usability testing.
Time and number:
- time to complete a task (after a specified time away);
- Number of errors per task or per time unit;
- Number of views of online help / manuals;
- Number of users making a particular error;
- Number of users completing a task successfully.
Number of users needed in usability testing.
5 - 12 considered acceptable.
Can use less if lower budget / schedule, or for quick feedback on eg. logo placement - 2 or 3 users.
Remote usability testing (including advantage).
Users perform tasks in their own setting, logged remotely.
Advantage - many users tested at once, and logged data automatically compiled into statistical packages for analysis. Eg. no of clicks per page.
Usability testing (UB)
Doesn’t have to result in accurate measures of user performance, or be under controlled conditions, but does involve some form of performance assessment and a level of evaluator control, most commonly the use of pre-set tasks.
Several methods may be used - eg. indirect observation, user test to measure performance, questionnaire.
User tests often conducted along experimental lines, involving hypothesis testing and statistics.
Doing user testing. (Intro)
How to do user testing is peculiar to usability testing.
Controlling test conditions is central - careful planning; ensure all material is prepared, conditions are the same for each user, what is being measured is indicative of what is being tested and that assumptions are made explicit in the test design.
Elements of doing user testing. (4)
Design typical tasks.
Select typical users.
Prepare testing conditions.
Plan how to run the tests.
Design typical tasks (user testing).
Appropriate task to test users’ performance is critical.
‘Completion’ tasks set, eg. find website, create spreadsheet.
Quantitative performance measures obtained.
User tests most suitable for hi-fi prototypes, simulations and working products.
Task type depends on system type and evaluation goals / questions.
Eg. whether paper-prototypes, simulation or limited part of system’s functionality will influence breadth and complexity of tasks set.
Generally, tasks are 5 - 10 mins and designed to probe a problem.
Often straightforward, but occasionally more complex, eg. join an online community or solve a problem.
Select typical users (user testing)
What characteristics?
Some products aimed at eg. old / young, novice / expert.
Equal male / female split, unless aimed at one.
Previous experience with similar systems?
If user population is large - short questionnaire can help identify testers.
Prepare testing conditions (user testing)
Environment controlled to prevent unwanted influences and noise that distorts results.
Plan how to run the tests (user testing).
Schedule an script, equipment tested, pilot test.
Start with easy task to build confidence and familiarisation.
Avoid long tasks; keep session under 1 hour.
Don’t create too much data to analyse.
Potential sources of bias (user testing)
Users:
if don’t match profile, may get misleading results.
Evaluation tasks:
If users get unintentional clues from tasks, this may prevent exposing UPs.
Test setting:
Lab is artificial, and may need conditions adding to more closely simulate actual environment of use.
Evaluator / observer bias:
Behaviour may be altered when being watched. Observer may forget they shouldn’t ask leading questions, or help if user gets stuck.
Methodology:
‘Think-aloud’ alters behaviour.
Avoid ‘order effects’ - trying one design may give ‘practice’ to trying another. (So, counterbalance).
Reporting / analysis:
Data analysis and interpretation will be undertaken based on evaluator’s knowledge and experience - so biased to some extent.
Field studies (intro)
Typically used to find how a product / prototype is adopted an used in people’s working / everyday lives.
Such settings are ‘messy’ - activities overlap and constantly get interrupted.
The way people interact with products is often different than in a laboratory setting; a better sense of the success of a product is gained in the real world.
Trade-off of field studies.
Can’t test specific hypothesis about an interface or account, with the same degree of certainty, for how people react or use a product (than in a lab).
So, harder to determine what causes certain behaviour or what is problematic about the usability of a product.
Instead, qualitative accounts and descriptions of people’s behaviour and activities are obtained to reveal how they used the product and reacted to the design.
Some characteristics of field studies. (3)
Can be minutes, months or years.
Primary data collection - observing and interviewing people; video, audio and field notes.
May ask for paper or electronic diaries to be filled in at points in the day - when interrupted, encounter a problem or are in a particular location.
Things to do for field studies. (4)
Important to inform people of the length of study / session.
Need to agree part of site to be recorded and how.
Need to set up cameras unobtrusively, and if eg. in someone’s home - how to turn on / off.
What happens if product breaks?
Planning an observational evaluation (intro).
Expand DECIDE to present a procedural approach to ‘doing’ observational evaluation.
Primary focus - evaluation of low-fi prototypes (especially card-based and interface sketches).
Kinds of task to consider (‘I’)
Core tasks (frequent)
Those very important to users or the business.
Those that have some new design features / functionality.
Critical tasks.
Ones you feel have to be validated with users for greater clarity and understanding of the design team.
Those that scenarios were based on, which you used for developing the conceptual design of the UI.