Evaluation


This module covers weeks 12-13 of the course.

Note that this material is subject to ongoing refinements and updates!

Evaluation

In IxD, everything “from low-tech prototypes to complete systems, from a particular screen function to the whole workflow, and from aesthetic design to safety features” can be evaluated (Sharp et al., 2019, p. 498). Evaluation can take place in a research lab, but also “in the wild”, and can be undertaken through direct engagement with users, but also through more passive, indirect methods.

Field studies "in the wild" can:

Why conduct evaluations?

Evaluations are usually conducted after designers have produced something (either a prototype or more concrete design) based on user requirements. These designs are then "evaluated to see whether the designers have interpreted the users’ requirements correctly and embodied them in their designs appropriately" (Sharp et al., 2019, p. 499).

Types of evaluation

Three broad types of evaluation:

Some methods used in evaluation are the same as those used for discovering requirements, such as observation, interviews, and questionnaires. A common type of evaluation conducted in a controlled setting is usability testing, where the “primary goal is to determine whether an interface is usable by the intended user population to carry out the tasks for which it was designed” (Sharp et al., 2019, p. 501).

The aforementioned System Usability Scale (SUS) is still a common way of testing usability of a system. See the original version here:
Brooke, J. (1986). SUS - A quick and dirty usability scale. https://digital.ahrq.gov/sites/default/files/docs/survey/systemusabilityscale%2528sus%2529_comp%255B1%255D.pdf

Ways of evaluation

Three ways to evaluate:

Design guidelines, policies,heuristics, principles, rules, laws, policies, etc.

Some examples:

Shneiderman's Eight Golden Rules of Interface Design are a seminal set principles of interface design:


A cognitive walk-through to create a Spotify playlist:
Dalrymple, B. (2018). Cognitive Walkthroughs. https://medium.com/user-research/cognitive-walkthroughs-b84c4f0a14d4

Web analytics

Google analytics a common platform for web analytics:
https://marketingplatform.google.com/about/analytics/

Google analytics visitors
Figure: Example of functionality from Google Analytics, showing visitors to a page according to country


A/B testing

A/B testing is a "way to evaluate a website, part of a website, an application, or an app running on a mobile device... by carrying out a large-scale experiment to evaluate how two groups of users perform using two different designs—one of which acts as the control and the other as the experimental condition, that is, the new design being tested" (Sharp et al., 2019, p. 574).

simple A/B test
Figure: A basic visual explanation of A/B testing. A is an original design, while B is a variation of it. Users are split into two groups, one who tests the original and one who tests the variation.

A list of AB tests towards the bottom of this page: https://business.adobe.com/blog/basics/learn-about-a-b-testing

Predictive modelling, such as Fitt’s Law

We have covered Fitt’s Law previously, but it has some interesting applications to evaluation, such as:

In simpler terms, we can say that big objects that are near are easier to access, while those that are small and further away are harder to access:

© Interaction Design Foundation, CC BY-SA 4.0

  


Back to top | © Paul Haimes at Ritsumeikan University | View template on Github