Saturday, February 4, 2012

Reading Report 4 – “Usability Testing: What Have We Overlooked?”

Reading Report 4 – “Usability Testing: What Have We Overlooked?”

Citation:
Lindgaard, G., & Chattratichart, J. (2007). Usability Testing: What Have We Overlooked? CHI 2007, (pp. 1415-1424). Retrieved from http://www.itu.dk/~hulda/Usability/artikler/p1415-lindgaard.pdf


Key words:
Usability testing, Metrics, UEM (Usability Evaluation Method) , participant recruitment, number of users, number of tasks, recruitment of users

This article is about (main discussion):
The article explains research into usability testing to see if “The Magic Number 5” for users is accurate. The research is to find:
  1. That there is a correlation between number of users and the proportion of problems found.
  2. That there is a correlation between number of user tasks and the proportion of problems found.
In order to study these two ideas, the authors did not perform usability tests, rather they were allowed to use the raw data from a previous study from a previous study (CUE-4 from CHI 2003) into the number of users needed. They took data from nine of the usability teams. The nine teams all completed usability tests on the same hotel booking website, but the nine groups did not have the same user tasks or number of users. Lindgaard and Chattratichart analyzed the number of users, the number of user tasks and scenarios, and the number of problems that each group identified (p. 1417).

Upon completing their statistical analysis they found:
  • The first hypothesis that there is a correlation between the number of users and the proportion of problems found was not supported.
  • The second hypothesis that there is a correlation between the number of user tasks and the proportion of problems found was supported.
 The authors argue and state that:
  • There has been too great of a focus in ID on the number of users when the focus should really be on the number of user tasks and also the quality of participants.
  • The Magic Number of 5 does not hold. In this study, “the percentage of problems found by the nine teams ranged from 7% to 43% - nowhere near the predicted 85%. The argument [the 5 users will find 85% of the problems] is therefore not upheld by the present data” (p. 1422)
  • Giving users a persona to imagine during the tasks is helpful because “[t]he persona might have helped their test users place themselves in the real users’ shoes and hence carry out the required tasks the way real users would do. As a result, Team S [the only team with a persona] performed better than expected” (p. 1423)
  • Rather than the number of users, the focus should instead be on “careful participant recruitment […]. It [also] pays off to give many sets of user tasks to a small number of users in a usability test rather than giving many users the same limited set of user tasks in a usability test” (p. 1423).
The authors conclude that:
Less resources and research should be spent on studying the number of users to use for usability testing. Instead the focus of further research should be on “the role of user tasks on improving usability testing approaches as well as into the importance of recruitment of test users” (p. 1424).

References: 16, (1981-2004)

Relating this to our project:
  • When creating our usability tests, we should focus on increasing the number of user tasks
  • The quality of the user tasks and scenarios must also be a consideration
  • It will be more fruitful to give a higher quality usability test to a small number of users
  • Care and attention must be taken to ensure that we have a good variety of users. The users should have varying levels of computer savvy and from different age groups. We should not just focus on using one group of users because they may all find different errors or problems. For example, a more tech savvy user may uncover issues with the usability than a less skilled user. They may approach the slide collection in a much different way
  • It may be beneficial to have the users take on a persona while completing the user tasks

Reading Report 3 – “Usability testing for web redesign: a UCLA case study”

Reading Report 3 – “Usability testing for web redesign: a UCLA case study”

Citation: Dominique Turnbow, Kris Kasianovitz, Lise Snyder, David Gilbert, David Yamamoto, (2005) "Usability testing for web redesign: a UCLA case study", OCLC Systems & Services, Vol. 21 Iss: 3, pp.226 - 234


Keywords: usability testing, library web design, card-sorting, surveys, think-aloud, structured analysis

Summary and Arguments made in the article:
This article describes the process and usability testing that was done to redesign the UCLA Library website. The original website had many issues that were identified by users and also employees of the libraries. Some of the issues included:
  • Different graphics and layout on different departments’ websites
  • Inconsistent nomenclature and heavy use of library jargon, rather than terms that the users would actually use
Usability Testing Methodology Employed
  • Structured analysis – an inventory of all the different library pages were taken and entered into a spreadsheet. This allowed for easy comparison of the different pages.
  • Surveys – there was an online survey created. Anyone who completed the survey was entered into a $250 draw. They had 300 responses, but found that since the survey was not well-designed, most of the information was not useful. The most useful information came from the open-response questions.
  • Second survey – a second survey was done to investigate the terms that should be used on the library website. This survey was done on paper and distributed evenly at the different library locations in order to ensure that the responses didn’t all come from the same group of users.
  • Card-sort protocol – this was done to decide how to organize the website categories before thinking about the actual design. From the structured analysis the team came up with 76 “essential links.” These links were put on cards and users were asked to organize the cards into meaningful groups and name the groups. UCLA used 40 participants in the card sort, although research suggests 15-20 or 30 is the ideal number. To recruit users, they offered gift bags. It took two weeks to find enough participants. Users were given one hour to sort the 76 cards. During the initial tests, they found that users used a wide variety of category titles; therefore, they needed to standardize the terms. To do that, they did a second card sort protocol.
  • Think aloud protocol – from the previous stages enough info was gathered to build a prototype website. UCLA decided to use ten participants in the think-aloud, even though research suggests that five users is enough. They used more than one participant from each user group, for example undergraduate students and graduate students. Users were given a list of tasks to complete on the website. Two facilitators were used, “one read the questions and interacted with the participant while the other recorded the participant’s actions, including: the path taken to find the answer; anything said while navigating the site; and any observations of the participant’s behavior” (p. 232). After the task, users were asked, “their general impression of the site, suggestions for the designer, or any other comments about the website. Finally, there was a brief survey about the participant’s previous use and knowledge of the library and the library’s website” (p. 232). No major changes were made to the site based on the think-aloud.
Following the think-aloud usability testing, the full website was developed. A logo was designed and a standardized page template was created to be used on all pages. Once the site was created, the team solicited feedback through emails to faculty, comments through a link on the new page, and library staff feedback sessions (p. 234). For any revision made to the site, a think-aloud protocol will be used.

References: 3, (2000-2004)

Key Ideas and their relation to our project:
  • Surveys need to be well-made in order for them to provide useful information. Based on this Case Study, the open-ended questions provide more valuable data than selected responses.
  • It may take a while to find participants for testing; therefore, we need to keep this in mind when planning our timeline for prototypes
  • Card-sorting could be used to help with categories or set names
  • Use a variety of users for the think-aloud. Make sure not all users are of the same type
  • Give users a list of specific tasks to complete while doing usability testing. For example, we could ask them to find a specific slide, find slides for a specific location, etc.