Thinking Aloud for Digitized Books Interview Analysis

From MPDLMediaWiki
Revision as of 08:22, 4 October 2010 by Rkiefl (talk | contribs) (→‎Analysis)
Jump to navigation Jump to search

This is a protected page.

Results from 'Thinking Aloud' interviews for ViRR Release 3[edit]

The usability evaluation method 'think aloud' is the most affordable and efficient among numerous test methods. If one follows human centred design principles to go for ISO 13407[1], it is good way to get relevant issues fast.

Participants[edit]

According to the main groups defined in Personas 2 + 11 Persons were interviewed. 2 Persons were interviewed at the MPDL in Munich and 11 at the MPIER in Frankfurt:

Group Amount
Administrator 0
Librarian 7
Scientist 2
Secretary 4
Guest 0
All 13

Working conditions[edit]

Working conditions are quite well among all 13 Participants although there is some distraction and more urgent tasks need to be tackled while screen handling. This is a quite good rate compared to tests from 2008 and 2009 even if we partial the two MPDL staff members out.

Desktop space Rating
Lot of space 13
Good 0
Crowded 0
External Help Rating
Available 13
IT Services 0
Poor 0
Noise and distraction Rating
Distraction 0
Little 4
None 7
Urgency of task Rating
Lower 5
Average 6
Higher 2

Experience with Screen Interfaces[edit]

Participants already do have a lot of experience. They are well experienced with web applications already. The rate is quite high compared to other institutes.

Percentage of Work spent with applications Percentage of web applications involved
88% 57%

Terms and definitions[edit]

Issues

Indicate where users were hesitating, asking, put a comment or missed the right action at the first attempt. (A hint for User Interface Engineering where problems exist and if they can be related to wording, process or position). Issues are rated as Minor, Serious or Failed but do not count as 'failed' or 'ok'. They only provide hints to the interface engineering team how to judge severity of issues.

Failed

The task couldn't be completed with the interface. This is the case when users get lost within the interface, do something wrong or can not continue because something is missing.

OK

The task leads to a valid result. This is the case when no step shows a single 'Failed'. It does not mean that single steps can have issues of type serious or minor.

Analysis[edit]

Non Functional Tasks[edit]

Error creating thumbnail: /bin/bash: line 1: /usr/bin/convert: No such file or directory Error code: 127
Analysis Result

These tasks are performed as a warm up only and are usually carried out fine.

Browse & Display[edit]

Analysis Result

Summary[edit]

  • The filter dropdown is not recognized; participants focus on items in the list
  • Participants would have failed without filtering (there were only view items and the status was visible at the item)
  • Medium View was used rarely; Participants rather click on the title for detail view
  • Participants get stuck with numbers; Scan numbers and page numbers can not be distinguished; additional numbers in the dropdown (e.g. 123/56) and the filename (e.g. e00001) leads to confusion
  • The metadata section of structural elements can not be accessed by most users; participants keep their focus on the tree and do not search in the menu

View Item/Details[edit]

Analysis Result

Summary[edit]

  • Breadcrumb is not used; either the back button is used or they start over from "Inhaltsverzeichnis"
  • Participants hesitate quite long because most of them are in a different volume and the way up is not obvious
  • The way to the detail section of a structural element is still difficult even if it was found previously

Export and Miscellaneous[edit]

Analysis Result
Analysis Result

Summary[edit]

  • The maximum page size was not found by most users because participants didn't click on the image; most users searched in the menu or near the picture

Edit Book Meta Data[edit]

Analysis Result
These are less important tasks. They are performed to as a warm up for participants and are usually carried out flawless.

Summary[edit]

  • half of the participants were not able to find the edit mode; especially when the menu was in a state where no edit was offered
  • participants especially had problems when moving back to browse and then to edit mode again
  • users were not aware of the view mode and edit mode at all

Edit Table of Content[edit]

Analysis Result
These are less important tasks. They are performed to as a warm up for participants and are usually carried out flawless.

Summary[edit]

Measures taken from interviews do not lead necesserily to a solution of the problem in one go. Old issues need to be tested again. As R4 comes with a new interface it needs to be tested completely.

References[edit]