Heuristic Evaluation of Groupware. How to do Heuristic Evaluation of Groupware. Benefits

Size: px
Start display at page:

Download "Heuristic Evaluation of Groupware. How to do Heuristic Evaluation of Groupware. Benefits"

Transcription

1 Kimberly Tee CPSC 681 Topic Heuristic Evaluation of Groupware Heuristic evaluation [9] is a discount evaluation method for finding usability problems in a singleuser interface design. With this methodology, a small set of evaluators examine the interface and judge its compliance with recognized usability principles (the heuristics ). Evaluators describe each problem found, rate its severity, and can suggest possible fixes. In 2001, Kevin Baker adapted the heuristic evaluation methodology to find usability problems in real-time, shared workspace groupware systems [1]. Real-time, shared workspace groupware systems support real-time collaboration between distance-separated people through a shared workspace a medium sized flat work surface where people collaborate by manipulating visible tools and task artefacts [6]. Some examples of shared workspace groupware systems are shared editors and multi-player games. The activities in a shared workspace can be divided into taskwork and teamwork, where taskwork is the actual execution of a task (by a single user) and teamwork is the work of working together [6]. Thus, groupware usability is defined as the degree to which a system supports both taskwork (single-user usability) and teamwork [3]. How to do Heuristic Evaluation of Groupware Baker s adapted heuristic evaluation methodology uses a set of groupware heuristics (see Groupware Heuristics below) designed to evaluate shared workspace interfaces for teamwork usability problems. Like Nielsen s heuristic evaluation methodology, it consists of three stages: 1. Orientation: If required, evaluat ors are introduced to the heuristics and the methodology. They are given a short description of the system under test, and (optionally) a scenario to help them explore and evaluate the system. 2. Evaluation: Evaluators go through the interface several times and inspect specific interface elements for compliance with the list of heuristics. They describe each usability problem found, rate its severity, and can suggest possible fixes (see Problem Report section). 3. Debriefing: After all evaluations are complete, evaluators, observers, and representatives of the design team group together all duplicate problem reports. They discuss their findings, prioritize the problems they found, and suggest solutions. Heuristic evaluation of groupware may be used to evaluate medium or high-fidelity prototypes, or almost-finished products. Unlike traditional heuristic evaluation which can be done on low-fidelity prototypes, heuristic evaluation of groupware is restricted to systems that are partly functional. However, Baker suggests that the groupware heuristics can be used to guide the design of groupware so that many potential problems can be addressed prior to any formal evaluation [3]. Benefits is a discount evaluation method quick and cheap [2] requires a small number of evaluators (three to five evaluators will typically identify about 40-60% of the total known teamwork problems) [2] doesn t require end-users [9] most of the problems found will be major ones, and most of the missed problems will be minor ones [2] non-usability experts can use heuristic evaluation fairly successfully [2]

2 Problems is limited to real-time, shared workspace groupware systems [1] some problems will go unnoticed, particularly if few evaluators are used [2] since real users aren t involved, surprise problems relating to their needs are not found [9] some problems reported are not actually usability problems [9] there is quite a bit of redundancy in the problems found by evaluators (the ratio of consolidated problems to original raw problem reports is about 1:6) so results synthesis can take some effort [2] does not provide a systematic way to generate fixes to the usability problems or a way to assess the quality of any redesigns [9] Groupware Heuristics The following heuristics (based on the mechanics of collaboration framework [1, 3]) are designed to evaluate shared workspace groupware systems for teamwork usability problems. These heuristics are tailored for workspaces that are used by groups of approximately five people [3]: 1. Provide the means for intentional and appropriate verbal communication: the groupware system should provide a channel for intentional communication so that collaborators can effectively communicate with one another. 2. Provide the means for intentional and appropriate gestural communication: the groupware system should support intentional gestural communication by making all explicit gestures and other visual actions visible to all group members, as well as maintaining the association between the gesture/action and its author. The system should also display the gestures in relation to the objects or people they are referencing, when applicable. 3. Provide consequential communication of an individual s embodiment: the groupware system should capture and transmit the explicit and subtle visual and verbal cues that are given by participants during collaboration, such as locations, gazes and views, actions, presences, and identities. 4. Provide consequential communication of shared artifacts: the groupware system should transmit the local user's feedback to all remote users, and should indicate when an artifact is being manipulated to help users determine what others are doing to it and how it has changed. 5. Provide protection: the groupware system should enable users to concurrently interact with the shared virtual workspace with minimal conflicts. 6. Manage the transitions between tightly and loosely-coupled collaboration: the groupware system should allow each individual to navigate the workspace independently. The system should also enable smooth transitions between joint work and individual work. 7. Support people with the coordination of their actions: the groupware system should allow people to effectively coordinate their activities by providing awareness of others and their actions in the workspace, by providing context for the actions which take place, and by providing collaborators a means of communication. 8. Facilitate finding collaborators and establishing context: the groupware system should provide information on potential collaborators so that they can be easily found and their availability for group work can be determined. The system should allow collaborators to initiate contact with minimal effort.

3 Problem Report The following form (adapted from [3]) can be used by an evaluator to report a groupware usability problem: The heuristic violated: Provide the means for intentional and appropriate verbal communication Provide the means for intentional and appropriate gestural communication Provide consequential communication of an individual s embodiment Provide consequential communication of shared artifacts Provide protection Manage the transition between tightly and loosely-coupled collaboration Support people with the coordination of their actions Facilitate finding collaborators and establishing contact Other Description of usability problem: Severity rating: Major obstacle to effective collaboration Minor obstacle; people can work around it Solution (optional):

4 Example Notification Collage is a groupware system where members of a community can post media elements such as Sticky Notes and Photo Shows onto a real-time collaborative surface that all members can see. For this example, we will only be looking at the NC Sketch media element, which allows people to simultaneously draw on a shared canvas (figure below). We will not be conducting a full heuristic evaluation of NC Sketch since most of the process is the same as heuristic evaluation of single-user applications. Instead, we will concentrate on finding usability problems using the groupware heuristics. Example problem reports are on the next page. Exercise /Homework Find at least 3 more usability problems using the groupware heuristics before next class (so that Phuoc can use them in his Results Synthesis presentation). Instructions for downloading and connecting to the Notification Collage will be posted later.

5 Example Problem Reports for NC Sketch Problem Report #1 The heuristic violated: Provide the means for intentional and appropriate verbal communication Description of usability problem: The chat window isn t resizable so it s hard to follow conversations when lots of people are involved Severity rating: Major obstacle to effective collaboration Problem Report #2 The heuristic violated: Provide the means for intentional and appropriate gestural communication Description of usability problem: When two or more people have the same paint colour, it is impossible to see who is painting what Severity rating: Major obstacle to effective collaboration Problem Report #3 The heuristic violated: Provide consequential communication of an individual s embodiment Description of usability problem: When someone isn t drawing, can t tell if they have stopped work, are typing into the chat window, or are choosing a control Severity rating: Minor obstacle; people can work around it Problem Report #4 The heuristic violated: Provide protection Description of usability problem: No undo function; easy to lose sketch Severity rating: Major obstacle to effective collaboration Problem Report #5 The heuristic violated: Manage the transitions between tightly and loosely-coupled collaboration Description of usability problem: Can t work privately; all work is immediately visible to others Severity rating: Minor obstacle; people can work around it

6 Resources on Heuristic Evaluation of Groupware 1. Baker, K., Greenberg, S. and Gutwin, C. (2001). Heuristic Evaluation of Groupware Based on the Mechanics of Collaboration. I M. R. Little and L. Nigay (Eds.), Engineering for Human- Computer Interaction, LNCS 2254, London, UK: Springer-Verlag. This is the best resource for a quick overview of heuristic evaluation of groupware. This paper presents a set of groupware heuristics for identifying teamwork problems in shared workspace groupware systems. For each of the heuristics, it provides an explanation of how the heuristic was derived from the mechanics of collaboration framework [6] along with some examples of how groupware applications typically realize and support its criteria. This publication was based on research from [3]. 2. Baker, K., Greenberg, S. and Gutwin, C. (2002). Empirical Development of a Heuristic Evaluation Methodology for Shared Workspace Groupware. Proceedings of CSCW 2002, ACM Press, This paper justifi es that heuristic evaluation of groupware [1] is a discount evaluation method. It reports on a study where 27 evaluators used the groupware heuristics to evaluate two groupware applications. It was found that individual evaluators uncover around 20% of the total known teamwork problems and that three to five evaluators will find 40-60% of the total known teamwork problems. These results are similar to Nielsen s [7, 8]. It concludes that the groupware heuristics can be effectively applied in heuristic evaluation of groupware to identify teamwork usability problems in a low-cost and practical way. It also speculates that the results could be even better if evaluators had better training and more practice evaluating groupware. This publication was based on research from [3]. 3. Baker, K. (2002). Heuristic Evaluation of Shared Workspace Groupware based on the Mechanics of Collaboration. MSc Thesis, Department of Computer Science, University of Calgary, Calgary, Alberta, Canada. This is the best resource for an in-depth look at heuristic evaluation of groupware. Baker adapts Nielsen s heuristic evaluation methodology [9] for evaluating real-time, shared workspace groupware by developing a set of groupware heuristics based on Gutwin and Greenberg s mechanics of collaboration framework [6]. He also revises another set of groupware heuristics based on the Locales Framework [4]. To demonstrate that the adapted heuristic evaluation for groupware remains a discount usability technique, he shows that individuals who are not necessarily experts in groupware can learn and apply the heuristics to identify groupware usability problems in a study replicating Nielsen s original work [7, 8]. While inspectors were able to use the mechanics of collaboration heuristics with minimal expertise and training, the Locales Framework heuristics were found to be less successful in that they were more difficult to understand and thus more difficult to effectively use. [1, 2] are publications from this research. 4. Greenberg, S., Fitzpatrick, G., Gutwin, C. and Kaplan, S. (2000). Adapting the Locales Framework for Heuristic Evaluation of Groupware. Australian Journal of Information Systems (AJIS), 7(2), This paper describes a set of groupware heuristics based on the Locales framework [5]. It reports on a heuristic evaluation done on a commercial groupware system using these heuristics. The authors were able to identify groupware usability problems, but suggest that these groupware heuristics may require a greater level of evaluator training and experienc e (compared with the conventional heuristics, which tend to be easy to learn and apply) to use effectively. They also suggest that these heuristics are best used for evaluating general groupware environments rather than specific groupware applications.

7 Resources on Groupware Theory 5. Fitzpatrick, G. (2002). The Locales Framework: making social thinking accessible for software practitioners?. In C. Floyd, Y. Dittrich, and R. Klischewski (Eds.), Social Thinking - Software Practice, Cambridge, MA: MIT Press. This paper describes the Locales Framework, a sociological theory of how people interact, collaborate and perform work in a complex social environment. It includes discussion on the strengths and weaknesses of this framework. 6. Gutwin, C. and Greenberg, S. (2000). The Mechanics of Collaboration: Developing Low Cost Usability Evaluation Methods for Shared Workspaces. IEEE 9 th International Workshop on Enabling Technologies: Infrastructure for Collaborative Enterprises (WET-ICE 00). June 14-16, held at NIST, Gaithersburg, MD USA. Awarded Best Paper of the workshop. This paper suggests that a groupware system must support both taskwork and teamwork to be usable. It further breaks teamwork down into two areas: the social and affective elements that make up group dynamics, and the mechanics of collaboration. It identifies and describes seven activities that comprise the mechanics of collaboration and proposes evaluating groupware for the mechanics of collaboration by examining if a group can perform them effectively, efficiently, and pleasantly. It illustrates ways existing discount usability evaluation methodologies can incorporate this framework to uncover groupware usability problems. Resources on Heuristic Evaluation 7. Nielsen, J. and Molich, R. (1990). Heuristic evaluation of user interfaces. Proceedings of CHI 1990, ACM Press, This paper reports on four experiments run to evaluate how practical heuristic evaluation is. It was found that while individual evaluators only find 20-50% of the usability problems in the interfaces they evaluate, three to five evaluators can be expected to find at least two-thirds of the usability problems. 8. Nielsen, J. (1992). Finding usability problems through heuristic evaluation. Proceedings of CHI 1992, ACM Press, This paper extends previous work [7] evaluating how practical heuristic evaluation is. It suggests that usability specialists are better than non-specialists at performing heuristic evaluation, and usability specialists with expertise in the kind of interface being developed perform even better (so fewer of them can be used). It also discusses the types of usability problems that tend to be found and missed by evaluators. 9. Nielsen, J. (1993). Usability Engineering, & 224. Cambridge, MA: Academic Press. This text provides a useful overview of heuristic evaluation and justifies why it is a discount evaluation method. Examples are also provided.

CS 160: Evaluation. Professor John Canny Spring /15/2006 1

CS 160: Evaluation. Professor John Canny Spring /15/2006 1 CS 160: Evaluation Professor John Canny Spring 2006 2/15/2006 1 Outline User testing process Severity and Cost ratings Discount usability methods Heuristic evaluation HE vs. user testing 2/15/2006 2 Outline

More information

CS 160: Evaluation. Outline. Outline. Iterative Design. Preparing for a User Test. User Test

CS 160: Evaluation. Outline. Outline. Iterative Design. Preparing for a User Test. User Test CS 160: Evaluation Professor John Canny Spring 2006 2/15/2006 1 2/15/2006 2 Iterative Design Prototype low-fi paper, DENIM Design task analysis contextual inquiry scenarios sketching 2/15/2006 3 Evaluate

More information

How to Conduct a Heuristic Evaluation

How to Conduct a Heuristic Evaluation Page 1 of 9 useit.com Papers and Essays Heuristic Evaluation How to conduct a heuristic evaluation How to Conduct a Heuristic Evaluation by Jakob Nielsen Heuristic evaluation (Nielsen and Molich, 1990;

More information

Interaction Design. Heuristic Evaluation & Cognitive Walkthrough

Interaction Design. Heuristic Evaluation & Cognitive Walkthrough Interaction Design Heuristic Evaluation & Cognitive Walkthrough Interaction Design Iterative user centered design and development Requirements gathering Quick design Build prototype Evaluate and refine

More information

NPTEL Computer Science and Engineering Human-Computer Interaction

NPTEL Computer Science and Engineering Human-Computer Interaction M4 L5 Heuristic Evaluation Objective: To understand the process of Heuristic Evaluation.. To employ the ten principles for evaluating an interface. Introduction: Heuristics evaluation is s systematic process

More information

Assistant Professor Computer Science. Introduction to Human-Computer Interaction

Assistant Professor Computer Science. Introduction to Human-Computer Interaction CMSC434 Introduction to Human-Computer Interaction Week 12 Lecture 24 Nov 21, 2013 Intro to Evaluation Human Computer Interaction Laboratory @jonfroehlich Assistant Professor Computer Science Hall of Fame

More information

Heuristic Evaluation. Hall of Fame or Shame? Hall of Fame or Shame? Hall of Fame! Heuristic Evaluation

Heuristic Evaluation. Hall of Fame or Shame? Hall of Fame or Shame? Hall of Fame! Heuristic Evaluation 1 USER INTERFACE DESIGN + PROTOTYPING + EVALUATION Hall of Fame or Shame? Heuristic Evaluation Prof. James A. Landay University of Washington Pocket By Read It Later 11/1/2012 2 Hall of Fame or Shame?

More information

USER INTERFACE DESIGN + PROTOTYPING + EVALUATION. Heuristic Evaluation. Prof. James A. Landay University of Washington CSE 440

USER INTERFACE DESIGN + PROTOTYPING + EVALUATION. Heuristic Evaluation. Prof. James A. Landay University of Washington CSE 440 USER INTERFACE DESIGN + PROTOTYPING + EVALUATION Heuristic Evaluation Prof. James A. Landay University of Washington CSE 440 February 19, 2013 Hall of Fame or Shame? Pocket By Read It Later Jan. 14-18,

More information

User Experience Report: Heuristic Evaluation

User Experience Report: Heuristic Evaluation User Experience Report: Heuristic Evaluation 1 User Experience Report: Heuristic Evaluation Created by Peter Blair for partial fulfillment of the requirements for MichiganX: UX503x Principles of Designing

More information

Design Heuristics and Evaluation

Design Heuristics and Evaluation Design Heuristics and Evaluation Rapid Evaluation Selected material from The UX Book, Hartson & Pyla Heuristic Evaluation Another method for finding usability problems in a UI design Validation during

More information

Redesign Workshop: Involving Software Developers Actively in Usability Engineering

Redesign Workshop: Involving Software Developers Actively in Usability Engineering Redesign Workshop: Involving Software Developers Actively in Usability Engineering Nis Bornoe nis@cs.aau.dk Jane Billestrup jane@cs.aau.dk Jesper Lumbye Andersen TC Electronic Aarhus, Denmark jesperla@tcelectronic.com

More information

Nektarios Kostaras, Mixalis Xenos. Hellenic Open University, School of Sciences & Technology, Patras, Greece

Nektarios Kostaras, Mixalis Xenos. Hellenic Open University, School of Sciences & Technology, Patras, Greece Kostaras N., Xenos M., Assessing Educational Web-site Usability using Heuristic Evaluation Rules, 11th Panhellenic Conference on Informatics with international participation, Vol. B, pp. 543-550, 18-20

More information

Heuristic Evaluation! Hall of Fame or Shame?! Hall of Fame or Shame?! Hall of Fame or Shame?! Hall of Fame!!

Heuristic Evaluation! Hall of Fame or Shame?! Hall of Fame or Shame?! Hall of Fame or Shame?! Hall of Fame!! CS 147: HCI+D UI Design, Prototyping, and Evaluation, Autumn 2014 HCI+D: USER INTERFACE DESIGN + PROTOTYPING + EVALUATION Hall of Fame or Shame? Heuristic Evaluation Computer Science Department Autumn

More information

HCI and Design SPRING 2016

HCI and Design SPRING 2016 HCI and Design SPRING 2016 Topics for today Heuristic Evaluation 10 usability heuristics How to do heuristic evaluation Project planning and proposals Usability Testing Formal usability testing in a lab

More information

Additional reading for this lecture: Heuristic Evaluation by Jakob Nielsen. Read the first four bulleted articles, starting with How to conduct a

Additional reading for this lecture: Heuristic Evaluation by Jakob Nielsen. Read the first four bulleted articles, starting with How to conduct a Additional reading for this lecture: Heuristic Evaluation by Jakob Nielsen. Read the first four bulleted articles, starting with How to conduct a heuristic evaluation and ending with How to rate severity.

More information

Addition about Prototypes

Addition about Prototypes Vorlesung Mensch-Maschine-Interaktion Evaluation Ludwig-Maximilians-Universität München LFE Medieninformatik Heinrich Hußmann & Albrecht Schmidt WS2003/2004 http://www.medien.informatik.uni-muenchen.de/

More information

Assignment 5 is posted! Heuristic evaluation and AB testing. Heuristic Evaluation. Thursday: AB Testing

Assignment 5 is posted! Heuristic evaluation and AB testing. Heuristic Evaluation. Thursday: AB Testing HCI and Design Topics for today Assignment 5 is posted! Heuristic evaluation and AB testing Today: Heuristic Evaluation Thursday: AB Testing Formal Usability Testing Formal usability testing in a lab:

More information

CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018

CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018 CPSC 444 Project Milestone III: Prototyping & Experiment Design Feb 6, 2018 OVERVIEW... 2 SUMMARY OF MILESTONE III DELIVERABLES... 2 1. Blog Update #3 - Low-fidelity Prototyping & Cognitive Walkthrough,

More information

SFU CMPT week 11

SFU CMPT week 11 SFU CMPT-363 2004-2 week 11 Manuel Zahariev E-mail: manuelz@cs.sfu.ca Based on course material from Arthur Kirkpatrick, Alissa Antle and Paul Hibbits July 21, 2004 1 Analytic Methods Advantages can be

More information

iscreen Usability INTRODUCTION

iscreen Usability INTRODUCTION INTRODUCTION Context and motivation The College of IST recently installed an interactive kiosk called iscreen, designed to serve as an information resource for student/visitors to the College of IST. The

More information

Perfect Timing. Alejandra Pardo : Manager Andrew Emrazian : Testing Brant Nielsen : Design Eric Budd : Documentation

Perfect Timing. Alejandra Pardo : Manager Andrew Emrazian : Testing Brant Nielsen : Design Eric Budd : Documentation Perfect Timing Alejandra Pardo : Manager Andrew Emrazian : Testing Brant Nielsen : Design Eric Budd : Documentation Problem & Solution College students do their best to plan out their daily tasks, but

More information

Generating and Using Results

Generating and Using Results Background Generating and Using Results from Usability Evaluations Kasper Hornbæk University of Copenhagen www.kasperhornbaek.dk Associate professor in the Human computer Interaction group at Copenhagen

More information

User-Centered Design. SWE 432, Fall 2017 Design and Implementation of Software for the Web

User-Centered Design. SWE 432, Fall 2017 Design and Implementation of Software for the Web User-Centered Design SWE 432, Fall 2017 Design and Implementation of Software for the Web In class exercise As you come in and take a seat Write down at least 3 characteristics that makes something usable

More information

IPM 10/11 T1.6 Discount Evaluation Methods

IPM 10/11 T1.6 Discount Evaluation Methods IPM 10/11 T1.6 Discount Evaluation Methods Licenciatura em Ciência de Computadores Miguel Tavares Coimbra Acknowledgements: Most of this course is based on the excellent course offered by Prof. Kellogg

More information

Improve the User Experience on Your Website

Improve the User Experience on Your Website Forrester Consulting Approach Document Improve the User Experience on Your Website Improving the usability of your website will provide a better customer experience. By providing site visitors with a better

More information

User Centered Design And Prototyping

User Centered Design And Prototyping User Centered Design And Prototyping Why User Centered Design is important Approaches to User Centered Design Rapid prototype techniques The Design Of Well Crafted Tools The All Too Common Approach In

More information

Heuristic Evaluation. Heuristic evaluation evaluates the interface to identify usability problems against recognized usability design heuristics.

Heuristic Evaluation. Heuristic evaluation evaluates the interface to identify usability problems against recognized usability design heuristics. Heuristic Evaluation Heuristic evaluation evaluates the interface to identify usability problems against recognized usability design heuristics. Usability heuristics are best practices developed and identified

More information

White Paper. Incorporating Usability Experts with Your Software Development Lifecycle: Benefits and ROI Situated Research All Rights Reserved

White Paper. Incorporating Usability Experts with Your Software Development Lifecycle: Benefits and ROI Situated Research All Rights Reserved White Paper Incorporating Usability Experts with Your Software Development Lifecycle: Benefits and ROI 2018 Situated Research All Rights Reserved Learnability, efficiency, safety, effectiveness, memorability

More information

CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation

CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation Lecture 12: Inspection-Based Methods James Fogarty Daniel Epstein Brad Jacobson King Xia Tuesday/Thursday 10:30 to 11:50

More information

Cognitive Walkthrough

Cognitive Walkthrough 1 Cognitive Walkthrough C. Wharton, J. Rieman, C. Lewis and P. Polson, The Cognitive Walkthrough Method: A Practitioner s Guide, in J. Nielsen and R. Mack (eds.), Usability Inspection Methods, John Wiley

More information

cs465 principles of user interface design, implementation and evaluation

cs465 principles of user interface design, implementation and evaluation cs465 principles of user interface design, implementation and evaluation Karrie G. Karahalios 24. September 2008 1. Heuristic Evaluation 2. Cognitive Walkthrough 3. Discuss Homework 3 4. Discuss Projects

More information

Design, Ideation, and Prototyping

Design, Ideation, and Prototyping Design, Ideation, and Prototyping Why User Centered Design is important How Prototyping helps User Centered Design System Centered Design System Centered Design What can be built easily on this platform?

More information

Component-Based Technologies for End-User Development

Component-Based Technologies for End-User Development Paper to be published in Communications of the ACM, special issue on End-User Development, Sept. 2004 Component-Based Technologies for End-User Development Anders I. Mørch 1, Gunnar Stevens 2, Markus Won

More information

Overview of Today s Lecture. Analytical Evaluation / Usability Testing. ex: find a book at Amazon.ca via search

Overview of Today s Lecture. Analytical Evaluation / Usability Testing. ex: find a book at Amazon.ca via search Overview of Today s Lecture Analytical Evaluation / Usability Testing November 17, 2017 Analytical Evaluation Inspections Recapping cognitive walkthrough Heuristic evaluation Performance modelling 1 2

More information

Page 1. Ideas to windows. Lecture 7: Prototyping & Evaluation. Levels of prototyping. Progressive refinement

Page 1. Ideas to windows. Lecture 7: Prototyping & Evaluation. Levels of prototyping. Progressive refinement Ideas to windows Lecture 7: Prototyping & Evaluation How do we go from ideas to windows? Prototyping... rapid initial development, sketching & testing many designs to determine the best (few?) to continue

More information

Usability & User Centered Design. SWE 432, Fall 2018 Design and Implementation of Software for the Web

Usability & User Centered Design. SWE 432, Fall 2018 Design and Implementation of Software for the Web Usability & User Centered Design SWE 432, Fall 2018 Design and Implementation of Software for the Web Review: Mental models Only single temperature sensor. Controls not independent, need to adjust both.

More information

15/16 CSY2041 Quality and User-Centred Systems

15/16 CSY2041 Quality and User-Centred Systems 15/16 CSY2041 Quality and User-Centred Systems INTERACTION DESIGN 1 Heuristic evaluation and walkthroughs 2 1 Aims: Describe the key concepts associated with inspection methods. Explain how to do heuristic

More information

HEURISTIC EVALUATION WHY AND HOW

HEURISTIC EVALUATION WHY AND HOW HEURISTIC EVALUATION WHY AND HOW REF: Scott Klemmer Jacob Nielsen James Landay HEURISTIC EVALUATION Multiple ways to evaluate Empirical: Assess with real users trying s/w Formal: Models and formulas to

More information

CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation

CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation Lecture 11: Inspection Tuesday / Thursday 12:00 to 1:20 James Fogarty Kailey Chan Dhruv Jain Nigini Oliveira Chris Seeds

More information

1. Select/view stores based on product type/category- 2. Select/view stores based on store name-

1. Select/view stores based on product type/category- 2. Select/view stores based on store name- COMPETITIVE ANALYSIS Context The world of mobile computing is fast paced. There are many developers providing free and subscription based applications on websites like www.palmsource.com. Owners of portable

More information

Introducing Evaluation

Introducing Evaluation Chapter 12 Introducing Evaluation 1 The aims Explain the key concepts used in evaluation. Introduce different evaluation methods. Show how different methods are used for different purposes at different

More information

Hyper Mesh Code analyzer

Hyper Mesh Code analyzer Hyper Mesh Code analyzer ABSTRACT Hyper Mesh Code Analyzer (HMCA) is a text based programming environment, designed for programmers to write their source code in a more organized and manageable fashion.

More information

Evaluation of Commercial Web Engineering Processes

Evaluation of Commercial Web Engineering Processes Evaluation of Commercial Web Engineering Processes Andrew McDonald and Ray Welland Department of Computing Science, University of Glasgow, Glasgow, Scotland. G12 8QQ. {andrew, ray}@dcs.gla.ac.uk, http://www.dcs.gla.ac.uk/

More information

A Heuristic Evaluation of Ohiosci.org

A Heuristic Evaluation of Ohiosci.org A Heuristic Evaluation of Ohiosci.org Executive Summary Site evaluated: Goal: Method: The Ohio Academy of Science http://www.ohiosci.org/ The goal of this heuristic evaluation is to test the overall usability

More information

Heuristic evaluation is a usability inspection technique developed by Jakob Nielsen. The original set of heuristics was derived empirically from an

Heuristic evaluation is a usability inspection technique developed by Jakob Nielsen. The original set of heuristics was derived empirically from an Heuristic evaluation is a usability inspection technique developed by Jakob Nielsen. The original set of heuristics was derived empirically from an analysis of 249 usability problems (Nielsen, 1994). -Preece

More information

Jakob Nielsen s Heuristics (

Jakob Nielsen s Heuristics ( Jakob Nielsen s Heuristics (http://www.useit.com/papers/heuristic/heuristic_list.html) What is heuristic evaluation: Heuristic evaluation is done as a systematic inspection of a user interface design for

More information

IBM s approach. Ease of Use. Total user experience. UCD Principles - IBM. What is the distinction between ease of use and UCD? Total User Experience

IBM s approach. Ease of Use. Total user experience. UCD Principles - IBM. What is the distinction between ease of use and UCD? Total User Experience IBM s approach Total user experiences Ease of Use Total User Experience through Principles Processes and Tools Total User Experience Everything the user sees, hears, and touches Get Order Unpack Find Install

More information

GUI Design Principles

GUI Design Principles GUI Design Principles User Interfaces Are Hard to Design You are not the user Most software engineering is about communicating with other programmers UI is about communicating with users The user is always

More information

The Information Technology Program (ITS) Contents What is Information Technology?... 2

The Information Technology Program (ITS) Contents What is Information Technology?... 2 The Information Technology Program (ITS) Contents What is Information Technology?... 2 Program Objectives... 2 ITS Program Major... 3 Web Design & Development Sequence... 3 The Senior Sequence... 3 ITS

More information

User Interface Evaluation

User Interface Evaluation User Interface Evaluation Heuristic Evaluation Lecture #17 Agenda Evaluation through Expert Analysis Cognitive walkthrough Heuristic evaluation Model-based evaluation Cognitive dimension of notations 2

More information

Lecture 14: Heuristic Evaluation. Fall UI Design and Implementation 1

Lecture 14: Heuristic Evaluation. Fall UI Design and Implementation 1 Lecture 14: Heuristic Evaluation Fall 2006 6.831 UI Design and Implementation 1 1 UI Hall of Fame or Shame? Fall 2006 6.831 UI Design and Implementation 2 The next version of Microsoft Office (Office 2007)

More information

Heuristic Evaluation Report. The New York Philharmonic Digital Archives archives.nyphil.org

Heuristic Evaluation Report. The New York Philharmonic Digital Archives archives.nyphil.org Heuristic Evaluation Report The New York Philharmonic Digital Archives archives.nyphil.org Cassie Hickman Wednesday, October 14, 2015 Table of Contents Executive Summary... 3 Introduction... 4 Methodology...

More information

Evaluation and Design Issues of Nordic DC Metadata Creation Tool

Evaluation and Design Issues of Nordic DC Metadata Creation Tool Evaluation and Design Issues of Nordic DC Metadata Creation Tool Preben Hansen SICS Swedish Institute of computer Science Box 1264, SE-164 29 Kista, Sweden preben@sics.se Abstract This paper presents results

More information

Expert Reviews (1) Lecture 5-2: Usability Methods II. Usability Inspection Methods. Expert Reviews (2)

Expert Reviews (1) Lecture 5-2: Usability Methods II. Usability Inspection Methods. Expert Reviews (2) : Usability Methods II Heuristic Analysis Heuristics versus Testing Debate Some Common Heuristics Heuristic Evaluation Expert Reviews (1) Nielsen & Molich (1990) CHI Proceedings Based upon empirical article

More information

STUDY OF THE IMPACT OF THE RAPID PROTOTYPING METHOD ON THE PERFORMANCES OF A DESIGN PROCESS

STUDY OF THE IMPACT OF THE RAPID PROTOTYPING METHOD ON THE PERFORMANCES OF A DESIGN PROCESS STUDY OF THE IMPACT OF THE RAPID PROTOTYPING METHOD ON THE PERFORMANCES OF A DESIGN PROCESS Daniel-Constantin Anghel, Nadia Belu University of Pitesti, Romania KEYWORDS Rapid prototyping, DSM, design experiment,

More information

Adding Usability to Web Engineering Models and Tools

Adding Usability to Web Engineering Models and Tools Adding Usability to Web Engineering Models and Tools Richard Atterer 1 and Albrecht Schmidt 2 1 Media Informatics Group Ludwig-Maximilians-University Munich, Germany richard.atterer@ifi.lmu.de 2 Embedded

More information

Heuristic Evaluation of Covalence

Heuristic Evaluation of Covalence Heuristic Evaluation of Covalence Evaluator #A: Selina Her Evaluator #B: Ben-han Sung Evaluator #C: Giordano Jacuzzi 1. Problem Covalence is a concept-mapping tool that links images, text, and ideas to

More information

Ryan Parsons Chad Price Jia Reese Alex Vassallo

Ryan Parsons Chad Price Jia Reese Alex Vassallo Ryan Parsons - Paper Prototype, Writing Chad Price - Paper Prototype, Digital Mockup Jia Reese - Paper Prototype, Usability Testing Alex Vassallo - Usability Testing, Writing All we have to decide is what

More information

Lecture 14: Heuristic Evaluation

Lecture 14: Heuristic Evaluation Lecture 14: Heuristic Evaluation 1 1 U IH alloffam e or Sham e? Courtesy of Snapfish. Used with permission. 2 For today s UI Hall of Fame and Shame, we ll focus on the Rotate commands in photo browsers

More information

Web-based Interactive Support for Combining Contextual and Procedural. design knowledge

Web-based Interactive Support for Combining Contextual and Procedural. design knowledge Web-based Interactive Support for Combining Contextual and Procedural Design Knowledge J.-H. Lee & Z.-X. Chou Graduate School of Computational Design, NYUST, Touliu, Taiwan ABSTRACT: Design study can take

More information

Design Principles. Overview. User-Center Design. SMD157 Human-Computer Interaction Fall User-center design Guidelines

Design Principles. Overview. User-Center Design. SMD157 Human-Computer Interaction Fall User-center design Guidelines INSTITUTIONEN FÖR SYSTEMTEKNIK LULEÅ TEKNISKA UNIVERSITET Design Principles SMD157 Human-Computer Interaction Fall 2005 Nov-4-05 SMD157, Human-Computer Interaction 1 L Overview User-center design Guidelines

More information

Assignments. Assignment 2 is due TODAY, 11:59pm! Submit one per pair on Blackboard.

Assignments. Assignment 2 is due TODAY, 11:59pm! Submit one per pair on Blackboard. HCI and Design Assignments Assignment 2 is due TODAY, 11:59pm! Submit one per pair on Blackboard. Today Paper prototyping An essential tool in your design toolbox! How do we design things that actually

More information

Concept. Shared Editing. The notion of congruence. Types of shared editors. Master Recherche - Université Paris-Sud. M. Beaudouin-Lafon - Groupware 1

Concept. Shared Editing. The notion of congruence. Types of shared editors. Master Recherche - Université Paris-Sud. M. Beaudouin-Lafon - Groupware 1 Concept Shared Editing Michel Beaudouin-Lafon Collaborative creation and editing of shared computer artifacts - Typically a shared document - All users have the illusion that they edit the same document

More information

Design Principles. Overview. User-Center Design. SMD157 Human-Computer Interaction Fall User-center design Guidelines

Design Principles. Overview. User-Center Design. SMD157 Human-Computer Interaction Fall User-center design Guidelines INSTITUTIONEN FÖR SYSTEMTEKNIK LULEÅ TEKNISKA UNIVERSITET Design Principles SMD157 Human-Computer Interaction Fall 2003 Nov-6-03 SMD157, Human-Computer Interaction 1 L Overview User-center design Guidelines

More information

Lose It! Weight Loss App Heuristic Evaluation Report

Lose It! Weight Loss App Heuristic Evaluation Report Lose It! Weight Loss App Heuristic Evaluation Report By Manuel Ryan Espinsa Manuel Ryan Espinosa 1-27-2017 Heuristic Evaluation IN4MATX 283 LEC A: USER EXPERIENCE (37000) TABLE OF CONTENTS EXECUTIVE SUMMARY

More information

Improving user interfaces through a methodological heuristics evaluation framework and retrospective think aloud with eye tracking

Improving user interfaces through a methodological heuristics evaluation framework and retrospective think aloud with eye tracking Improving user interfaces through a methodological heuristics evaluation framework and retrospective think aloud with eye tracking Progress Report Supervisors: Dr. Tom Gedeon Mr. Christopher Chow Principal

More information

Artifacts as Instant Messenger Buddies 1

Artifacts as Instant Messenger Buddies 1 Artifacts as Instant Messenger Buddies Saul Greenberg, Nathan Stehr and Kimberly Tee Department of Computer Science, University of Calgary Calgary, Alberta CANADA T2N 1N4 saul.greenberg@ucalgary.ca ABSTRACT

More information

SkillSwap. A community of learners and teachers

SkillSwap. A community of learners and teachers Team: Jacob Yu Villa, Dana Murphy, Tuan Tran SkillSwap A community of learners and teachers Problem During our needfinding process, we found that many people felt discouraged about learning due to the

More information

Inspection Overview Massood Towhidnejad Computer & Software Engineering Dept. Embry-Riddle University

Inspection Overview Massood Towhidnejad Computer & Software Engineering Dept. Embry-Riddle University Inspection Overview Massood Towhidnejad Computer & Software Engineering Dept. Embry-Riddle University Inspection Is used to verify intellectual products by manually examining the developed product, a piece

More information

Challenges in evaluating complex IT security management systems

Challenges in evaluating complex IT security management systems Challenges in evaluating complex IT security management systems Pooya Jaferian, Kirstie Hawkey, Konstantin Beznosov University of British Columbia, Vancouver, Canada {pooya,hawkey,beznosov}@ece.ubc.ca

More information

Cindy Fan, Rick Huang, Maggie Liu, Ethan Zhang November 6, c: Usability Testing Check-In

Cindy Fan, Rick Huang, Maggie Liu, Ethan Zhang November 6, c: Usability Testing Check-In Cindy Fan, Rick Huang, Maggie Liu, Ethan Zhang November 6, 2014 3c: Usability Testing Check-In HEURISTIC EVALUATION Our group did two heuristic evaluations. For each issue discovered during evaluation,

More information

Evaluation techniques 1

Evaluation techniques 1 IMS5302 Human-computer interaction Lecture 6 Other Evaluation Techniques Overview Other evaluation methods Expert reviews Field studies Developing scenarios Selecting an evaluation method IMS5302 2 Scenarios

More information

EVALUATION OF PROTOTYPES USABILITY TESTING

EVALUATION OF PROTOTYPES USABILITY TESTING EVALUATION OF PROTOTYPES USABILITY TESTING CPSC 544 FUNDAMENTALS IN DESIGNING INTERACTIVE COMPUTATIONAL TECHNOLOGY FOR PEOPLE (HUMAN COMPUTER INTERACTION) WEEK 9 CLASS 17 Joanna McGrenere and Leila Aflatoony

More information

Evaluation techniques 1

Evaluation techniques 1 IMS5302 Human- computer interaction Lecture 6 Other Evaluation Techniques Overview Other evaluation methods Expert reviews Field studies Developing scenarios Selecting an evaluation method IMS5302 2 Scenarios

More information

PRAVTA A Light-Weight WAP Awareness Client

PRAVTA A Light-Weight WAP Awareness Client PRAVTA A Light-Weight WAP Awareness Client Tom Gross GMD German National Research Center for Information Technology (tom.gross@gmd.de) Abstract. Despite huge progress in information and communication technology

More information

Evaluation Plan for Fearless Fitness Center Website

Evaluation Plan for Fearless Fitness Center Website Evaluation Plan for Fearless Fitness Center Website A three-step process will be used to prepare the Fearless Fitness website for use. First, at the outset of the project, Back-of-the-Envelope Task Analysis

More information

Heuristic Evaluation of [ Quest ]

Heuristic Evaluation of [ Quest ] Heuristic Evaluation of [ Quest ] 1. Problem Quest is an app that allows you to stay involved in, participate in, and create local clubs and events happening in your community 2. Violations Found 1. [H10.

More information

3 Prototyping and Iterative Evaluations

3 Prototyping and Iterative Evaluations 3 Prototyping and Iterative Evaluations Viktoria Pammer-Schindler March 15, 2016 Prototyping and Iterative Evaluations 1 Days and Topics March 1 March 8 March 15 April 12 April 19/21 April 26 (10-13) April

More information

Combining Different Business Rules Technologies:A Rationalization

Combining Different Business Rules Technologies:A Rationalization A research and education initiative at the MIT Sloan School of Management Combining Different Business Rules Technologies:A Rationalization Paper 116 Benjamin Grosof Isabelle Rouvellou Lou Degenaro Hoi

More information

Interactive (High-fi) Prototype (Group)

Interactive (High-fi) Prototype (Group) Interactive (High-fi) Prototype (Group) Midway Milestone due at the start of your studio (Thursday/Friday Dec 1-2) Final Prototype due at the start of your studio (Thursday/Friday Dec 8-9) Writeup due

More information

Übung zur Vorlesung Mensch-Maschine-Interaktion. e5: Heuristic Evaluation

Übung zur Vorlesung Mensch-Maschine-Interaktion. e5: Heuristic Evaluation Übung zur Vorlesung Mensch-Maschine-Interaktion e5: Heuristic Evaluation Sara Streng Ludwig-Maximilians-Universität München Wintersemester 2007/2008 Ludwig-Maximilians-Universität München Sara Streng Übung

More information

EVALUATION OF THE USABILITY OF EDUCATIONAL WEB MEDIA: A CASE STUDY OF GROU.PS

EVALUATION OF THE USABILITY OF EDUCATIONAL WEB MEDIA: A CASE STUDY OF GROU.PS EVALUATION OF THE USABILITY OF EDUCATIONAL WEB MEDIA: A CASE STUDY OF GROU.PS Turgay Baş, Hakan Tüzün Hacettepe University (TURKEY) turgaybas@hacettepe.edu.tr, htuzun@hacettepe.edu.tr Abstract In this

More information

Business Analysis for Practitioners - Requirements Elicitation and Analysis (Domain 3)

Business Analysis for Practitioners - Requirements Elicitation and Analysis (Domain 3) Business Analysis for Practitioners - Requirements Elicitation and Analysis (Domain 3) COURSE STRUCTURE Introduction to Business Analysis Module 1 Needs Assessment Module 2 Business Analysis Planning Module

More information

8.1 Goals of Evaluation 8.2 Analytic Evaluation 8.3 Empirical Evaluation 8.4 Comparing and Choosing Evaluation Techniques

8.1 Goals of Evaluation 8.2 Analytic Evaluation 8.3 Empirical Evaluation 8.4 Comparing and Choosing Evaluation Techniques 8 Evaluation 8.1 Goals of Evaluation 8.2 Analytic Evaluation 8.3 Empirical Evaluation 8.4 Comparing and Choosing Evaluation Techniques Ludwig-Maximilians-Universität München Prof. Hußmann Mensch-Maschine-Interaktion

More information

Cognitive Walkthrough

Cognitive Walkthrough 1 1 Cognitive Walkthrough C. Wharton, J. Rieman, C. Lewis and P. Polson, The Cognitive Walkthrough Method: A Practitioner s Guide, in J. Nielsen and R. Mack (eds.), Usability Inspection Methods, John Wiley

More information

EVALUATION OF PROTOTYPES USABILITY TESTING

EVALUATION OF PROTOTYPES USABILITY TESTING EVALUATION OF PROTOTYPES USABILITY TESTING CPSC 544 FUNDAMENTALS IN DESIGNING INTERACTIVE COMPUTATION TECHNOLOGY FOR PEOPLE (HUMAN COMPUTER INTERACTION) WEEK 9 CLASS 17 Joanna McGrenere and Leila Aflatoony

More information

All LJMU programmes are delivered and assessed in English

All LJMU programmes are delivered and assessed in English PROGRAMME SPECIFICATION Master of Science in Computer Forensics Awarding institution Teaching institution JACS Code Programme Duration Language of Programme Subject benchmark statement Programme accredited

More information

APPLYING HUMAN FACTORS ENGINEERING TO IMPROVE USABILITY AND WORKFLOW IN PATHOLOGY INFORMATICS

APPLYING HUMAN FACTORS ENGINEERING TO IMPROVE USABILITY AND WORKFLOW IN PATHOLOGY INFORMATICS Proceedings of the 2017 International Symposium on Human Factors and Ergonomics in Health Care 23 APPLYING HUMAN FACTORS ENGINEERING TO IMPROVE USABILITY AND WORKFLOW IN PATHOLOGY INFORMATICS Austin F.

More information

THE USE OF PARTNERED USABILITY TESTING TO HELP TO IDENTIFY GAPS IN ONLINE WORK FLOW

THE USE OF PARTNERED USABILITY TESTING TO HELP TO IDENTIFY GAPS IN ONLINE WORK FLOW THE USE OF PARTNERED USABILITY TESTING TO HELP TO IDENTIFY GAPS IN ONLINE WORK FLOW Dianne Davis Fishbone Interactive Gordon Tait Department of Surgery, University of Toronto Cindy Bruce-Barrett Strategic

More information

3d: Usability Testing Review

3d: Usability Testing Review Balance Ryan Parsons, Chad Price, Jia Reese, Alex Vassallo 3d: Usability Testing Review Usability Test #1 Our first usability testing participant, Mary, was an undergraduate student in Computer Science

More information

Applying Usability to elearning

Applying Usability to elearning Applying Usability to elearning 6 December 08 John Wooden, PhD Director of Usability Services Fredrickson Communications jwooden@fredcomm.com We make information useful! About Fredrickson Communications

More information

Lecture 22: Heuristic Evaluation. UI Hall of Fame or Shame? Spring User Interface Design and Implementation 1

Lecture 22: Heuristic Evaluation. UI Hall of Fame or Shame? Spring User Interface Design and Implementation 1 Lecture 22: Heuristic Evaluation Spring 2008 6.831 User Interface Design and Implementation 1 UI Hall of Fame or Shame? From Shauni Deshmukh: Kayak.com is a website that allows people to search for flights.

More information

Improving Awareness with Remote Control Point Movement Prediction in Real-Time Collaborative Graphics Editing Systems

Improving Awareness with Remote Control Point Movement Prediction in Real-Time Collaborative Graphics Editing Systems IJCSNS International Journal of Computer Science and Network Security, VOL.6 No.3A, March 2006 203 Improving Awareness with Remote Control Point Movement Prediction in Real-Time Collaborative Graphics

More information

Cognitive Disability and Technology: Universal Design Considerations

Cognitive Disability and Technology: Universal Design Considerations Cognitive Disability and Technology: Universal Design Considerations Clayton Lewis Coleman Institute for Cognitive Disabilities RERC-ACT clayton.lewis@colorado.edu Prepared for AUCD Training Symposium,

More information

Current Issues in the Determination of Usability Test Sample Size: How Many Users is Enough?

Current Issues in the Determination of Usability Test Sample Size: How Many Users is Enough? Current Issues in the Determination of Usability Test Sample Size: How Many Users is Enough? Carl W. Turner (State Farm Insurance, carl.turner.hxyf@statefarm.com) Jakob Nielsen (Nielsen Norman Group, nielsen@nngroup.com)

More information

CMSC434 Intro to Human-Computer Interaction. Visual Design #3 and Evaluation #1 Monday, April 8th, 2012 Instructor: Jon Froehlich TA: Kotaro Hara

CMSC434 Intro to Human-Computer Interaction. Visual Design #3 and Evaluation #1 Monday, April 8th, 2012 Instructor: Jon Froehlich TA: Kotaro Hara CMSC434 Intro to Human-Computer Interaction Visual Design #3 and Evaluation #1 Monday, April 8th, 2012 Instructor: Jon Froehlich TA: Kotaro Hara #inspiration [Applied Sciences Group: High Performance Touch,

More information

Prototyping. Unit 5. Zeno Menestrina, MSc Prof. Antonella De Angeli, PhD

Prototyping. Unit 5. Zeno Menestrina, MSc Prof. Antonella De Angeli, PhD Prototyping Unit 5 Zeno Menestrina, MSc zeno.menestrina@unitn.it Prof. Antonella De Angeli, PhD antonella.deangeli@unitn.it Assessment Pervasive game to support social life in/from the campus 2 Assessment

More information

Published in: Proceedings of the Eighth Danish Human-Computer Interaction Research Symposium, 20th November 2008

Published in: Proceedings of the Eighth Danish Human-Computer Interaction Research Symposium, 20th November 2008 Downloaded from vbn.aau.dk on: marts 26, 2019 Aalborg Universitet The WPU Project Jensen, Janne Jul; Skov, Mikael B.; Stage, Jan Published in: Proceedings of the Eighth Danish Human-Computer Interaction

More information

What is a prototype?

What is a prototype? Prototyping Unit 4 Learning outcomes Understand the uses of different types of prototypes for different kinds/stages of design and be able to choose appropriately Know the basic techniques for low-fidelity

More information

University of Maryland. fzzj, basili, Empirical studies (Desurvire, 1994) (Jeries, Miller, USABILITY INSPECTION

University of Maryland. fzzj, basili, Empirical studies (Desurvire, 1994) (Jeries, Miller, USABILITY INSPECTION AN EMPIRICAL STUDY OF PERSPECTIVE-BASED USABILITY INSPECTION Zhijun Zhang, Victor Basili, and Ben Shneiderman Department of Computer Science University of Maryland College Park, MD 20742, USA fzzj, basili,

More information