iscreen Usability INTRODUCTION
|
|
- Shawn Tucker
- 5 years ago
- Views:
Transcription
1 INTRODUCTION Context and motivation The College of IST recently installed an interactive kiosk called iscreen, designed to serve as an information resource for student/visitors to the College of IST. The content is designed to answer some of the common questions students/visitors might have, for example: faculty office locations, building hours, etc. The hardware comprises of a touch-screen capable 42 Plasma monitor, featuring a walk-up-and-use interface. The interface needs to support novice as well as repeat users in terms of content and accessibility. A user-based evaluation and an expertbased cognitive walkthrough of the interface revealed issues that could hinder usability of the iscreen. In this report, we use heuristic evaluation (Nielsen 1994) to investigate usability issues. Each of the methods mentioned addresses different aspects of usability. Using multiple approaches for studying usability would enable us to identify problems at various levels. Goal and findings Our goal is to study the interface using heuristics or guidelines to find usability problems that would generate recommendations/design implications for the (re)design of the information kiosk, the iscreen. The findings reveal that there are a number of things that violate one of more heuristics. The majority of the problems encountered relate more to interface elements like icons and feedback. We also found that the choice of information presentation/layout may lead to some confusion for the user in finding/processing the information requested. METHODOLOGY Heuristic evaluation involves using a small set of evaluators to examine an interface and judge it s compliance with a set of recognized usability heuristics or principles (Nielsen 1994). We used Jakob Nielsen s (1994) ten usability heuristics seen in table 1 to conduct the heuristic evaluation. H 1. Visibility of system status The system should always keep users informed about what is going on, through appropriate feedback within reasonable time. H 2. Match between system and the real world The system should speak the users' language, with words, phrases and concepts familiar to the user, rather than systemoriented terms. Follow real-world conventions, making information appear in a natural and logical order. H 3. User control and freedom Users often choose system functions by mistake and will need a clearly marked "emergency exit" to leave the unwanted state without having to go through an extended dialogue. Support undo and redo. H 4. Consistency and standards Users should not have to wonder whether different words, situations, or actions mean the same thing. Follow platform conventions. 1
2 H 5. Error prevention Even better than good error messages is a careful design which prevents a problem from occurring in the first place. Either eliminate error-prone conditions or check for them and present users with a confirmation option before they commit to the action. H 6. Recognition rather than recall Minimize the user's memory load by making objects, actions, and options visible. The user should not have to remember information from one part of the dialogue to another. Instructions for use of the system should be visible or easily retrievable whenever appropriate. H 7. Flexibility and efficiency of use Accelerators -- unseen by the novice user -- may often speed up the interaction for the expert user such that the system can cater to both inexperienced and experienced users. Allow users to tailor frequent actions. H 8. Aesthetic and minimalist design Dialogues should not contain information which is irrelevant or rarely needed. Every extra unit of information in a dialogue competes with the relevant units of information and diminishes their relative visibility. H 9. Help users recognize, diagnose, and recover from errors Error messages should be expressed in plain language (no codes), precisely indicate the problem, and constructively suggest a solution. H 10. Help and documentation Even though it is better if the system can be used without documentation, it may be necessary to provide help and documentation. Any such information should be easy to search, focused on the user's task, list concrete steps to be carried out, and not be too large. Table 1. Ten Usability Heuristics Although it is recommended to use three to five evaluators for heuristic evaluation in our situation we only had two evaluators. We each conducted the heuristic evaluation first individually. We each reviewed the ten usability heuristics, then explored and inspected the interface at least three times with the list of the ten usability heuristics in front of us. Each evaluator noted discovered problems on a form (Appendix A). For each problem recorded the evaluator also recorded what heuristic the problem violated and where the problem occurred in the system. After each evaluator had conducted the heuristic evaluation individually we came together to discuss the problems we found. We discussed each problem and assigned each problem a severity rating from 0-4 based on the ratings seen below. 2
3 0 No problem 1 Cosmetic problem 2 Minor usability problem 3 Major usability problem; important to fix 4 Usability catastrophe; imperative to fix Table 2. Severity Ratings After each problem was assigned a severity rating we discussed possible solutions for each usability problem. The problems found during the heuristic evaluation can be seen in table 3 below. No. Problem Description Heuristic Violated Places(s) where problem occurs Severity Rating Touch screen to begin not visible at all times. No affordance regarding purpose/content of the screen. Users cannot search for a person by name to find their phone number or office number. The user must know first whether someone is faculty or staff (or that there are phone numbers in Building Info) in order to find their phone number or office location. 4 No icons to supplement text in the buttons Information displayed does not give any aid to the user on how to use the system. No indication if the user is on the main screen, especially when other screens provide a button "Main". No permanent navigation bar to let the user know about the other information categories. System does not provide any help to the user to help them choose between faculty and staff. Phone numbers are shown in extension format only. Some pages do not present any additional information after selection has been made. Main menu button is not on all pages. Therefore the users must wait for the screen to re-enter the inert phase in order to choose another option. system status Inert Screen 4 H* Lack of Affordance Inert screen 4 H10 Help and documentation H2 Match between system and the real world H6 Recognition Faculty / Staff / Phone Numbers 3 rather than recall All screens 3 Main / Directory / H8 Aesthetic and Building info / IST minimalist design calendar / Course H10 Help and schedule / documentation Academic FAQ 3 system status All screens 3 system status All screens 3 H5 Error prevention H6 Recognition rather than recall Directory 3 H2 Match between system Staff / faculty / and the real world telephone numbers 3 H8 Aesthetic and minimalist design Main / Directory / Building info / IST calendar / Course schedule / Academic FAQ 3 H3 User control and freedom Fall Schedule 2 3
4 No Problem Description The user cannot tell where he/she is in the system. There are no titles on the pages. Some pages say iscreen at the top and others do not. Abbreviations and acronyms are used that people may not understand Position and size of "touch screen to begin" icon not standardized. No feedback to communicate a selection has been made (Buttons inactive) Too many names presented together inhibit the ability of the user to find information quickly. Does not inform the use when the i-screen would revert to its default state. The screen is not visually divided to depict categories. Links for the "graduate fall term" not consistent. User may not know if its clickable. Similarity of transitions between screens during the inert state and active state are misleading. There is a screen that only presents one option to the user. When the user selects course schedule they are presented with a blank screen with one button for the Fall course schedule. Heuristic Violated Places(s) where problem occurs Severity Rating system status All screens 2 H2 Match between system and the real world Phone numbers 2 H4 Consistency and standard Inert screen 2 system status All screens 2 H8 Aesthetic and minimalist design Faculty / Staff 2 system status All screens 2 Telephone numbers / H8 Aesthetic and Directory / fall minimalist design courses 2 Graduate / fall courses / academic calendar. 2 H4 Consistency and standard system status H4 Consistency and standard All screens 1 H8 Aesthetic and minimalist design Course schedule 1 Table 3. Heuristic Evaluation Problems The second problem listed in the table did not violate one of the Nielsen s ten known heuristics rather it is specific to the system we evaluated. We noted this as H* Lack of Affordance because there are no affordances that inform the user of the purpose or content of the system. The design implications that we discussed are included in the following section. DESIGN IMPLICATIONS We will discuss the design implications based on the results of the heuristic evaluation in terms of the organization of the information in the iscreen. We will first discuss design implications for the system in its inert phase and then for the system as a whole. The order in which we discuss the problems and their recommendations indicates the severity of the problem. The most severe problems and the recommendations will be discussed first. 4
5 Inert system state The touch screen to begin text and accompanying hand should always be visible from a distance in addition to always being in the same place on the screen. Problem 1 and 14: Touch screen to begin is not visible at all times. The inert screens should contain information regarding what type of information the system contains. This could be accomplished by using the content or the main menu buttons as the inert screens, for example one of the inert screens could display the advising walk in hours or the faculty and staff directory. Problem 2: The purpose of the iscreen and its content are not readily apparent. It is also not apparent that the iscreen may be touched. System wide design implications A search function should be added to the directory in order to allow users to search for a faculty of staff member without knowing whether they are faculty of staff. Another possible solution is to simply display faculty and staff information on the same screen and distinguish between faulty or staff by supplying a title or differentiating by color. Problem 3 and 8: Users must know the position a person hold in order to find their phone number and office location. System does not provide any help to the user to aid them in choosing between faculty and staff. All phone numbers should be displayed in the directory. Problem 3: Some phone numbers are included in Building info. Icons should be added to supplement the text only buttons. Icons with labels may be most useful. Problem 4: The main menu buttons are plain text. 5
6 A help option should be added to the system or descriptions of what each button will lead you to may be helpful, for example if there was text that appeared that read click here (and pointed to the directory button) if you would like to find information about faculty and staff. Problem 5: Information displayed does not give any aid to the user on how to use the system. A title should be added to the main menu screen that reads Main Menu. All other screens should also be given appropriate titles. Problems 6 and 12: The system does not inform the user of where they are in the system. A permanent navigation bar should always be displayed to allow for easier navigation. Problem 7: There is also no permanent navigation the user must return to the main menu to navigate to another part of the system. The main menu button is not on all pages. Phone numbers should be displayed in full format, rather than just the extension, for example (555) Problem 9: Phone numbers are displayed in extension format. Any unnecessary pages should not be included in the design or pages that currently only present one button should provide additional information. Problem 10 and 21: Some pages do not present any additional information after selection has been made. Abbreviations should not be used. Everything should be fully spelled out. Problem 13: Abbreviations and acronyms are used that people may not understand. 6
7 Buttons should depress when they are touched or make an appropriate clicking sound to give users feedback that the button was actually pressed. Problem 15: The main menu buttons and all other system buttons do not provide any feedback to users to communicate that a selection has been made. Text on screens should be at least a minimum font size and there should be a limit to how much is displayed. There should be minimum spacing between lines. Problem 16 and 18: It is difficult read screens when there is too much text presented. There should be some mechanism for informing the users that the screen will return to its inert phase after a period of time. This could be done by asking the user to make another selection. Problem 17: Users are also not informed of when the system will revert back to the inert phase, therefore the screen a user is viewing simply disappears after a number of second without any notice. The names in the directory should be spaced out in order to facilitate browsing and easier searching. The text size may need to be increased to allow for easily finding a name. Problem 18: The screen is not visually divided to depict categories. All links should be standardized. Buttons should be used throughout the system instead of switching between buttons and simply text. Problem 19: Links are not standardized throughout the system. Transitions between the inert screens and active screens should be differentiated in some way. Problem 20: Transitions between screens during the inert phase and active state are the same and therefore misleading. 7
8 Appendix A Heuristic Evaluation Data Form Problem description: Heuristic(s): violated: Average severity rating: Place(s) where problem occurs: Problem description: Heuristic(s): violated: Average severity rating: Place(s) where problem occurs: Problem description: Heuristic(s): violated: Average severity rating: Place(s) where problem occurs: Problem description: Heuristic(s): violated: Average severity rating: Place(s) where problem occurs: 8
9 References Nielsen, J. (1994). Heuristic evaluation. In Nielsen, J., and Mack, R.L. (Eds.), Usability Inspection Methods, John Wiley & Sons, New York, NY. 9
1. Select/view stores based on product type/category- 2. Select/view stores based on store name-
COMPETITIVE ANALYSIS Context The world of mobile computing is fast paced. There are many developers providing free and subscription based applications on websites like www.palmsource.com. Owners of portable
More informationJakob Nielsen s Heuristics (
Jakob Nielsen s Heuristics (http://www.useit.com/papers/heuristic/heuristic_list.html) What is heuristic evaluation: Heuristic evaluation is done as a systematic inspection of a user interface design for
More informationUser Interface Evaluation
User Interface Evaluation Heuristic Evaluation Lecture #17 Agenda Evaluation through Expert Analysis Cognitive walkthrough Heuristic evaluation Model-based evaluation Cognitive dimension of notations 2
More informationHeuristic Evaluation. Jon Kolko Professor, Austin Center for Design
Heuristic Evaluation Jon Kolko Professor, Austin Center for Design Heuristic Evaluation Compare an interface to an established list of heuristics best practices to identify usability problems. Heuristic
More informationCraigslist Heuristic Evaluation
Heuristic Evaluation Introduction This document is a heuristic evaluation of part of the website. Due to the size of the website, this is a partial examination ed to showcase Al Strauss analytical capabilities.
More informationINTRODUCTION USER POPULATION
iscreen Usability Cognitive Walkthrough Report Christine Wania George Abraham INTRODUCTION Context and motivation The College of IST recently installed an interactive kiosk called iscreen, designed to
More informationUser Experience Report: Heuristic Evaluation
User Experience Report: Heuristic Evaluation 1 User Experience Report: Heuristic Evaluation Created by Peter Blair for partial fulfillment of the requirements for MichiganX: UX503x Principles of Designing
More informationComputer Systems & Application
For updated version, please click on http://ocw.ump.edu.my Computer Systems & Application Computer System and Application Development Process By Mr. Nor Azhar Ahmad Faculty of Computer Systems & Software
More informationAssignment 5 is posted! Heuristic evaluation and AB testing. Heuristic Evaluation. Thursday: AB Testing
HCI and Design Topics for today Assignment 5 is posted! Heuristic evaluation and AB testing Today: Heuristic Evaluation Thursday: AB Testing Formal Usability Testing Formal usability testing in a lab:
More informationHCI and Design SPRING 2016
HCI and Design SPRING 2016 Topics for today Heuristic Evaluation 10 usability heuristics How to do heuristic evaluation Project planning and proposals Usability Testing Formal usability testing in a lab
More informationHyper Mesh Code analyzer
Hyper Mesh Code analyzer ABSTRACT Hyper Mesh Code Analyzer (HMCA) is a text based programming environment, designed for programmers to write their source code in a more organized and manageable fashion.
More informationCO328- Human Computer Interaction Michael Kölling Caroline Li. Heuristic Evaluation
CO328- Human Computer Interaction Michael Kölling Caroline Li Heuristic Evaluation Signage: What does this tells you? - History, what went earlier there before - Tells you more about a problematic situation
More informationHeuristic Evaluation
Heuristic Evaluation For CS4760 & HU4628 Group 3 -- BAER Can Li (canli@mtu.edu) 2/17/2015 Background: The primary mission of BAER Teams is to protect lives, property and sensitive habitats from post-fire
More informationCSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation
CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation Lecture 12: Inspection-Based Methods James Fogarty Daniel Epstein Brad Jacobson King Xia Tuesday/Thursday 10:30 to 11:50
More informationCSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation
CSE 440: Introduction to HCI User Interface Design, Prototyping, and Evaluation Lecture 11: Inspection Tuesday / Thursday 12:00 to 1:20 James Fogarty Kailey Chan Dhruv Jain Nigini Oliveira Chris Seeds
More informationDesign Heuristics and Evaluation
Design Heuristics and Evaluation Rapid Evaluation Selected material from The UX Book, Hartson & Pyla Heuristic Evaluation Another method for finding usability problems in a UI design Validation during
More informationDue on: May 12, Team Members: Arpan Bhattacharya. Collin Breslin. Thkeya Smith. INFO (Spring 2013): Human-Computer Interaction
Week 6 Assignment: Heuristic Evaluation of Due on: May 12 2013 Team Members: Arpan Bhattacharya Collin Breslin Thkeya Smith INFO 608-902 (Spring 2013): Human-Computer Interaction Group 1 HE Process Overview
More informationLose It! Weight Loss App Heuristic Evaluation Report
Lose It! Weight Loss App Heuristic Evaluation Report By Manuel Ryan Espinsa Manuel Ryan Espinosa 1-27-2017 Heuristic Evaluation IN4MATX 283 LEC A: USER EXPERIENCE (37000) TABLE OF CONTENTS EXECUTIVE SUMMARY
More informationUsability. Daniela Rosner. Web Architecture, October 9, School of Information UC Berkeley
Usability Daniela Rosner Web Architecture, 290-03 October 9, 2007 School of Information UC Berkeley Outline Introduction what is usability Best Practices common solutions Design Patterns shared languages
More informationHeuristic Evaluation of NUIG Participate Module 1
Heuristic Evaluation of NUIG Participate Module 1 Nielsen s 10 Usability Heuristics (Nielsen & Mack, 1994) 1. Aesthetic & Minimalist Design Pages should not contain information which is irrelevant or rarely
More informationIntroduction to Internet Applications
to Internet Applications Internet Applications, ID1354 1 / 36 Contents 2 / 36 Section 3 / 36 Local Application We are familiar with an architecture where the entire application resides on the same computer.
More informationHeuristic Evaluation. Ananda Gunawardena. Carnegie Mellon University Computer Science Department Fall 2008
Heuristic Evaluation Ananda Gunawardena Carnegie Mellon University Computer Science Department Fall 2008 Background Heuristic evaluation is performed early in the development process to understand user
More informationCrab Shack Kitchen Web Application
Crab Shack Kitchen Web Application EVALUATION ASSIGNMENT 2 HEURISTIC EVALUATION Author: Sachin FERNANDES Graduate 8 Undergraduate Team 2 Instructor: Dr. Robert PASTEL February 16, 2016 LIST OF FIGURES
More informationHeuristic Evaluation of Enable Ireland
Heuristic Evaluation of Enable Ireland Aesthetic and minimalist design Pages should not contain information which is irrelevant or rarely needed. Currently, there is too much content on the Enable Ireland
More informationUsability in Multimedia. By Pınar Koçer Aydın and Özgür Bayram
Usability in Multimedia By Pınar Koçer Aydın and Özgür Bayram 1 OUTLINE: Part 1: What is usability and what it isn t Part 2: Jakob Nielsen s 10 principles Part 3: The user experience testing Part 4: Application
More informationGroup #: Evaluation Proposal For Digital Libraries, LIS 5472, Fall 2010, Professor Sanghee Oh
Group #: Evaluation Proposal For Digital Libraries, LIS 5472, Fall 2010, Professor Sanghee Oh Background constraints. Group 3 worked within set background constraints which may or may not apply to other
More informationAssistant Professor Computer Science. Introduction to Human-Computer Interaction
CMSC434 Introduction to Human-Computer Interaction Week 12 Lecture 24 Nov 21, 2013 Intro to Evaluation Human Computer Interaction Laboratory @jonfroehlich Assistant Professor Computer Science Hall of Fame
More informationUX DESIGN BY JULIA MITELMAN
UX DESIGN BY JULIA MITELMAN TOPICS Why should we care? Usability Heuristics It s all about Context The Visual Applied Psychology The Pursuit of Product Design WHY SHOULD WE CARE? (IT S ALWAYS THE DESIGNER
More information1. The Best Practices Section < >
DRAFT A Review of the Current Status of the Best Practices Project Website and a Proposal for Website Expansion August 25, 2009 Submitted by: ASTDD Best Practices Project I. Current Web Status A. The Front
More informationCogSysIII Lecture 9: User Modeling with GOMS
CogSysIII Lecture 9: User Modeling with GOMS Human Computer Interaction Ute Schmid Applied Computer Science, University of Bamberg last change June 26, 2007 CogSysIII Lecture 9: User Modeling with GOMS
More informationUsability Inspection Methods. Overview. Usability Measures. SMD157 Human-Computer Interaction Fall 2003
INSTITUTIONEN FÖR SYSTEMTEKNIK LULEÅ TEKNISKA UNIVERSITET Usability Inspection Methods SMD157 Human-Computer Interaction Fall 2003 Nov-20-03 SMD157, Usability Inspection Methods 1 L Overview Usability
More informationSoftware Quality. Martin Glinz. Thomas Fritz. Lecture 7 UI Design, Usability & Testing. Many thanks to Meghan Allen and Daniel Greenblatt.
Institut für Informatik Software Quality Lecture 7 UI Design, Usability & Testing Thomas Fritz Martin Glinz Many thanks to Meghan Allen and Daniel Greenblatt. Overview Introduction to UI design User-centered
More informationCS 160: Evaluation. Outline. Outline. Iterative Design. Preparing for a User Test. User Test
CS 160: Evaluation Professor John Canny Spring 2006 2/15/2006 1 2/15/2006 2 Iterative Design Prototype low-fi paper, DENIM Design task analysis contextual inquiry scenarios sketching 2/15/2006 3 Evaluate
More informationCS 160: Evaluation. Professor John Canny Spring /15/2006 1
CS 160: Evaluation Professor John Canny Spring 2006 2/15/2006 1 Outline User testing process Severity and Cost ratings Discount usability methods Heuristic evaluation HE vs. user testing 2/15/2006 2 Outline
More informationcs465 principles of user interface design, implementation and evaluation
cs465 principles of user interface design, implementation and evaluation Karrie G. Karahalios 24. September 2008 1. Heuristic Evaluation 2. Cognitive Walkthrough 3. Discuss Homework 3 4. Discuss Projects
More informationUsability analysis and inspection
Usability analysis and inspection Why and how? 1MD113 Why? Iterative design Prototyping Measuring usability Objective/subjective feedback Quick and dirty Slow and clean With or without users 1 Evaluation
More informationHeuristic Evaluation of [Slaptitude]
Heuristic Evaluation of [Slaptitude] 1. Problem I am evaluating Slaptitude, a mobile app that allows you to set a timer and monitor leaderboards to help achieve and improve focus. 2. Violations Found 1.
More informationWikitude Usability testing and heuristic evaluation
Wikitude Usability testing and heuristic evaluation O. Perutka a a Faculty of Information Technology, CTU, Prague, Czech Republic Abstract Since augmented reality mobile applications depend on surrounding
More informationWhy? Usability analysis and inspection. Evaluation. Evaluation. Measuring usability. Evaluating usability
Usability analysis and inspection Why and how? Iterative design Prototyping Measuring usability Why? Objective/subjective feedback Quick and dirty Slow and clean With or without users 1MD113 Evaluation
More informationHeuristic Evaluation. Hall of Fame or Shame? Hall of Fame or Shame? Hall of Fame! Heuristic Evaluation
1 USER INTERFACE DESIGN + PROTOTYPING + EVALUATION Hall of Fame or Shame? Heuristic Evaluation Prof. James A. Landay University of Washington Pocket By Read It Later 11/1/2012 2 Hall of Fame or Shame?
More informationA Heuristic Evaluation of Ohiosci.org
A Heuristic Evaluation of Ohiosci.org Executive Summary Site evaluated: Goal: Method: The Ohio Academy of Science http://www.ohiosci.org/ The goal of this heuristic evaluation is to test the overall usability
More informationUSER INTERFACE DESIGN + PROTOTYPING + EVALUATION. Heuristic Evaluation. Prof. James A. Landay University of Washington CSE 440
USER INTERFACE DESIGN + PROTOTYPING + EVALUATION Heuristic Evaluation Prof. James A. Landay University of Washington CSE 440 February 19, 2013 Hall of Fame or Shame? Pocket By Read It Later Jan. 14-18,
More informationHeuristic Evaluation. Heuristic evaluation evaluates the interface to identify usability problems against recognized usability design heuristics.
Heuristic Evaluation Heuristic evaluation evaluates the interface to identify usability problems against recognized usability design heuristics. Usability heuristics are best practices developed and identified
More informationHeuristic evaluation is a usability inspection technique developed by Jakob Nielsen. The original set of heuristics was derived empirically from an
Heuristic evaluation is a usability inspection technique developed by Jakob Nielsen. The original set of heuristics was derived empirically from an analysis of 249 usability problems (Nielsen, 1994). -Preece
More informationUser Experience Design
User Experience Design PRESENTED BY Morgan Bibbs Director of Creative Services J. William Fulbright College of Arts & Sciences John C. Dailey, Ph.D. Content Strategist University Relations WHAT IS USER
More informationE2: Heuristic Evaluation A usability analysis of decorativethings.com. Jordana Carlin LIS Spring 2014
E2: Heuristic Evaluation A usability analysis of decorativethings.com Jordana Carlin LIS-644-01 Spring 2014 2 E2: HEURISTIC EVALUATION Executive Summary Decorative Things is an online retailer of unique
More information1. Problem Mix connects people who are interested in meeting new people over similar interests and activities.
1. Problem Mix connects people who are interested in meeting new people over similar interests and activities. 2. Violations Found 1. [H2 1 Visibility of Status][Severity 2][Found by: A] The interface
More informationHeuristic Evaluation of igetyou
Heuristic Evaluation of igetyou 1. Problem i get you is a social platform for people to share their own, or read and respond to others stories, with the goal of creating more understanding about living
More informationInteraction Design. Heuristic Evaluation & Cognitive Walkthrough
Interaction Design Heuristic Evaluation & Cognitive Walkthrough Interaction Design Iterative user centered design and development Requirements gathering Quick design Build prototype Evaluate and refine
More informationApplying Usability to elearning
Applying Usability to elearning 6 December 08 John Wooden, PhD Director of Usability Services Fredrickson Communications jwooden@fredcomm.com We make information useful! About Fredrickson Communications
More informationHeuristic Evaluation of Covalence
Heuristic Evaluation of Covalence Evaluator #A: Selina Her Evaluator #B: Ben-han Sung Evaluator #C: Giordano Jacuzzi 1. Problem Covalence is a concept-mapping tool that links images, text, and ideas to
More informationUser Experience Research Report: Heuristic Evaluation
User Experience Research Report: Heuristic Evaluation SI 622 003 Group 3: Yunqi Hu, Diane Pham, Chieh-Lin Wu, Ruofan Zhang Date: March 31, 2016 Word Count: 2,610 Table of Contents Executive Summary...
More informationChapter 10 Interactive Systems And Usability Organizational Requirements Engineering
Chapter 10 Interactive Systems And Usability Organizational Requirements Engineering Prof. Dr. Armin B. Cremers Sascha Alda Overview Introduction: What is usability? Why is usability an important non-functional
More informationHeuristic Evaluation! Hall of Fame or Shame?! Hall of Fame or Shame?! Hall of Fame or Shame?! Hall of Fame!!
CS 147: HCI+D UI Design, Prototyping, and Evaluation, Autumn 2014 HCI+D: USER INTERFACE DESIGN + PROTOTYPING + EVALUATION Hall of Fame or Shame? Heuristic Evaluation Computer Science Department Autumn
More informationHEURISTIC EVALUATION WHY AND HOW
HEURISTIC EVALUATION WHY AND HOW REF: Scott Klemmer Jacob Nielsen James Landay HEURISTIC EVALUATION Multiple ways to evaluate Empirical: Assess with real users trying s/w Formal: Models and formulas to
More informationSFU CMPT week 11
SFU CMPT-363 2004-2 week 11 Manuel Zahariev E-mail: manuelz@cs.sfu.ca Based on course material from Arthur Kirkpatrick, Alissa Antle and Paul Hibbits July 21, 2004 1 Analytic Methods Advantages can be
More informationHeuristic Evaluation of MPowered Entrepreneurship s Slack Workspace. Presented By Dylan Rabin. SI 110 Section 003
Heuristic Evaluation of MPowered Entrepreneurship s Slack Workspace Presented By Dylan Rabin SI 110 Section 003 You have so little time to attract their attention every little thing counts. Stewart Butterfield
More informationHeuristic Evaluation of PLATELIST
1. Problem Heuristic Evaluation of PLATELIST https://platelist.proto.io/share/?id=5793e1ea-5fd2-4f9c-9af9-4f745e2e30f2&v=1 This is an evaluation of Platelist, a mobile application that aims to facilitate
More informationHeuristic Evaluation of Math Out of the Box
Heuristic Evaluation of Math Out of the Box Evaluator #A: Joanie Hollberg Evaluator #B: Rassan Walker Evaluator #C: Alex Wang Evaluator #D: Carlos Araujo 1. Problem Math Out of The Box is a mobile application
More informationHeuristic Evaluation of [ Quest ]
Heuristic Evaluation of [ Quest ] 1. Problem Quest is an app that allows you to stay involved in, participate in, and create local clubs and events happening in your community 2. Violations Found 1. [H10.
More informationNPTEL Computer Science and Engineering Human-Computer Interaction
M4 L5 Heuristic Evaluation Objective: To understand the process of Heuristic Evaluation.. To employ the ten principles for evaluating an interface. Introduction: Heuristics evaluation is s systematic process
More informationUsability & User Centered Design. SWE 432, Fall 2018 Design and Implementation of Software for the Web
Usability & User Centered Design SWE 432, Fall 2018 Design and Implementation of Software for the Web Review: Mental models Only single temperature sensor. Controls not independent, need to adjust both.
More informationEvaluation in Information Visualization. An Introduction to Information Visualization Techniques for Exploring Large Database. Jing Yang Fall 2005
An Introduction to Information Visualization Techniques for Exploring Large Database Jing Yang Fall 2005 1 Evaluation in Information Visualization Class 3 2 1 Motivation What are the advantages and limitations
More informationDesign Principles. Overview. User-Center Design. SMD157 Human-Computer Interaction Fall User-center design Guidelines
INSTITUTIONEN FÖR SYSTEMTEKNIK LULEÅ TEKNISKA UNIVERSITET Design Principles SMD157 Human-Computer Interaction Fall 2005 Nov-4-05 SMD157, Human-Computer Interaction 1 L Overview User-center design Guidelines
More informationUser-Centered Design. SWE 432, Fall 2017 Design and Implementation of Software for the Web
User-Centered Design SWE 432, Fall 2017 Design and Implementation of Software for the Web In class exercise As you come in and take a seat Write down at least 3 characteristics that makes something usable
More informationDesign Principles. Overview. User-Center Design. SMD157 Human-Computer Interaction Fall User-center design Guidelines
INSTITUTIONEN FÖR SYSTEMTEKNIK LULEÅ TEKNISKA UNIVERSITET Design Principles SMD157 Human-Computer Interaction Fall 2003 Nov-6-03 SMD157, Human-Computer Interaction 1 L Overview User-center design Guidelines
More informationCMSC434 Intro to Human-Computer Interaction. Visual Design #3 and Evaluation #1 Monday, April 8th, 2012 Instructor: Jon Froehlich TA: Kotaro Hara
CMSC434 Intro to Human-Computer Interaction Visual Design #3 and Evaluation #1 Monday, April 8th, 2012 Instructor: Jon Froehlich TA: Kotaro Hara #inspiration [Applied Sciences Group: High Performance Touch,
More informationHeuristic Evaluation Google Play Store
Heuristic Evaluation Google Play Store Daniel J. Black April 2016 Contents EXECUTIVE SUMMARY... 3 INTRODUCTION... 4 Product Information... 4 Product Description... 4 Target Population... 4 HEURISTIC EVALUATION
More informationNektarios Kostaras, Mixalis Xenos. Hellenic Open University, School of Sciences & Technology, Patras, Greece
Kostaras N., Xenos M., Assessing Educational Web-site Usability using Heuristic Evaluation Rules, 11th Panhellenic Conference on Informatics with international participation, Vol. B, pp. 543-550, 18-20
More informationÜbung zur Vorlesung Mensch-Maschine-Interaktion. e5: Heuristic Evaluation
Übung zur Vorlesung Mensch-Maschine-Interaktion e5: Heuristic Evaluation Sara Streng Ludwig-Maximilians-Universität München Wintersemester 2007/2008 Ludwig-Maximilians-Universität München Sara Streng Übung
More informationevaluation techniques goals of evaluation evaluation by experts cisc3650 human-computer interaction spring 2012 lecture # II.1
topics: evaluation techniques usability testing references: cisc3650 human-computer interaction spring 2012 lecture # II.1 evaluation techniques Human-Computer Interaction, by Alan Dix, Janet Finlay, Gregory
More informationSeverity Definitions:
Heuristic Evaluation: Death by Hypothermia Overall, this project does a good job keeping the user interface simple. It is not difficult to go from one place to another within the application, but navigation
More informationEvaluation of Interactive Systems. Inspection methods
Evaluation of Interactive Systems Inspection methods Caroline Appert - 2018/2019 Inspection Methods Walkthroughs Norman and Nielsen s usability heuristics Green and Blackwell s cognitive dimensions Predictive
More informationAnalytical Evaluation
Analytical Evaluation November 7, 2016 1 Questions? 2 Overview of Today s Lecture Analytical Evaluation Inspections Performance modelling 3 Analytical Evaluations Evaluations without involving users 4
More informationIntroducing Evaluation
Chapter 12 Introducing Evaluation 1 The aims Explain the key concepts used in evaluation. Introduce different evaluation methods. Show how different methods are used for different purposes at different
More informationProperty of Shree Dhavale. Not for public distribution. Practicum Report. Shree Dhavale
Practicum Report By Shree Dhavale Submitted in fulfillment of the HIT 5000: Practicum in Applied Health Informatics Shree Dhavale Student Signature Faculty: Susan H. Fenton, PhD, RHIA, FAHIMA Associate
More informationactivated is a platform that allows students to create, organize, and share the steps on their road to college.
Heuristic Evaluation of [ activated ] 1. Problem activated is a platform that allows students to create, organize, and share the steps on their road to college. 2. Violations Found 1. H7: Flexibility and
More information15/16 CSY2041 Quality and User-Centred Systems
15/16 CSY2041 Quality and User-Centred Systems INTERACTION DESIGN 1 Heuristic evaluation and walkthroughs 2 1 Aims: Describe the key concepts associated with inspection methods. Explain how to do heuristic
More informationCognitive Walkthrough
1 Cognitive Walkthrough C. Wharton, J. Rieman, C. Lewis and P. Polson, The Cognitive Walkthrough Method: A Practitioner s Guide, in J. Nielsen and R. Mack (eds.), Usability Inspection Methods, John Wiley
More informationASSIGNMENT 06: Heuristic Evaluation
Group 4 Heuristic Evaluation 1 ASSIGNMENT 06: Heuristic Evaluation SI 622: Evaluation of Systems & Services Prof. Mark Newman, Winter 2012 March 15, 2012 Group 4: Charles Adams Jane Davis Nina Elias Liu
More informationCS 147 Autumn 2017: Assignment 9 (Heuristic Evaluation Group Template) Instructor: James Landay. Fix: make web also block the desktop screen.
Fix: make web also block the desktop screen. 5. H10 Help and documentation / Severity: 2 / Found by: A On the web calendar, there aren t any detailed instructions on how to engage in self-care activities,
More informationIPM 10/11 T1.6 Discount Evaluation Methods
IPM 10/11 T1.6 Discount Evaluation Methods Licenciatura em Ciência de Computadores Miguel Tavares Coimbra Acknowledgements: Most of this course is based on the excellent course offered by Prof. Kellogg
More informationHeuristic Evaluation as A Usability Engineering Method
Department of Computer Science Heuristic Evaluation as A Usability Engineering Method Prepared by: YASER GHANAM Revised by: SAUL GREENBERG Oct 2007 Table of Contents Table of Contents... 2 Table of Figures...
More informationIntroduction to Usability and its evaluation
Introduction to Usability and its evaluation ~ Greg Ralph What is usability? Attributes ISO 9241-11 (1998) The extent to which a product can be used: by specified users to achieve specified goals with
More informationNeon Carrot Prototype I Evaluation. Dan Cody, Logan Dethrow, Ben Fisher, Jeff Stanton April 6, Preamble
Neon Carrot Prototype I Evaluation Dan Cody, Logan Dethrow, Ben Fisher, Jeff Stanton April 6, 2009 Preamble Overall, we were impressed with the prototype's visual style, although some interface elements
More informationLearnability of software
Learnability of software Aim: Evaluating learnability Writing on-line help Basis for Assignment 2 1 Qualities of IT applications Learnability From novice to expert user Time from first encounter to use
More informationHeuristic Evaluation
Heuristic Evaluation Assignment 11: HE of Prototypes (Individual) PROBLEM PlateList is a mobile application designed to help people overcome small obstacles when trying to cook by allowing users to (1)
More informationUSER RESEARCH Website portfolio prototype
USER RESEARCH Website portfolio prototype Researcher & Author: Álvaro Ibargüen Villa UX, UI & Visual Designer Tel. E-mail Online +34 695 42 17 92 alvaroiv1@gmail.com aivweb.es INTRODUCTION 2 OBJECTIVES
More informationUsability. HCI - Human Computer Interaction
Usability HCI - Human Computer Interaction Computer systems optimization for easy access and communication Definition design Basic principles Testing assessment implementation Quality Utility funcionality
More informationPage 1. Ideas to windows. Lecture 7: Prototyping & Evaluation. Levels of prototyping. Progressive refinement
Ideas to windows Lecture 7: Prototyping & Evaluation How do we go from ideas to windows? Prototyping... rapid initial development, sketching & testing many designs to determine the best (few?) to continue
More informationPROJECT 1. Heuristic Evaluation and Cognitive Walkthrough of Goroo.com
PROJECT 1. Heuristic Evaluation and Cognitive Walkthrough of Goroo.com Cherese Cooper, Tatiana Iegorova, Andrew Wasowicz HCI 460 Usability Evaluations Spring 2013 1 Executive Summary In this report we
More informationDesign Reviews. Scott Klemmer. stanford hci group / cs147. tas: Marcello Bastea-Forte, Joel Brandt, Neil Patel, Leslie Wu, Mike Cammarano
stanford hci group / cs147 Design Reviews Scott Klemmer tas: Marcello Bastea-Forte, Joel Brandt, Neil Patel, Leslie Wu, Mike Cammarano 25 October 2007 http://cs147.stanford.edu Design Applied Psychology
More informationCognitive Walkthrough
1 1 Cognitive Walkthrough C. Wharton, J. Rieman, C. Lewis and P. Polson, The Cognitive Walkthrough Method: A Practitioner s Guide, in J. Nielsen and R. Mack (eds.), Usability Inspection Methods, John Wiley
More informationHeuristic Evaluation of Team Betamax
Heuristic Evaluation of Team Betamax Eric Gallimore Connor Riley Becky Scholl Chris Stone November 4, 2006 Overview Evaluation Let s just state for the record that we like this a whole lot better than
More informationLecture 14: Heuristic Evaluation. Fall UI Design and Implementation 1
Lecture 14: Heuristic Evaluation Fall 2006 6.831 UI Design and Implementation 1 1 UI Hall of Fame or Shame? Fall 2006 6.831 UI Design and Implementation 2 The next version of Microsoft Office (Office 2007)
More informationOverview of Today s Lecture. Analytical Evaluation / Usability Testing. ex: find a book at Amazon.ca via search
Overview of Today s Lecture Analytical Evaluation / Usability Testing November 17, 2017 Analytical Evaluation Inspections Recapping cognitive walkthrough Heuristic evaluation Performance modelling 1 2
More informationHCI CA1 (ii) Redesign Implementation
HCI CA1 (ii) Redesign Implementation A redesign of Waitron KatedeBoeAgnew N00100104 Overview Review of the Research Original Redesign Ideas Focus Group #1 Feedback from Focus Group #1 Redesign #1 Focus
More informationHeuristic Evaluation of [Pass It On]
Heuristic Evaluation of [Pass It On] Evaluator #A: Janette Evaluator #B: John Evaluator #C: Pascal Evaluator #D: Eric 1. Problem Pass It On aims to transform some of the numerous negative and stressful
More informationWhat is interaction design? What is Interaction Design? Example of bad and good design. Goals of interaction design
What is interaction design? What is Interaction Design? Designing interactive products to support people in their everyday and working lives Sharp, Rogers and Preece (2002) The design of spaces for human
More informationAnalytical &! Empirical Evaluation
Analytical &! Empirical Evaluation Informatics 132 5/22/2012 TODAY Evaluation Due: A3 Paper Prototyping UPCOMING Friday: Group Project Time Monday: Memorial Day, No Class Wednesday: HCI in the Real World
More information