Upload
vincci-kwong
View
2.368
Download
0
Embed Size (px)
Citation preview
October 20, 2009
Web Site Usability
Vincci KwongFranklin D. Schurz Library
Indiana University South BendIndiana Library Federation Annual Conference
Website Usability
“The extent to which a product can be used by specified users to achieve specified goals with effectiveness, efficiency and satisfaction in a specified context of use.” (British Standards Institution, 1998)
“Making whatever you’re working on easier to use for who ever is going to use it.” (Steve Krug, 2001)
“An element of design, focusing on ‘can this be used when it’s done?’ rather than just making it look good.” (Steve Krug, 2001)
“On the Web, usability is a necessary condition for survival. If a website is difficult to use, people leave” (Jakob Nielsen, 2003)
Components of Website Usability
The five Es:
• Effective• Efficient• Engaging• Error tolerant• Easy to learn
Types of Usability Testing
• Scenario-Based Inspection• Heuristic Evaluation• User Observation• Prototyping Test• Card Sorting
Perquisites for all Usability Testing
• Information gathering• Who are the users?• What do users do?• What do users want?• Limitation of current system
• Understanding of design principles and rules
Scenario-Based Inspection
• Evaluator examines the website in which people perform their tasks
Goal
Task Task Task
Action Action Action Action
Setting Up a Scenario-Based Inspection
1. Come up with scenarios
2. Explore different ways in order to accomplish each scenario
3. Make note on problems encounter when working through each scenario
Heuristic Evaluation
• Inspector examines the website to check whether it complies with a set of design principles
- Nielsen’s Heuristics- MILE+ Heuristics
- Triacca, Luca, Alessandro Inversini, and Davide Bolchini. “Quality of Web Usability Evaluation Methods: An Empirical Study on MiLE+.” Seventh IEEE International Symposium on Web Site Evolution. 2005. 22-29. Print.
- W3C Content Accessibility Guidelines 1.0- Section 508 Standards
Setting Up a Heuristic Inspection
1. Choose inspectors
2. Decide on heuristic criteria
3. Prepare data collection and analysis form
4. Prepare task description
5. Reserve a room
Task Scenario No.: 1Inspector’s Name: KellySession Date: 3/20/2009Session Start Time: 9:00 amSession End Time: 10:00 am
Heuristic violated Usability defect Comments
Navigation Inconsistent overall navigation
The user was confused by the different navigation systems.
Semiotics Unclear links labels The user would like simple wordings without the use of jargon.
Graphics Contrast between text and background
The user found it difficult to read the wordings on the webpage.
Fig. 1 An example of data collection and analysis form.
Pro and Con of Heuristic EvaluationPros
• Less expensive• Inspectors usually suggest solutions to the problem
they identified• Helps to reduce obvious errors
Cons• Inspector’s view points may not reflect those of actual
users• Inspector may have their own preferences, biases,
and views• Results depends on skills and experiences of
inspectors
User Observations
• User carries out specific tasks while under observation
Setting Up a User Observation Session
1. Choose participants
2. Select facilitator and observers (if any)
3. Prepare task description
4. Create evaluation script
5. Prepare permission form
6. Decide on how to record session
7. Pick a location
Choose participants
• Real user Vs Representative user• How many participants?• Availability of participant
Select facilitators and observers
• Facilitator- Maintains a welcome and relaxed
atmosphere- Keeps participant talking- Ensure the purpose of evaluation is fulfilled
• Observer
- Observe testing session
Prepare task description
• Decide on what function to test• Come up with scenarios
• 1 scenario per function• Elaborate scenarios
• Clear wordings• Document task description
• Task card
Create evaluation script
• A script used by evaluator to guide participant throughout the testing session
• Popular components of evaluation script- Welcome participant- Explain the evaluation and confidentiality- Perform preliminary interview- Administer tasks- Conduct exit interview- Thank the participant
Prepare permission form
• Participate Consent Form- Permission to record- Permission for use of data- Permission to share test results
- Nondisclosure agreements
Recording a session
• Paper Note• Audio recording• Video recording• Eye tracking
Our Experience with User Observation Session
Choose Users
• 4 faculty members• 4 students• 2 staff
Prepare task description
• Each user completed five tasks1. You have a citation for a journal article that you are interested in and
you want to find out if the library has a copy of the journal article.
2. You are working on a research related to economic recession and you want to find some books that provide information on it.
3. You are working on a term paper and you need to cite resources used in your paper, however you have no idea on how to do citation. You want to find out if the library provides information on how to do citation.
4. You just recognized two of the books you checked out from the library are overdue, you want to find out how much you own in the fines.
5. You have some photos that you want to resize, so you want to find out if Adobe Photoshop is available on the computers at the library.
Create evaluation script
1. Introduction• Welcome and explain the purpose of the usability test
2. Preliminary Interview• How much time do you normally spend on the Web in a given week?
• Have you used any library website to look up information in the past?
3. Evaluation Instruction
4. Task
5. Wrap Up & Brainstorm• Additional questions you would like to ask
• Invite user to provide any suggestion/comment for improvement
• Express thanks to user
Decide on how to record data
• Record session through video and screen capture
- Laptop
- Webcam
- Microphone
- Camtasia Studio
Testing is over, now what?
Data analysis and Interpretation
1. Review and summarize data• Quantitative data• Qualitative data
2. Group findings
3. Assign Severities
4. Write evaluation report
Review and summarize data
• Quantitative data
- Easy to understand
- More objective than qualitative data
• Qualitative data- Provide insights into
cause of problemo Labels and headings
not intuitiveo Crowded, clustered,
clumsyo Nothing stand out to
me
Group findings
• Chronological order• Severity of defect• Type of issue• Difficulty of fix
Assigning Severities• Helps to
prioritize the work list
• Severities scale varies
Severity Rating Issue3 Unclear link labels [Semiotics]
Inconsistent overall navigation [Navigation]Information overload [Cognitive]Inefficient search function [Others]
Insufficient system reaction to errors for a user [Technology/Performance]
Unmatched/unexpected information [Content]2 Contrast between text and background [Graphics]
Backward navigation [Navigation]Duplicate Information [Navigation]Item not in category as expected [Navigation]
Image information [Graphics]Scripting errors [Technology/Performance]Video is not able to accommodate all users [Others]
1 Grouping of left navigation bar [Navigation]Visited vs unvisited states [Navigation]Inflexible layout [Graphics]Uncertainty on currency of information [Content]Pop-up windows [Others]Unexpected file format [Others]
Write evaluation report
• Document what you did• Serves as a communication tool
Summing up our experience
• Pre-test play around• Different task sets• Wordings of task description
Helpful Readings
• Stone, Debbie, et al. User Interface Design and Evaluation. Boston: Elsevier, 2005.
• Krug, Steve. Don’t Make Me Think. Indianapolis: New Riders, 2000.
• Nielsen, Jakob, and Marie Tahir. Homepage Usability: 50 Websites Deconstructed. Indianapolis: New Riders, 2002.
• Carole, George. User-Centered Library Websites. Oxford: Chandos, 2008.
Helpful Resources
• Paper Prototyping a How-To Video. Fremont: Nielsen Norman Group, 2003.
• http://cardsorting.pbworks.com/ by Ellyssa Kroski
Questions?
Feel free to contact me at• Email: [email protected]• AIM: himffy• Yahoo: vincci_kwong• MSN: [email protected]• Phone: 574-520-4444