390 likes | 404 Views
Explore statistical analysis, A/B testing, standardized usability questionnaires, and usability testing to enhance product usability and user satisfaction. Includes case studies, sampling methods, and planning guidelines.
E N D
Statistical User Research IPw Jenny Huang, PraisePop
Outline • Introduction • Statistical Analysis for A/B Testing • Standardized Usability Questionnaires • Usability Testing – Plan • Usability Testing – Data Collection • Usability Testing – Graphical Display • Usability Case Studies
Quantifying User research • “User research is the systematic study of goals, needs, and capabilities of users so as to specify design, construction, or improvement of tools to benefit how users work and live.” – Schumacher, Ph.D., User Researcher
Usability Testing • Usability – “the extent to which a product can be used by specified users to achieve specified goals with effectiveness, efficiency, and satisfaction in a specified context of use” • Things to Measure • Completion Rates • Task Time • Errors • Satisfaction Ratings
SampLing • Sample Size • Though it is common belief that we need large samples to use statistics to interpret data, in user research, a small sample can still inform design decisions. • Stratified Sampling • It is more important to choose a small sample representative of the target market than a large sample that does not represent it.
A/B Testing • Method for comparing alternate designs • Users randomly work with one of two deployed design alternatives and the data differences are analyzed to tell whether they are statistically significant • Report button designs • Goal: To find a button design that makes reporting clear, encourages users to filter negative posts, and does not reduce enjoyability of the app • Random Name Generator button designs • Goal: To find a random name generator that encourages users to click it and also use the name it gives • Different menu designs • Goal: To find a menu design that is easily navigable and engaging for users • “Was this post sweet or salty?” (pop-up vs. no pop-up) • Goal: To determine whether the benefit of reducing negativity outweighs the annoyance users experience from the pop-up
Chapter 5 of Quantifying the User Experience • “Is the difference between designs greater than what we’d expect from chance?” • A/B Testing • Overview • t-test (comparing means), chi-square test of independence (comparing proportions) • Find p-value • Determine if statistically significant
Chapter 8 of Quantifying the User Experience • Standardized Questionnaire – form designed to assess participants’ satisfaction with the perceived usability of products • SUS • developed in the mid 1980s to become a popular questionnaire • 10 items, five scale steps, odd numbered items have a positive tone while the even numbered items have a negative tone • shown to be most sensitive post-task questionnaire
Planning Usability Tests Week 5
Essential Elements • Scope • Purpose • Schedule & Location • Sessions • Equipment • Participants • Scenarios • Metrics • Quantitative Metrics (Planning guidelines fromusability.gov.)
Scope • Indicate what you are testing: Give the name of the Web site, Web application, or other product. Specify how much of the product the test will cover. • We are testing PraisePop v1.1 (updated June 2015). The test will cover navigation and usability.
PurPose Identify the concerns, questions, and goals for this test. These can be quite broad or quite specific. In each round of testing, you will probably have several general and several specific concerns to focus on. Your concerns should drive the scenarios you choose for the usability. • Can users easily scroll through the feed and read the posts? • Can users easily find how to write a post? • Can users easily write a post to a specific person? • Can users easily send a post to the feed? • Can users easily pop the kernel? • Can users easily report a negative post? • Can users easily find how to delete posts about themselves or one they wrote? • Can users easily find the time and date a post was written? • Can users easily respond with the pre-set messages on the post? • Can users easily export posts to social media, email, and messages? • Can users easily find the settings? • Can users easily sign up? • Can users easily confirm their emails? • Can users easily sign in? • Can users easily log out? • Do users see that they can swipe right to see stats about the community feed? • Do users know how to user the random name banner? • Do users know how to use the random name generator button?
Schedule & Location • Indicate when and where you will do the test. If you have the schedule set, you may want to be specific about how many sessions you will hold in a day and exactly what times the sessions will be. • Library (5/5): 5-5:30pm, 5:40-6:10pm, 6:20-6:50pm • Day Hall Dorm (5/5): 9:30-10:00pm, 10:10-10:40pm
Sessions • You will want to describe the sessions, the length of the sessions. • Sessions will occur at the specified times, asking users to fill out a background questionnaire, do the tasks, and afterwards, fill out the SUS.
Equipment • Indicate the type of equipment you will be using in the test; desktop, laptop, mobile/Smartphone. • questionnaires and pencil • Phone with PraisePop app • Laptop to record data
Participants • Indicate the number and types of participants to be tested you will be recruiting. Describe how these participants were or will be recruited. • 5 participants • Stratified Sampling • at least one participant from each grade • at least 3 males and 3 females • Convenience Sampling • In the library and dorm, I will just find people there who are there and willing to participate
Scenarios • Indicate the number and types of tasks included in testing. Typically, you should end up with approximately 8 (+/- 2) scenarios for a mobile/smartphone test. You may want to include more in the test plan so the team can choose the appropriate tasks. • Write a post to John Palfrey complimenting him on his ASM on Wednesday. • Delete the post you wrote. • Upvote a post. • Report a post. • Find the time and date a post was written. • Respond to a post in the feed with a pre-set message. • Export a post to iMessage. • Find your and the community’s stats in the feed. • Find the rules of PraisePop. • Find two ways to choose a random name of a person to write about. • Log out of the account.
Subjective Metrics • Include the questions you are going to ask the participants prior to the sessions (e.g., background questionnaire), after each task scenario is completed (ease and satisfaction questions about the task), and overall ease, satisfaction and likelihood to use/recommend questions when the sessions is completed. • Background Questionnaire • Name? • Grade? • Gender? • Experience using PraisePop? • iPhone or Android phone?
Quantitative Metrics • Indicate the quantitative data you will be measuring in your test (e.g., successful completion rates, error rates, time on task). • I measure: • Task time • Error rates (error = 1, no error = 0) • Participant writes down: • Completion rates (success = 1, fail = 0) • Satisfaction ratings • “It was easy to complete this task.” (1-5) • At the end: • SUS questionnaire
My Directions • Explain the test: “Hi, my name is Jenny and I am part of the PraisePop team. Today, you will be participating in a usability test that will help us understand how to make PraisePop better. You will follow certain tasks that I ask you to do and rate your satisfaction. While you are completing tasks, I will be collecting data like task times and error rates. To begin, please fill out this questionnaire.” • Give participant questionnaire (page 1 of Questionnaires) • “Are you ready for your first task?” • Give tasks to complete and time them and count the errors • After each task, have them rate their satisfaction on the scale (page 2) • At the end, have them fill out the SUS (page 3) • Thank them for their time!
Results • Five participants (2 males, 3 females; 4 seniors, 1 upper) • Four of the users had never used PraisePopbefore • The five SUS scores obtained were: 72.5, 95, 87.5, 97.5, 97.5. • These give a sample mean of 90. • This score actually ranks in the A+ range of the Suaro/Lewis SUS grading curve. Yay PraisePop!! • Bias • Convenience sampling • Sample not representative of population
Usability case Studies Week 8