UX Research Term

System Usability Scale

System Usability Scale (SUS): Complete Guide to Usability Measurement

The System Usability Scale (SUS) is a standardized 10-question survey that measures users' perceived usability of digital products, providing a numerical score from 0-100 where scores above 68 indicate above-average usability. Developed by John Brooke in 1986 at Digital Equipment Corporation, SUS remains the most widely adopted usability assessment tool across industries due to its reliability, speed, and ability to generate actionable insights with small sample sizes.

Key Takeaways

  • Industry Standard: SUS is used by 43% of UX professionals according to research, making it the most common usability measurement tool
  • Sample Size Efficiency: Reliable results can be obtained with just 8-12 participants, unlike other metrics requiring 30+ users
  • Benchmark Scoring: The average SUS score across all products is 68, with scores above 80 considered excellent usability
  • Cross-Platform Validity: SUS works equally well for websites, mobile apps, software interfaces, and physical products
  • Quick Implementation: Takes only 2-3 minutes to complete after user testing sessions

What is the System Usability Scale?

The System Usability Scale measures subjective usability through 10 carefully crafted statements that users rate on a 5-point Likert scale. The questionnaire alternates between positively and negatively worded statements to eliminate response bias and ensure accurate measurement of user perceptions.

SUS delivers consistent results across different product types and user demographics. The scoring system allows for direct comparison between product versions, competitor analysis, and industry benchmarking. Research demonstrates that SUS scores correlate strongly with task completion rates and user satisfaction metrics, making it a validated predictor of overall user experience quality.

How the System Usability Scale Works

The SUS questionnaire requires users to rate 10 standardized statements immediately after completing tasks with your product. Users respond using a 5-point scale from "Strongly Disagree" (1) to "Strongly Agree" (5).

The 10 standard SUS statements are:

  1. I think that I would like to use this system frequently.
  2. I found the system unnecessarily complex.
  3. I thought the system was easy to use.
  4. I think that I would need the support of a technical person to use this system.
  5. I found the various functions in this system were well integrated.
  6. I thought there was too much inconsistency in this system.
  7. I would imagine that most people would learn to use this system very quickly.
  8. I found the system very cumbersome to use.
  9. I felt very confident using the system.
  10. I needed to learn a lot of things before I could get going with this system.

Calculating the SUS Score

SUS scoring follows a specific mathematical formula that normalizes responses to a 0-100 scale:

  1. For odd-numbered questions (positive statements), subtract 1 from the user's response
  2. For even-numbered questions (negative statements), subtract the user's response from 5
  3. Sum all converted responses and multiply by 2.5
  4. The result is your final SUS score between 0 and 100

Interpreting SUS Scores

SUS scores translate into five distinct usability categories based on extensive research by Bangor, Kortum, and Miller:

  • Below 50: Poor usability requiring immediate attention
  • 50-68: Below average, significant improvements needed
  • 68-80: Good usability, above industry average
  • 80-90: Excellent usability, top quartile performance
  • Above 90: Exceptional, best-in-class user experience

Academic research also converts SUS scores to letter grades: 90+ (A), 80-89 (B), 70-79 (C), 60-69 (D), Below 60 (F).

Best Practices for Using the System Usability Scale

Effective SUS implementation requires following established protocols to ensure valid, reliable results according to usability research standards.

Timing and Context: Administer SUS immediately after users complete representative tasks with your interface. This captures authentic usability perceptions while the experience remains fresh. Research shows that delayed SUS administration produces less accurate results due to memory decay and interference from other activities.

Maintain Standardization: Preserve the original question wording to maintain validity and enable benchmark comparisons. Minor adaptations like changing "system" to "website" are acceptable, but substantial modifications invalidate the scoring system and decades of comparative research data.

Combine with Qualitative Methods: SUS identifies usability problems but does not explain their causes. Pair SUS with user interviews, think-aloud protocols, or task analysis to understand why scores are high or low and prioritize improvements effectively.

Establish Baselines and Track Progress: Use initial SUS scores as benchmarks for measuring improvement across design iterations. Document testing conditions to ensure valid comparisons over time and demonstrate ROI of UX improvements with quantifiable data.

Common Mistakes When Using SUS

Modifying Questions Significantly: Substantial changes to question wording or order invalidate SUS reliability and prevent comparison with industry benchmarks. According to John Brooke's original research, even small modifications can affect score validity if not applied consistently across all participants.

Insufficient Sample Sizes: SUS requires 8-12 participants minimum per user segment for statistically reliable results. While this is smaller than other usability metrics, teams that test with fewer than 8 users risk unreliable scores that lead to poor design decisions.

Premature Testing: SUS requires functional interfaces that users can meaningfully evaluate through realistic tasks. Testing wireframes or early mockups produces unreliable scores that don't predict actual user experience according to Nielsen Norman Group research.

Score Misinterpretation: The industry average is 68, not 50 as many teams assume. Scores in the 60s indicate acceptable performance, not poor usability requiring immediate intervention. This misunderstanding leads teams to over-invest in fixing products that already meet usability standards.

How Card Sorting Complements SUS Testing

Card sorting exercises identify specific information architecture problems that directly impact SUS scores related to system complexity and learnability. When SUS scores fall below 68, card sorting reveals whether poor organization contributes to user confusion and low confidence scores.

The optimal workflow combines both methods strategically: conduct card sorting to inform information architecture decisions, build prototypes based on those insights, then measure overall usability with SUS. This approach ensures you identify root causes rather than just measuring symptoms of poor user experience.

Card sorting validates navigation structures, terminology choices, and content categorization that directly influence SUS metrics like perceived system complexity, integration of functions, and learning requirements.

Take Action with System Usability Scale

Implement SUS testing in your next usability evaluation to establish baseline usability metrics that drive design decisions. Begin by selecting representative tasks, recruit 8-12 participants from your target audience, and administer the standard SUS questionnaire immediately after task completion. Use the resulting scores to identify improvement opportunities and track progress across design iterations with quantifiable data that demonstrates UX impact to stakeholders.


Frequently Asked Questions

What is a good SUS score?

A SUS score of 68 or higher indicates above-average usability, while scores above 80 represent excellent user experience. The industry average across all digital products is 68, making this the key benchmark for acceptable usability performance according to research by Sauro and Lewis.

How many users do I need for reliable SUS results?

SUS produces reliable results with 8-12 participants per user segment, significantly fewer than traditional usability metrics. Research by Tullis and Stetson demonstrates that SUS scores stabilize quickly due to the questionnaire's statistical properties and standardized format.

Can I modify the SUS questions for my specific product?

Minor modifications like changing "system" to "website" or "app" are acceptable and won't affect reliability according to John Brooke's guidelines. However, substantial changes to question wording or meaning invalidate the scoring system and prevent comparison with industry benchmarks established over decades of research.

How often should I conduct SUS testing?

Conduct SUS testing after major design changes, product launches, or quarterly as part of ongoing UX measurement programs. Consistent testing intervals help track usability trends and validate improvement efforts with quantifiable data over time, providing clear ROI metrics for UX investments.

What's the difference between SUS and other usability metrics?

SUS measures subjective usability perceptions through standardized questions, while metrics like task completion rate or time-on-task measure objective performance. SUS provides the user's overall satisfaction and confidence, complementing behavioral data with attitudinal insights for comprehensive usability assessment that predicts long-term product adoption.

Try it in practice

Start a card sorting study and see how it works

Related UX Research Resources

Explore related concepts, comparisons, and guides