The System Usability Scale (SUS) is a standardized 10-question survey that measures users' perceived usability of digital products, providing a numerical score from 0-100 where scores above 68 indicate above-average usability. Developed by John Brooke in 1986 at Digital Equipment Corporation, SUS remains the most widely adopted usability assessment tool across industries due to its reliability, speed, and ability to generate actionable insights with small sample sizes.
The System Usability Scale measures subjective usability through 10 carefully crafted statements that users rate on a 5-point Likert scale. The questionnaire alternates between positively and negatively worded statements to eliminate response bias and ensure accurate measurement of user perceptions.
SUS delivers consistent results across different product types and user demographics. The scoring system allows for direct comparison between product versions, competitor analysis, and industry benchmarking. Research demonstrates that SUS scores correlate strongly with task completion rates and user satisfaction metrics, making it a validated predictor of overall user experience quality.
The SUS questionnaire requires users to rate 10 standardized statements immediately after completing tasks with your product. Users respond using a 5-point scale from "Strongly Disagree" (1) to "Strongly Agree" (5).
The 10 standard SUS statements are:
SUS scoring follows a specific mathematical formula that normalizes responses to a 0-100 scale:
SUS scores translate into five distinct usability categories based on extensive research by Bangor, Kortum, and Miller:
Academic research also converts SUS scores to letter grades: 90+ (A), 80-89 (B), 70-79 (C), 60-69 (D), Below 60 (F).
Effective SUS implementation requires following established protocols to ensure valid, reliable results according to usability research standards.
Timing and Context: Administer SUS immediately after users complete representative tasks with your interface. This captures authentic usability perceptions while the experience remains fresh. Research shows that delayed SUS administration produces less accurate results due to memory decay and interference from other activities.
Maintain Standardization: Preserve the original question wording to maintain validity and enable benchmark comparisons. Minor adaptations like changing "system" to "website" are acceptable, but substantial modifications invalidate the scoring system and decades of comparative research data.
Combine with Qualitative Methods: SUS identifies usability problems but does not explain their causes. Pair SUS with user interviews, think-aloud protocols, or task analysis to understand why scores are high or low and prioritize improvements effectively.
Establish Baselines and Track Progress: Use initial SUS scores as benchmarks for measuring improvement across design iterations. Document testing conditions to ensure valid comparisons over time and demonstrate ROI of UX improvements with quantifiable data.
Modifying Questions Significantly: Substantial changes to question wording or order invalidate SUS reliability and prevent comparison with industry benchmarks. According to John Brooke's original research, even small modifications can affect score validity if not applied consistently across all participants.
Insufficient Sample Sizes: SUS requires 8-12 participants minimum per user segment for statistically reliable results. While this is smaller than other usability metrics, teams that test with fewer than 8 users risk unreliable scores that lead to poor design decisions.
Premature Testing: SUS requires functional interfaces that users can meaningfully evaluate through realistic tasks. Testing wireframes or early mockups produces unreliable scores that don't predict actual user experience according to Nielsen Norman Group research.
Score Misinterpretation: The industry average is 68, not 50 as many teams assume. Scores in the 60s indicate acceptable performance, not poor usability requiring immediate intervention. This misunderstanding leads teams to over-invest in fixing products that already meet usability standards.
Card sorting exercises identify specific information architecture problems that directly impact SUS scores related to system complexity and learnability. When SUS scores fall below 68, card sorting reveals whether poor organization contributes to user confusion and low confidence scores.
The optimal workflow combines both methods strategically: conduct card sorting to inform information architecture decisions, build prototypes based on those insights, then measure overall usability with SUS. This approach ensures you identify root causes rather than just measuring symptoms of poor user experience.
Card sorting validates navigation structures, terminology choices, and content categorization that directly influence SUS metrics like perceived system complexity, integration of functions, and learning requirements.
Implement SUS testing in your next usability evaluation to establish baseline usability metrics that drive design decisions. Begin by selecting representative tasks, recruit 8-12 participants from your target audience, and administer the standard SUS questionnaire immediately after task completion. Use the resulting scores to identify improvement opportunities and track progress across design iterations with quantifiable data that demonstrates UX impact to stakeholders.
A SUS score of 68 or higher indicates above-average usability, while scores above 80 represent excellent user experience. The industry average across all digital products is 68, making this the key benchmark for acceptable usability performance according to research by Sauro and Lewis.
SUS produces reliable results with 8-12 participants per user segment, significantly fewer than traditional usability metrics. Research by Tullis and Stetson demonstrates that SUS scores stabilize quickly due to the questionnaire's statistical properties and standardized format.
Minor modifications like changing "system" to "website" or "app" are acceptable and won't affect reliability according to John Brooke's guidelines. However, substantial changes to question wording or meaning invalidate the scoring system and prevent comparison with industry benchmarks established over decades of research.
Conduct SUS testing after major design changes, product launches, or quarterly as part of ongoing UX measurement programs. Consistent testing intervals help track usability trends and validate improvement efforts with quantifiable data over time, providing clear ROI metrics for UX investments.
SUS measures subjective usability perceptions through standardized questions, while metrics like task completion rate or time-on-task measure objective performance. SUS provides the user's overall satisfaction and confidence, complementing behavioral data with attitudinal insights for comprehensive usability assessment that predicts long-term product adoption.
Explore related concepts, comparisons, and guides