UX Research Term

Survey Design

Survey Design

Survey design is the practice of writing research questions that collect accurate, unbiased data from respondents. It covers question wording, response format, question order, and survey length. Bad survey design doesn't just produce noisy data — it produces confidently wrong data, because you won't know your questions were flawed until you've already made decisions based on the answers.

Key Takeaways

  • One concept per question: Double-barreled questions ("Was this easy and enjoyable?") are the most common survey design mistake and produce uninterpretable data
  • Randomize option order: Position bias causes respondents to favor the first or last answer choice, skewing results by up to 20%
  • Keep post-sort surveys short: 5-8 questions after a card sort. Participants already spent their cognitive budget on the sorting task
  • Open-ended questions reveal the "why": After a card sort, ask "Why did you group these items together?" to capture reasoning that the sorting data alone can't show

Core Principles

Ask One Thing at a Time

The most frequent survey design failure is the double-barreled question. "Was this card sort easy and enjoyable?" sounds natural, but it's actually two questions wearing a trench coat. A participant who found the task easy but boring can't answer honestly. Split it: "How easy was the sorting task?" and "How enjoyable was the sorting task?" You'll get cleaner data and often discover that ease and enjoyment don't correlate the way you assumed.

Don't Lead the Witness

Leading questions embed the answer you expect. "How much did you enjoy using our intuitive navigation?" presumes the navigation is intuitive and that the participant enjoyed it. Neutral framing: "How would you describe your experience navigating the site?" Leading questions are especially dangerous in unmoderated testing where there's no moderator to notice confused facial expressions.

Provide Balanced Response Options

If your Likert scale runs from "Agree" to "Strongly Agree," you've eliminated the ability to disagree. Use balanced scales with equal positive and negative options and a genuine neutral midpoint. Five-point scales work for most UX research. Seven-point scales add granularity but increase cognitive load without proportional benefit for most sample sizes under 200.

Survey Design for Card Sorting

Post-study surveys are where survey design and card sorting intersect most directly. The card sort captures what participants did; the follow-up survey captures why.

Effective post-sort questions target three areas:

Reasoning: "Were there any cards you found difficult to place? Which ones and why?" This identifies cards with low agreement rates before you even run the analysis, and the qualitative explanations help you fix label problems rather than just detect them.

Confidence: "How confident are you that your groupings reflect how you'd look for this information?" Participants with low confidence often sorted strategically rather than intuitively — their data is still useful but should be interpreted differently.

Domain familiarity: "How often do you use [product/service type]?" This lets you segment results by expertise level. Novice and expert users frequently create different category structures, and knowing which is which changes your IA decisions.

Keep the whole survey under 5-8 questions. Your participants just finished a cognitively demanding task. A 25-question follow-up survey produces abandoned responses and resentful participants who rush through the remaining questions.

Common Mistakes

Asking about future behavior. "Would you use this navigation structure?" People are terrible at predicting their own behavior. Observe what they do in a tree test instead of asking what they would do in a survey.

Ignoring question order effects. Early questions frame how respondents interpret later ones. If you ask "Did you find the categories confusing?" before "How would you rate the overall experience?", you've primed them to think about confusion. Put general satisfaction questions before specific diagnostic ones.

Skipping the pilot test. Run your survey with 3-5 people before launching. Watch their faces as they read each question. If they pause, re-read, or ask for clarification, the question needs rewriting. This 30-minute investment saves you from collecting hundreds of responses to a question nobody understood the same way.

Further Reading

Frequently Asked Questions

What is a double-barreled question in survey design? A double-barreled question asks about two things at once, making it impossible to interpret the answer. For example, "Was this card sort easy and enjoyable?" combines difficulty and satisfaction into one question. A respondent who found it easy but tedious cannot answer accurately. Split double-barreled questions into separate items, each measuring one concept.

How many questions should a post-card-sort survey include? Keep post-card-sort surveys to 5-8 questions. Participants have already spent cognitive effort on the sorting task, so a long follow-up survey produces low-quality responses and increases abandonment. Focus on 2-3 questions about their reasoning during the sort, 1-2 about their confidence level, and 1-2 about their familiarity with the content domain.

Should you randomize question order in UX surveys? Randomize answer option order to prevent position bias, where respondents disproportionately select the first or last option. However, keep question order logical rather than random, as jumping between unrelated topics confuses respondents and reduces data quality. Group related questions together and move from general to specific topics.

Try it in practice

Start a card sorting study and see how it works

Related UX Research Resources

Explore related concepts, comparisons, and guides