A/B testing (also called split testing) is a method of comparing two versions of a webpage, app feature, or element to determine which one performs better. Traffic is randomly split between versions, and performance is measured with data.
Step 1 - Hypothesis "Changing button color from blue to green will increase conversions"
Step 2 - Create Variations
Step 3 - Split Traffic
Step 4 - Measure Results
High-Impact Elements:
Don't test everything at once - isolate one variable
Conversion Rate: Percentage who complete goal Click-Through Rate (CTR): Percentage who click Bounce Rate: Percentage who leave immediately Time on Page: How long users engage Revenue Per Visitor: Economic impact Form Completion Rate: For sign-ups, purchases
Why it matters:
Example:
Version A: 100 visitors, 10 conversions (10%)
Version B: 100 visitors, 11 conversions (11%)
Not significant - need more data!
Version A: 1000 visitors, 100 conversions (10%)
Version B: 1000 visitors, 150 conversions (15%)
Significant - B is clearly better!
Use together for IA optimization:
Card Sorting First: Discover user mental models
A/B Test Implementation: Validate in production
Example: Card sorting reveals users prefer "Plans" over "Pricing". A/B test proves "Plans" converts 23% better.
❌ Testing too many things: Can't tell what worked ❌ Stopping too early: Need statistical significance ❌ Ignoring segments: Different users behave differently ❌ No clear hypothesis: Just changing randomly ❌ Testing tiny changes: Button shade won't move needle ❌ Ignoring context: Seasonal effects, traffic sources
A/B Testing: One element, two versions Multivariate: Multiple elements, multiple versions
Example MVT:
When to use:
Enterprise: Optimizely, VWO, Adobe Target Mid-Market: Google Optimize (free), Unbounce DIY: Custom code with analytics E-commerce: Built into Shopify, BigCommerce
Factors that determine test duration:
Traffic: More traffic = faster results Baseline Conversion: Lower conversion needs more traffic Expected Lift: Bigger changes prove faster Confidence Level: 95% is standard
Typical test duration: 1-4 weeks
✅ One clear goal: Don't optimize multiple metrics ✅ Test high-traffic pages: Need sufficient sample ✅ Run full weeks: Account for weekly patterns ✅ Document everything: Learnings for future tests ✅ Test big changes: Small tweaks rarely matter ✅ Have a hypothesis: Know why you're testing
Don't test if:
Better approaches:
Obama Campaign 2008
Booking.com
Amazon
After using card sorting to design navigation:
Optimize your IA with card sorting first, then validate with A/B testing at freecardsort.com
Explore related concepts, comparisons, and guides