Conversion Rate Optimisation

A/B Testing for CRO: A Practical Framework That Drives Revenue

By Harrison Hill· Founder & Chief Strategist
12 min read

Most businesses treat A/B testing as a one-off exercise: change a button colour, see if conversions go up, move on. That approach wastes time and produces unreliable results.

Effective conversion rate optimisation requires a systematic testing framework rooted in data, not guesswork. The companies getting consistent 20-40% conversion improvements aren't running random tests, they're following a disciplined process.

At iNDEXHILL, we build testing programmes that compound over time. This guide covers the framework we use with clients, from hypothesis formation through to statistical analysis.

Why Most A/B Tests Fail

Industry data suggests 60-80% of A/B tests produce no statistically significant result. That's not because testing doesn't work, it's because most tests are poorly designed.

Common Failure Modes

  • Testing without a hypothesis — Changing random elements without understanding why you expect improvement
  • Insufficient sample size — Ending tests after a few hundred visitors rather than waiting for statistical significance
  • Testing too many variables — Multivariate tests that need millions of visits to reach significance
  • Ignoring segment differences — A test might win for mobile users but lose for desktop, and the aggregate masks both signals
  • Peeking at results — Checking daily and stopping when the graph looks good, before reaching the required confidence level

What Good Tests Look Like

The tests that produce reliable, actionable results share common characteristics: a clear hypothesis, sufficient traffic, isolated variables, and pre-defined success criteria.

A/B Test Conversion Rate Improvements

Average uplift across 200+ ecommerce and lead-gen tests

  • control
  • variant

CTA colour changes and form length reductions deliver the highest conversion uplift (62% and 59% respectively), while social proof additions show a more modest 29% improvement. Headline copy and page layout changes consistently land in the 44-47% range — strong enough to justify testing across most landing pages.

View full data table
TestControl (%)Variant (%)Uplift (%)
CTA colour2.1%3.4%+62%
Headline copy1.8%2.6%+44%
Form length3.2%5.1%+59%
Social proof2.4%3.1%+29%
Page layout1.9%2.8%+47%
Pricing display2.7%4%+48%

The chart above shows typical conversion improvements when tests follow a structured framework. Form length and CTA changes consistently deliver the largest uplifts because they directly reduce friction in the conversion path.

Building Testable Hypotheses

Every test should start with a structured hypothesis that connects observation, change, and expected outcome.

The Hypothesis Template

Use this format: "Because [observation from data], we believe [specific change] will [expected outcome], measured by [metric]."

  • Example 1 — "Because 68% of users abandon the checkout at the delivery options step, we believe simplifying delivery choices from 5 to 3 will increase checkout completion by 15%, measured by transaction rate"
  • Example 2 — "Because heatmap data shows users scroll past the CTA without clicking, we believe moving the CTA above the fold will increase click-through by 20%, measured by CTA click rate"
  • Example 3 — "Because exit survey data cites price uncertainty, we believe adding a price calculator will increase form submissions by 25%, measured by lead form completion rate"

Prioritising Tests: The ICE Framework

Not every hypothesis deserves a test. Score each idea on three dimensions:

  • Impact (1-10) — How much will this move the needle if it wins?
  • Confidence (1-10) — How strongly does data support this hypothesis?
  • Ease (1-10) — How quickly can we implement and run this test?

Multiply the scores. Tests scoring above 200 should run first. Below 100, park them for later.

Statistical Significance: Getting It Right

The number one mistake in A/B testing is calling a winner too early. Statistical significance is not optional, it's the difference between a real insight and noise.

Key Concepts

  • Confidence level — Aim for 95% minimum. This means there's only a 5% chance the result is random
  • Statistical power — Target 80%. This is the probability of detecting a real effect when one exists
  • Minimum detectable effect (MDE) — The smallest improvement worth detecting. Smaller MDE = larger sample needed
  • Sample size — Calculate before starting. A page with 1,000 monthly visitors testing for a 10% improvement needs roughly 4 weeks at 95% confidence

How Long to Run Tests

Minimum test duration depends on your traffic volume and the effect size you're trying to detect:

  • High-traffic pages (10,000+ daily visitors) — Most tests reach significance within 1-2 weeks
  • Medium-traffic pages (1,000-10,000 daily) — Allow 2-4 weeks
  • Low-traffic pages (under 1,000 daily) — Consider testing larger changes with bigger expected effects, or use qualitative research instead

Always run tests for full business cycles (minimum one full week) to account for day-of-week variation.

What to Test First: The Conversion Hierarchy

Start with elements that have the highest impact on conversions and work down:

Tier 1: High-Impact Elements

  • Value proposition — Does the headline communicate why someone should care?
  • Call-to-action — Is it clear, compelling, and visible?
  • Form length — Are you asking for more information than necessary?
  • Page speed — Every 100ms of delay costs conversions

Tier 2: Medium-Impact Elements

  • Social proof placement — Testimonials, reviews, client logos
  • Trust signals — Security badges, guarantees, accreditations
  • Visual hierarchy — Does the page guide the eye to the CTA?
  • Mobile experience — Touch targets, scroll depth, thumb-zone optimisation

Tier 3: Refinement Elements

  • Copy tone and length — Formal vs conversational, long vs short
  • Image selection — People vs products, lifestyle vs technical
  • Colour and typography — Brand-aligned variations
  • Micro-interactions — Button hover states, loading animations, progress indicators

Testing Tools and Implementation

The right tool depends on your traffic volume, technical capability, and budget.

Tool Comparison

  • Google Optimize (successor tools) — Free or low-cost, integrates with GA4, good for teams starting out
  • VWO — Strong visual editor, good for non-technical teams, solid statistical engine
  • Optimizely — Enterprise-grade, full-stack capability, server-side testing
  • AB Tasty — European-headquartered (GDPR-native), good personalisation features
  • Custom builds — Feature flags and server-side testing for technical teams wanting full control

Implementation Checklist

  1. Install tracking code on all pages (not just test pages)
  2. Set up goal tracking in your analytics platform
  3. Verify the test renders correctly across devices and browsers
  4. QA the variant against your original to ensure no broken elements
  5. Set a calendar reminder for the minimum test duration, resist checking early
  6. Document every test in a shared log: hypothesis, variant, results, learnings

Building a Testing Culture

The real value of A/B testing compounds over time. Individual tests produce incremental gains, but a culture of continuous testing produces transformational results.

The Compounding Effect

If you run 4 tests per month and 25% produce a 10% improvement, after 12 months you'll have achieved roughly 12 winning tests. At 10% each, that's a cumulative improvement of over 300% from your starting baseline.

Creating a Test Backlog

  • Analytics review — Where are users dropping off? What pages have high bounce rates?
  • Heatmap analysis — Where are users clicking? How far do they scroll?
  • User feedback — What do customers complain about? What questions do support teams get?
  • Competitor analysis — What are competitors doing differently on their conversion pages?
  • Industry benchmarks — Where does your conversion rate sit relative to your sector?

Document everything. A test that fails today might inform a winning test six months later. The learning is as valuable as the result.

How we do this at iNDEXHILL

Our Web Design & CRO services are built around this exact framework, designed for businesses that need predictable growth.

See how we applied this approach in our client case studies.

Frequently Asked Questions

As a rule of thumb, you need at least 1,000 conversions per variant to detect a 10% improvement at 95% confidence. For most B2B sites, that means testing on your highest-traffic pages first. Low-traffic sites should consider larger, bolder changes or use qualitative research methods instead.

Minimum one full business week to capture day-of-week variation. Most tests need 2-4 weeks. Never stop a test early just because results look good. Pre-calculate the required sample size and commit to running until you reach it.

Ideally, yes. Mobile and desktop users have different behaviours, and a change that helps mobile users might hurt desktop users. At minimum, segment your results by device after the test concludes.

Individual tests typically produce 5-15% improvements. Compounded over a systematic programme of 40-50 tests per year, total improvement of 100-300% over 12 months is achievable. The key is consistency and learning from every test.

Want help implementing this?

If you're looking to scale organic growth, we offer a free SEO audit to identify quick wins and growth opportunities.

Request a free SEO audit

Continue Reading