A/B testing

Compare two versions of a page, email, or feature to determine which performs better using statistical methods that isolate the impact of specific changes.

A/B testing

A/B testing

definition

Introduction

A/B testing (also called split testing) is a controlled experiment that compares two versions of a single element to measure which performs better against a defined metric. One version is the control; the other is the variant. By showing version A to one group and version B to another, you isolate the impact of the change and make decisions based on data rather than intuition.

A/B tests work because they control for variables. When you change only one element—a call-to-action button colour, email subject line, landing page headline, or checkout flow—you can confidently attribute performance differences to that specific change. This removes noise from the equation. A 15% lift in conversion rate on a redesigned form is meaningful. A 15% lift from changing two things simultaneously is meaningless; you don't know which one caused it.

The practice has become standard in B2B marketing because it compounds. A 5% improvement to email open rates, 3% to click-through rates, and 7% to landing page conversions stacks across your entire pipeline. Over a year, these incremental gains become significant revenue gains. But they only work if you run proper tests, reach statistical significance, and move deliberately rather than changing everything at once.

Why it matters

Prevents expensive guesses

Without testing, marketing decisions rest on opinion, trends, or what competitors do. A redesigned homepage might look beautiful but underperform. An email with personalisation might get lower engagement than expected. Testing removes the guesswork and validates assumptions before rolling out changes across your entire audience.

Builds evidence for larger decisions

A single A/B test might improve conversion by 2%. But when you run dozens of tests across your funnel, each small improvement multiplies. The tests also generate internal credibility—stakeholders see the data and buy in to further optimisation work. This accelerates decision-making across product, design, and marketing.

Uncovers unexpected insights

A/B tests often reveal counter-intuitive results. The longer form might outperform the short one. The urgent copy might underperform the educational copy. Testing exposes what your actual audience responds to, not what you assumed they would.

How to apply it

Define your hypothesis and metric

Start by identifying one element to test and the outcome you're measuring. Don't test 'everything looks better'—test 'changing the button from blue to green will increase form completions by 5%'. The metric must be trackable: conversion rate, click-through rate, email open rate, or time on page.

Split your audience randomly

Divide your traffic or user base equally between control and variant. Randomisation prevents selection bias. If your high-intent users all see version B, you can't claim B is better—it's just attracting higher-intent visitors.

Run the test long enough

Reach statistical significance before concluding. A 10% lift from 50 clicks is noise. A 10% lift from 5000 clicks is signal. Most platforms require at least 100-200 conversions per variant before results are reliable.

Document and iterate

Record every test, winner, and insight. This creates institutional memory and prevents repeating failed experiments. Winning tests often become the baseline for the next test—continuous improvement compounds.

Email subject line testing at a SaaS company

A B2B SaaS firm testing email subject lines to their database of 50,000 prospects found that 'Your team is losing £5,000 per week on X' significantly outperformed 'Learn how to improve X efficiency' (32% open rate vs 18%). The test was run to 5,000 people, well above the sample size needed for significance. The winning line was then used in all future campaigns to that segment.

Landing page headline testing for a consulting firm

A management consulting firm tested two headlines: 'Strategy consulting for growth-stage SaaS' vs 'Close your critical strategy gaps in 90 days'. The second variant converted at 8.2% compared to 6.1%. The specificity of the outcome ('close gaps') and the time constraint ('90 days') resonated more than the generic category positioning.

Call-to-action button colour test in a payment platform

A fintech company tested CTA button colours on their checkout page. The contrasting orange button outperformed the muted grey button by 4.3%, a seemingly small lift. Applied across 2 million annual transactions, this translated to an additional £180,000 in revenue annually with zero product changes.

Keep learning

Growth leadership

How do you make all four engines work together instead of in isolation?

Explore playbooks

Data & dashboards

Data & dashboards

Build the dashboards and data pipelines that show your growth engines in one view so you can spot bottlenecks and make decisions in minutes, not meetings.

Growth team tools

Growth team tools

The wrong tools create friction. The right ones multiply your output without adding complexity. These are the tools I recommend for growth teams that move fast.

Review and plan next cycle

Review and plan next cycle

Analyse last cycle's results across all twelve metrics, identify the highest-leverage improvements, and set priorities that compound into the next period.

Revisit quarterly

Revisit quarterly

Pressure-test your strategy against market shifts, performance data, and team capacity so your direction stays relevant and ambitious.

Related books

Hacking growth

Sean Ellis

Rating

Rating

Rating

Rating

Rating

Hacking growth

A practical framework for experiments and insights. Build loops, run tests and adopt a cadence that ships learning every week.

Lean Startup

Eric Ries

Rating

Rating

Rating

Rating

Rating

Lean Startup

A disciplined approach to experiments. Define hypotheses, design MVPs and learn before you scale.

Lean Analytics

Alistair Croll

Rating

Rating

Rating

Rating

Rating

Lean Analytics

Pick the One Metric that Matters for your stage. Build lean dashboards and use data to decide the next best move.

Related chapters

1

Building your backlog

Random testing wastes time and teaches you nothing. Learn how to collect experiment ideas systematically and prioritise them based on potential impact so you always know what to run next.

2

Creating strong hypotheses

Most experiments fail before they start because the hypothesis is vague or untestable. Learn how to write hypotheses that are specific enough to prove or disprove and tied to metrics that matter.

Wiki

Email sequence

Automate multi-touch email campaigns that adapt based on recipient behaviour to nurture leads consistently without manual follow-up from reps or marketers.

Buyer persona

Document your ideal customer's role, goals, and challenges to tailor messaging and prioritise features that solve real problems they actually pay for.

Sales-led growth

Win customers through direct sales conversations where reps guide prospects from discovery to close with personalised solutions and relationship building.

Trigger

Define events that start automation workflows so the right message reaches people at the right moment based on their actual behaviour not arbitrary timing.

Marketing stack

Organise the tools that capture leads, nurture prospects, and measure performance to automate repetitive work and connect customer data across systems.

Positioning statement

Define how you're different from alternatives in a way that matters to customers to guide all messaging and ensure consistent market perception.

Sales qualified lead velocity

Track how fast your pipeline of ready-to-buy leads grows to forecast sales capacity needs and spot when lead quality or sales efficiency changes.

Pirate metrics

Track your user journey through Acquisition, Activation, Retention, Referral, and Revenue to identify which stage constrains growth most.

Contact management

Organise customer and prospect information to track relationships, communication history, and next steps without losing context or duplicating effort.

OMTM (One Metric That Matters)

Focus your entire organisation on the single metric that best predicts success at your current growth stage, avoiding distraction and misalignment.

Statistical significance

Determine whether experiment results reflect real differences or random chance to avoid making expensive decisions based on noise instead of signal.

Attribution model

Assign credit to marketing touchpoints that influence conversions to understand which channels work together and deserve budget in multi-touch journeys.

Sample size

Calculate how many users you need in experiments to detect meaningful differences and avoid declaring winners prematurely based on insufficient data.

Cookie

Store information in browsers to track user behaviour across visits and enable personalised experiences without requiring login for every interaction.

Inbound Marketing

Attract prospects through valuable content that solves real problems, building trust and generating qualified leads who approach you.

P-value

Interpret experiment results to understand the probability that observed differences occurred by chance rather than because your changes actually work.

Annual Recurring Revenue (ARR)

Track predictable yearly revenue from subscriptions to measure business scale and growth trajectory in B2B SaaS and recurring revenue models.

Growth engine

Build self-reinforcing systems across demand generation, funnel conversion, sales pipeline, and customer value that create continuous momentum.

Net Revenue Retention (NRR)

Track revenue growth from existing customers through expansion and contraction to prove your product delivers increasing value over time.

Partner-led growth

Scale through partner relationships where other companies distribute your product to their customers in exchange for commissions or reciprocal value.