graph showing experiment duration comparison

You launch an A/B test on a Monday. Your traffic estimator says you need 12,000 visitors per variant. At your current volume, that takes five weeks. You wait. Week four, variant B is clearly ahead — +18% on your primary metric, posterior probability above 90%. But your stats tool says "not significant yet." So you keep running.

By week five you have your answer. But you also spent two weeks showing most of your visitors a worse experience. That is the hidden cost of fixed-horizon testing, and most teams are paying it without realizing it.

The fixed-horizon assumption

Classical A/B testing — the kind taught in statistics courses and baked into most early testing tools — is built on a specific mathematical contract. You decide before the test starts how many observations you will collect. You run the test until you hit that number. Then you look at the results exactly once. Any deviation breaks the statistical guarantees.

The required sample size comes from four variables: baseline conversion rate, minimum detectable effect, desired statistical power, and significance threshold. Plug those numbers in and you get a fixed target. The problem is that every one of those variables is an estimate made under uncertainty. Your baseline conversion rate might drift. Your effect size might be larger or smaller than guessed. Your traffic might spike or drop due to seasonality or marketing campaigns.

When the actual conditions differ from the planned ones — which they almost always do — the test either takes longer than expected or produces unreliable results when stopped early. Neither outcome is acceptable for teams trying to move fast.

Why "just use more traffic" does not solve it

The instinct is to throw more traffic at the problem. More visitors means faster results, right? Partially. Doubling traffic roughly halves the test duration. But it does not change the fundamental constraint: you are still locked into a commit-then-wait model.

For e-commerce sites with high daily traffic, this matters less. For B2B SaaS sites with 2,000 visitors per day, a test that needs 50,000 observations per arm takes over three weeks at 100% traffic allocation — and during those three weeks, traffic fluctuates, campaigns change, and your product team is waiting for permission to ship the next thing.

Traffic volume is a multiplier, not a solution. The solution is changing the testing methodology.

The peeking problem and why it gets worse under pressure

Every experimenter peeks. You know you should not check the dashboard until the test is complete, but day three comes along and the conversion rate on variant B is 22% higher. So you look. And looking changes behavior — if the number looks good, you feel pressure to call it early.

Early stopping under fixed-horizon methods inflates false-positive rates significantly. A test designed for 5% false-positive rate (alpha = 0.05) can hit 25-40% false-positive rates when experimenters peek repeatedly. This is not a discipline problem — it is a method mismatch. The tools were not built for the way humans actually behave under commercial pressure.

Sequential testing methods address this directly. They are designed to be checked at any point during the experiment without inflating error rates. The statistical guarantees hold whether you look once or every hour.

How adaptive traffic allocation changes the equation

Standard A/B tests split traffic 50/50 throughout the experiment. This is statistically efficient for measuring a difference, but commercially inefficient. During the test period, roughly half your visitors see the worse-performing variant. You are paying an opportunity cost to measure precisely.

Adaptive methods — specifically multi-armed bandit algorithms — allocate traffic dynamically based on observed performance. As data comes in, the system shifts more traffic to better-performing variants. A variant with 85% win probability might receive 70% of traffic by day five, while the test continues until statistical requirements are met.

This has two concrete effects. First, the experiment produces better business outcomes during the test period itself — fewer visitors exposed to lower-converting experiences. Second, because the system concentrates observations on likely winners, it can reach sufficient evidence faster than 50/50 splits for large effect sizes.

The tradeoff is measurement precision. Adaptive allocation is slightly less precise for estimating exact conversion rate differences, because traffic is uneven. For teams that need exact measurements for quarterly reports, this matters. For teams that need to ship the winner and move to the next experiment, it usually does not.

Bayesian methods and what "stopping early" actually means

Bayesian A/B testing reframes the question from "is this result statistically significant?" to "what is the probability that variant B is better than variant A?" The probability is calculated continuously as data arrives and can be checked at any time without the peeking problem.

A common Bayesian stopping rule is: stop when the posterior probability that the variant beats the control exceeds a threshold — commonly 95% or 97%. Unlike frequentist p-values, this probability is interpretable in plain language. "There is a 96% chance variant B has a higher conversion rate than variant A" means what it says.

In practice, Bayesian experiments with strong effects stop 30-50% earlier than equivalent fixed-horizon tests, because the stopping rule responds to how informative the data actually is. A large, clean effect produces decisive posterior probabilities quickly. A small, noisy effect requires more data — as it should.

What this means operationally

Switching to adaptive or Bayesian testing is not just a methodology change — it changes how your team plans and executes experiments. A few operational adjustments are necessary:

First, define your stopping criteria before the test starts, not during it. For Bayesian methods: what posterior probability threshold is required? What is the minimum runtime regardless of probability (to avoid calling tests that are still too early in noise)? These decisions belong in the experiment design document, not in the moment of pressure.

Second, set a maximum runtime as a fallback. Even adaptive experiments should have a ceiling — typically 8 weeks. If the experiment has not reached conclusive evidence by then, the effect is probably too small to matter practically, regardless of statistical outcome.

Third, separate "should we call this test?" from "what action should we take?" A 92% posterior probability is not a statistically decisive result, but it may be commercially sufficient if the expected uplift is large and the cost of shipping the variant is low. Teams that conflate statistical decision rules with business decisions tend to run experiments longer than necessary.

A rough comparison on experiment duration

Consider a landing page with a 3.2% baseline conversion rate. You are testing a new headline variant. Your minimum detectable effect is +0.5 percentage points. At 1,500 daily visitors, here is how the approaches compare:

Fixed-horizon (90% power, alpha 0.05): requires approximately 24,000 observations per variant. At 750 visitors per arm per day with 50/50 split, that takes 32 days.

Bayesian with adaptive allocation (95% posterior threshold, 20% minimum runtime buffer): if the true effect is +0.8 percentage points — larger than the minimum detectable — the experiment typically concludes in 18-22 days. If the true effect is smaller, say +0.3 points, it takes closer to 35-40 days, which is correct behavior — the evidence is genuinely weaker.

The gain is not guaranteed speed, it is efficiency. Tests that deserve to be called early get called early. Tests where evidence is weak run longer. The methodology matches the data rather than fighting it.

Getting started

If your current testing tool uses fixed-horizon frequentist statistics — which most older tools do — you are not out of options. Some tools allow switching to sequential or Bayesian analysis modes. Others require migrating to a platform that supports adaptive methods natively.

Before migrating, audit your last 10 experiments. How many ran to completion? How many were called early under commercial pressure? How many produced results that did not replicate when the winning variant was measured again in subsequent tests? That audit will tell you how much the methodology is costing you.

The speed of experimentation is one of the few variables that compounds over time. A team running 4 experiments per month instead of 2 does not learn twice as fast in year one — they learn three or four times as fast, because earlier findings inform later hypothesis generation. The methodology is not just a technical choice. It is a competitive one.

Webyn uses Bayesian updating with adaptive traffic allocation to reduce experiment duration without sacrificing reliability. Talk to our team about running your first experiment.

Back to Blog