Free A/B Test Calculator

Check if your A/B test results are statistically significant.

Control (A)

Variant (B)

How to Use the A/B Test Calculator

Enter the number of visitors and conversions for both your control (original) and variant (challenger). The calculator instantly computes conversion rates, statistical significance, and key test metrics. You'll see the confidence level, p-value, z-score, relative uplift, and a 95% confidence interval for the difference between the two variants. The result is color-coded: green for significant at 95%+, yellow for marginally significant (90-95%), and red for not significant. Optionally enter your daily traffic to estimate how many days you need to run the test to reach statistical significance with the current effect size.

What Is A/B Test Statistical Significance?

Statistical significance tells you whether the difference between your control and variant is real, or just due to random chance. When a result is "statistically significant at 95%," it means there's only a 5% probability that the observed difference happened by chance. This calculator uses a two-proportion z-test, the standard method for comparing two conversion rates. It calculates a z-score (how many standard deviations the difference is from zero), a p-value (the probability of seeing this result if there's no real difference), and a confidence interval (the range where the true difference likely falls). A common mistake is ending tests too early. Even if you see a winner after a few hundred visitors, the result may not hold up with more data. Always wait until you have enough traffic for statistical significance before making decisions.

Running A/B tests? Track which variant converts through the whole funnel.

Start Free

Frequently Asked Questions

What does "95% confidence" actually mean?

A 95% confidence level means there is only a 5% chance that the observed difference between your control and variant is due to random variation. It does NOT mean there's a 95% chance the variant is better — it means that if you ran this test 100 times with no real difference, you'd see a result this extreme only 5 times.

How long should I run my A/B test?

Run your test until you reach statistical significance (at least 95% confidence) AND have observed at least one full business cycle (typically 1-2 weeks). Ending a test early because it "looks like a winner" leads to false positives. Use the days-to-significance estimate in this calculator to plan ahead.

What is the minimum detectable effect (MDE)?

MDE is the smallest difference between control and variant that your test can reliably detect given the current sample size. A lower MDE requires more traffic. If your MDE is 2 percentage points, your test can detect a change from 5% to 7% but would likely miss a change from 5% to 5.5%.

Can I test more than two variants?

This calculator compares two variants (A vs B). If you're testing multiple variants (A/B/C/D), you need to adjust for multiple comparisons to avoid inflating your false positive rate. A common approach is the Bonferroni correction: divide your target p-value by the number of comparisons.

Why is my test not reaching significance?

The most common reason is insufficient traffic. Small differences need large sample sizes to detect reliably. If your baseline conversion rate is 3% and the true improvement is only 0.5 percentage points, you may need 15,000+ visitors per variant. Other reasons include high variance in your metric, or the change simply having no real effect.

More Free Tools