A/B testing for small traffic sites
Can you run A/B tests with 500 visitors a day? Yes โ but you need to adjust your approach. Standard methods that work for high-traffic sites will waste your time without modifications.
The reality of low-traffic testing
The math is unforgiving: detecting a 5% relative lift on a 3% conversion rate requires roughly 85,000 visitors per variant. At 250 visitors per day per variant, that is 340 days โ nearly a year.
This does not mean A/B testing is impossible with low traffic. It means you need to be strategic about what you test and how.
Strategies that work
- Test bigger changes โ Subtle tweaks (button color, copy changes) produce small effects that need massive samples to detect. Instead, test fundamentally different approaches: entirely new page layouts, different value propositions, different pricing structures. Larger effects need fewer visitors to detect.
- Focus on high-traffic pages โ Concentrate testing on your highest-traffic pages. A test on the homepage with 80% of traffic will complete much faster than one on a settings page with 2%.
- Use fewer variants โ Every variant divides your traffic. With 500 visitors/day, an A/B test (2 variants) gives you 250 per variant. An A/B/C/D test gives you only 125 per variant โ making detection even harder.
- Accept a larger MDE โ Instead of trying to detect a 5% relative lift, set your MDE to 15โ20%. You will miss small improvements, but you will catch the big wins โ which matter most anyway. Use the Sample Size Calculator to see how this affects required traffic.
- Run longer tests โ With 250 daily visitors per variant, a test for a 20% relative MDE on a 5% baseline needs about 2,500 visitors โ 10 days. That is perfectly feasible. Plan for 2โ4 week tests rather than 3-day sprints.
Statistical methods for small samples
- Bayesian analysis โ The Bayesian Calculator works well with smaller samples because it gives you a probability rather than a binary significant/not-significant answer. "72% chance B is better" is still useful information, even if a frequentist test would say "not significant."
- Sequential testing โ The Sequential Calculator lets you monitor results continuously and stop early if there is a clear winner. This can save weeks of testing time when the effect is large.
- Fisher's exact test โ For conversion rate tests with very small samples (under 100 per group), Fisher's exact test is more reliable than the standard z-test. The Conversions Calculator offers this option.
What not to do
- Do not lower your significance threshold โ Using ฮฑ = 0.20 instead of 0.05 to get more significant results just means you will ship more false positives. You will make changes that have no real effect โ or make things worse.
- Do not ignore sample size requirements โ Running a test for 3 days with 200 visitors and declaring a winner is worse than not testing at all โ it gives false confidence in a random result.
- Do not combine non-comparable traffic โ Pooling traffic across very different pages or user segments to reach sample size faster introduces confounds that invalidate the test.
A practical framework for small sites
- Pick your highest-traffic page
- Test a meaningful change (not a tweak)
- Set MDE to 15โ20% relative
- Use the Duration Calculator to estimate test length
- Run for at least 2 full weeks
- Analyze with the Bayesian Calculator for the most useful interpretation
Small traffic does not mean you cannot experiment. It means each experiment should count โ test bold ideas, accept larger MDE thresholds, and focus on learning.