- Conversion Current
- Posts
- What Looks Like a Winner Might Just Be Wednesday
What Looks Like a Winner Might Just Be Wednesday
Why every A/B test deserves (at least) two weeks
You launch a test.
Variant B jumps ahead by 20% by Day 3.
Slack is lit. The founder wants to ship it.
| "Can we just call it now?”
And honestly… it’s tempting.
But here’s the uncomfortable truth:
If you’re calling a test before it runs two full weeks, you’re not optimizing. You’re guessing.
📆 Time has a bias too
Most teams obsess over sample size and significance. But when that sample arrives is just as important as how many people are in it.
A 5-day test might accidentally capture:
A weekend spike in casual shoppers
A weekday lull from your 2nd largest customer cohort
A temporary traffic surge from an email or promo
Example: A DTC brand ran a 6-day test on a new homepage. The winning variant showed a 15% lift — but when they rolled it out, performance tanked. Why? The test had launched on a Friday and caught weekend shoppers, but lost effectiveness during the following workweek.
Two weeks isn’t magic, but it is practical:
Two full weekly cycles.
Enough to even out the weird days and normalize behavior across sources, devices, and intent levels but not so long as to create a non-starter obstacle.
🎲 Most early wins are just noise
Day 3 lift feels great - but that’s usually when variance outpaces truth.
You might be seeing:
Highly engaged users from a specific paid ad or a single creative
Early users clicking everything
A few big buyers tilting revenue metrics
This is when teams make the classic CRO mistake:
Confusing momentum for signal.
Example:
A B2B SaaS team tested two versions of their pricing page. The new version saw a huge spike in demo signups in the first 48 hours. But buried in the (ignored) data? Most signups were from one small LinkedIn campaign targeting current users, not new prospects. The “uplift” was just noise.
The antidote? Let the test run.
Let reality settle in. Let randomness average out.
🧪 Testing isn’t just math
Yes, you need statistical significance.
But you also need representative data - not just “enough” users, but the right mix of users, spread across the rhythms of time.
Think of your A/B test like fermentation.
Cut it short, and it looks done… but the flavor hasn’t developed.
Example:
An ecommerce brand testing a new product page design reached “95% significance” in 4 days. But post-test analysis showed mobile users (who typically make up 60% of traffic) were underrepresented due to a temporary bug in Meta ad delivery. The result wasn’t invalid, but it certainly wasn’t reliable either.
🛡️ The ROI of waiting
When you give a test 2+ weeks, you:
Avoid false positives
Catch inconsistent performance across days
Build real confidence that your “winner” will keep winning
Train your team to respect the process, not just the outcome
Example:
A brand tested a new headline on landing page. In Week 1, conversions spiked. But in Week 2, they dropped below baseline. The spike? It correlated with a single-day TikTok mention they didn’t even know had happened. Without a full 2-week run, they would’ve rolled out a loser.
Waiting protects your roadmap from rollouts based on false signals.
It gives your insights durability.
And in CRO, durability beats dopamine.
💡 TL;DR
If it’s really a winner,
It’ll still be one next week.
So slow down. Let the test breathe. And give time the respect it deserves, because it's the variable most teams forget, and the one most likely to burn them.
That’s all I got for today. The greatest compliment I can receive is sharing this with other people, so if you found value in this - forward this along.
‘Til next time,
Kanika
Reply