A/B testing has a dirty secret: most tests don't produce a winner. Per VWO and Optimizely research, only 22% of A/B tests produce a statistically significant result at the 95% confidence level. The other 78% are inconclusive — the difference between variants isn't large enough or the sample isn't big enough to declare a winner.
This doesn't mean testing is pointless. It means most teams are testing the wrong things, with too little traffic, for too short a duration. They change a button color, run it for a week on 300 visitors, see a "15% lift" with a p-value of 0.31, and declare victory. That's not optimization. That's self-deception.
The teams that consistently win at A/B testing share three traits: they test big changes (messaging, offers, and page structure — not button colors), they commit to adequate sample sizes before starting, and they accept that most tests will be inconclusive and that's fine. A null result is still information.
Here's what the data says about which tests work, how long they take, and what the winning teams do differently.