Lenny Distilled

The best experiments surprise you

Craft → Analytical Sense

Supporting
If you're very pessimistic, you might miss the surprising result that pops out of an experiment. You might force yourself to do a bunch of experiments grudgingly, but you're like, 'You know what? I hate this. I'm doing four experiments today because I have to do it because I want to be an entrepreneur, but it sucks and everything's miserable and black.' And then you won't notice that, oh, this thing didn't work, but it didn't work in an interesting way.
Supporting
Let's just pretend we ran that experiment. What do you think it'll come back with? Or let's pretend we ran that user study. And the PMs who have the ability to imagine those outcomes, I think, it helps us be much more efficient, too, because we're like, well, if we all think that it's going to go there and that's not going to compel us to take any action, why do it at all?
Nuanced
Twyman's law, the general statement is if any figure that looks interesting or different is usually wrong. If the result looks too good to be true, your normal movement of an experiment is under 1% and you suddenly have a 10% movement, hold the celebratory dinner.
With caveats

A P value is a statistical measure used in A/B testing to determine if experimental results are statistically significant, commonly set at 0.05 (5%).

Many people assign one minus P value as the probability that your treatment is better than control. That is wrong.
With caveats

The "success rate" refers to the historical percentage of A/B tests at Airbnb that showed positive results.

At Airbnb, where the success rate is only 8%, if you get a statistically significant result with a P value less than 0.05, there is a 26% chance that this is a false positive result. It's not 5%, it's 26%.

The Missing Stamp

Every episode of Lenny's Podcast, distilled into the insights that matter and the quotes that make them stick.

LENNY WAS HERE__STAMP_DATE__

Lenny, if you're reading this, the stamp's ready when you are. 🧡🔥