In the documentation here for a popular A/B testing they say,
For example, if you run an A/B testing experiment with a significance level of 95%, this means that if you determine a winner, you can be 95% confident that the observed results are real and not an error caused by randomness. It also means that there is a 5% chance that you could be wrong.
I believe this to be dead wrong.
Statistical significance at 95% means that $\text{p-value} < 0.05$. The p-value is defined as
$$\text{p-value} \equiv P( \text{reject the null} \;|\; \text{the null is true})$$
The complement of this is $$1 - \text{p-value} = P( \text{fail to reject the null} \;|\; \text{the null is true})$$
So 95% significance tells you $P( \text{fail to reject the null} \;|\; \text{the null is true}) > 0.95$.
Assuming "determining a winner" means rejecting the null, the part of the documentation that says "if you determine a winner, you can be 95% confident that the observed results are real and not an error caused by randomness" to me means $0.95 > P(\text{null is false} | \text{reject the null})$ which is flatly not equivalent.
Am I taking crazy pills here? Is there some reasonable way to convert whatever the heck they're saying with "95% confident that the observed results are real and not an error caused by randomness" into $P( \text{fail to reject the null} \;|\; \text{the null is true})$?