Here are some example data:
+----------------+-----------+---------------+--------+-----------------+
| Version | Converted | Not Converted | Total | Conversion Rate |
+----------------+-----------+---------------+--------+-----------------+
| Original | 300 | 10,000 | 10,300 | 2.9% |
| Test Variation | 175 | 5,000 | 5,175 | 3.4% |
| Total | 475 | 15,000 | 15,475 | 3.1% |
+----------------+-----------+---------------+--------+-----------------+
Context is a website where a test landing page was created diverting a percentage of traffic to the variation. Over the time period the variation showed a higher conversion rate. Is this real or is it just ebbs n flows?
To get the probability of seeing 475 conversions based on a null hypothesis I might use:
> dbinom(x = 475, size = 15475, p = 300/10300)
[1] 0.009591306
There's a 0.0096% chance of seeing 475 conversions assuming no difference in the test versions.
Or, would I do this:
> 1 - pbinom(q = 475, size = 15475, p = 300/10000)
[1] 0.2961731
In this case "There's a 29% chance we would have seen 475 or more conversions if the null is true.
I must be over thinking it because I cannot form my question correctly. I want to know if the test was really a success or not. So that my head doesn't explode I'd prefer an answer in terms of a binomial distribution approach that I've started on, if possible. But would appreciate pointers on any in built r functions for this kind of thing too.
Should I be looking for the probability of seeing exactly 475 conversions or should I be looking for the probability of seeing 475 or more conversions?