Statistical significance is a term used in hypothesis testing to refer to the likelihood that the relationship between two or more variables is caused by something other than chance. Statistical hypothesis testing is a method of making decisions using data, whether from a controlled experiment or an observational study (not controlled).
In hypothesis testing, a p-value is used to determine statistical significance. The p-value measures the probability of the metric lift you observe (or a more extreme lift) assuming that the variant you’re testing has no effect. The standard is to use a p-value less than 0.05 to identify variants that have a statistically significant effect. A p-value less than 0.05 implies that there’s less than 5% chance of seeing the observed metric lift (or a more extreme metric lift) if the variant had no effect. In practice, a p-value that's lower than your pre-defined threshold is treated as evidence for there being a true effect.
Another approach to assess statistical significance is using a confidence interval. A confidence interval examines whether the metric difference between the variant and control overlaps with zero. A 95% confidence interval is the range that covers the true difference 95% of the time. It is usually centered around the observed delta between the variant and control.
For example, if you are testing the effect of a new website design on the amount of time visitors spend on your website, you would compare the time spent by visitors who saw the new design (the treatment group) to the time spent by visitors who saw the old design (the control group). If the p-value is less than 0.05, you might conclude that the new design leads to a statistically significant increase in time spent on the website.
However, it's important to note that statistical significance does not always imply practical significance. A result might be statistically significant but the actual difference between the groups might be so small that it doesn't have any practical implications.
Also, when running multiple tests, the chance of observing a statistically significant result just by chance increases. This is known as the problem of multiple comparisons, and it is often addressed by adjusting the significance level using methods like the Bonferroni correction. The Bonferroni correction reduces the probability of Type I errors (false positives) by adjusting the significance level (α). The significance level is divided by the number of comparisons (equivalent to the number of test variants) in the experiment.