Minimum Detectable Effect

The Minimum Detectable Effect (MDE) is the smallest change in a metric that an experiment can reliably detect. For instance, if the MDE is 1%, it means that if there's a true effect of 1% or larger on our metric, we expect the experiment will show a statistically significant result. If the effect is smaller than 1%, then it will likely fall inside the confidence intervals and not be statistically significant.

The MDE is influenced by several factors, including the number of days or exposures (how long the experiment is active and the number of users enrolled in it), and the allocation (the percentage of traffic that participates in the experiment).

In terms of using the power analysis tool, you can select the population used to determine the metric mean and variance and to estimate the number of exposures over time. The population types include everyone (analysis is based on the entire user base), targeting gate (analysis is scoped to the set of users that pass the selected feature gate), and past experiment (analysis is based on data collected from in a previous experiment).

The tool also allows for fixed allocation analysis (to understand how the length of the experiment impacts the MDE) and fixed MDE analysis (if the smallest effect size that the experiment should detect is known).

Remember, the population selected directly impacts the inputs of the analysis (mean, variance, number of users). To obtain reliable power analysis estimates, the metric values of the selected population should roughly match those of the users you'll be targeting in the experiment.

Join the #1 Community for Product Experimentation

Connect with like-minded product leaders, data scientists, and engineers to share the latest in product experimentation.

Try Statsig Today

Get started for free. Add your whole team!

What builders love about us

OpenAI OpenAI
Brex Brex
Notion Notion
SoundCloud SoundCloud
Ancestry Ancestry
At OpenAI, we want to iterate as fast as possible. Statsig enables us to grow, scale, and learn efficiently. Integrating experimentation with product analytics and feature flagging has been crucial for quickly understanding and addressing our users' top priorities.
Dave Cummings
Engineering Manager, ChatGPT
Brex's mission is to help businesses move fast. Statsig is now helping our engineers move fast. It has been a game changer to automate the manual lift typical to running experiments and has helped product teams ship the right features to their users quickly.
Karandeep Anand
At Notion, we're continuously learning what our users value and want every team to run experiments to learn more. It’s also critical to maintain speed as a habit. Statsig's experimentation platform enables both this speed and learning for us.
Mengying Li
Data Science Manager
We evaluated Optimizely, LaunchDarkly, Split, and Eppo, but ultimately selected Statsig due to its comprehensive end-to-end integration. We wanted a complete solution rather than a partial one, including everything from the stats engine to data ingestion.
Don Browning
SVP, Data & Platform Engineering
We only had so many analysts. Statsig provided the necessary tools to remove the bottleneck. I know that we are able to impact our key business metrics in a positive way with Statsig. We are definitely heading in the right direction with Statsig.
Partha Sarathi
Director of Engineering
We use cookies to ensure you get the best experience on our website.
Privacy Policy