A/A Testing

A/A testing is a method used in the field of website optimization, where the same version of a webpage is shown to two different groups of users. This is done to check the accuracy of the testing tools and to ensure that the data collected is reliable and not due to random chance.

How A/A testing Works

In an A/A test, the same version of a webpage (version A) is shown to two groups of users (Group 1 and Group 2). Since both groups are seeing the same version of the webpage, there should ideally be no significant difference in user behavior between the two groups.


The main purpose of A/A testing is to check the accuracy of the testing tools being used. If there is a significant difference in user behavior between the two groups, it could indicate a problem with the testing tools or the data collection process.

Online and Offline A/A Tests

A/A tests can be either online or offline. An online A/A test is run on real users, with an engineer instrumenting your app with the Statsig SDK to check for experiment assignment. Since there is no difference in experience to the user, you expect to only see statistical noise.

An offline A/A test works by querying a representative sample of your data, randomly assigning subjects to Test or Control, computing relevant metrics for Test vs Control and running them through the stats engine. You're looking for the % of false positives.

A/A testing example

For instance, if you're using an A/A test to test a new feature on your website, you would show the same version of the webpage (with the new feature) to both Group 1 and Group 2. If the testing tools are accurate, there should be no significant difference in user behavior between the two groups.

Join the #1 Community for Product Experimentation

Connect with like-minded product leaders, data scientists, and engineers to share the latest in product experimentation.

Try Statsig Today

Get started for free. Add your whole team!

What builders love about us

OpenAI OpenAI
Brex Brex
Notion Notion
SoundCloud SoundCloud
Ancestry Ancestry
At OpenAI, we want to iterate as fast as possible. Statsig enables us to grow, scale, and learn efficiently. Integrating experimentation with product analytics and feature flagging has been crucial for quickly understanding and addressing our users' top priorities.
Dave Cummings
Engineering Manager, ChatGPT
Brex's mission is to help businesses move fast. Statsig is now helping our engineers move fast. It has been a game changer to automate the manual lift typical to running experiments and has helped product teams ship the right features to their users quickly.
Karandeep Anand
At Notion, we're continuously learning what our users value and want every team to run experiments to learn more. It’s also critical to maintain speed as a habit. Statsig's experimentation platform enables both this speed and learning for us.
Mengying Li
Data Science Manager
We evaluated Optimizely, LaunchDarkly, Split, and Eppo, but ultimately selected Statsig due to its comprehensive end-to-end integration. We wanted a complete solution rather than a partial one, including everything from the stats engine to data ingestion.
Don Browning
SVP, Data & Platform Engineering
We only had so many analysts. Statsig provided the necessary tools to remove the bottleneck. I know that we are able to impact our key business metrics in a positive way with Statsig. We are definitely heading in the right direction with Statsig.
Partha Sarathi
Director of Engineering
We use cookies to ensure you get the best experience on our website.
Privacy Policy