We’ve added metrics and metric source tools to the Statsig MCP, so your agents can now easily read and analyze Statsig metrics metadata from within their workflows.
List metrics and metric sources in your Statsig project
Retrieve metric definitions
This makes it much easier for MCP-powered workflows to find the right metric, inspect how it’s defined, and understand which metric sources are available before creating, updating, and analyzing experiments and gates.
Before this, the Statsig MCP supported adding existing metrics to gates and experiments, but had less visibility into the metrics layer itself. Now with these tools, agents can reason about Statsig metric definitions and sources directly, making it easier to discover the right metrics, understand their definitions, and set up experiments with confidence.
If you have the Statsig MCP set up, try the below example prompts and workflows to explore the new metrics functionality:
"What metrics do we have related to user retention? Pull their definitions and suggest which would work best for a 7-day activation experiment."
"Pull up the definition for the metric [metric name]"
(For Warehouse Native projects) "What metric sources do we have configured?"
To set up the Statsig MCP server and explore all the capabilities it supports today, see our docs page.
We’ve enhanced our CUPED variance reduction methodology in Cloud by automatically applying Stratification to better account for users with missing pre-experiment data.
Users are grouped into strata based on available pre-experiment information. Treatment effects are then calculated within each group before being aggregated into a single estimate and applying standard difference-in-means and variance calculations.
This improvement retains more users in analysis while still applying variance reduction wherever pre-experiment data exists. The result is tighter confidence intervals for more reliable decision-making.
We’re excited to announce the Statsig ChatGPT App — bringing Statsig into the place where more and more product work happens: ChatGPT.
Whether you’re using ChatGPT to draft analysis, experiment reviews, PRDs, user behavior studies, launch plans, or stakeholder updates, you can now pull in Statsig context directly into your workflow and take action without leaving the conversation.

With the Statsig ChatGPT App, you can:
Bring Statsig into your analysis by pulling experiment context and results, and asking questions like “what moved?” or “what should we do next?”
Compound knowledge across your connected tools by tying Statsig data to the docs, notes, and discussions you already have connected in ChatGPT
Build a flywheel of rapid iteration by generating insights and instantly creating or updating Statsig gates, experiments, and dynamic configs, all within ChatGPT.
ChatGPT is increasingly embedded in the day-to-day workflows of product builders. The Statsig ChatGPT App enhances that workflow, letting you plan with real results and ship the next step immediately—all in the same thread.
Install the Statsig ChatGPT App.
Connect your Statsig workspace (you’ll be prompted to authenticate).
Start exploring — and shipping.
For more setup instructions, see our docs page.
“I’m writing an experiment review — pull the latest results and context for [experiment name] and draft a summary.”
“Based on this PRD, create an experiment with two variants and the key metrics we mention.”
“Create a feature gate for [feature] and roll it out to 5% of users in [segment].”
“Create a dynamic config called [name] with these keys, and suggest a safe rollout plan.”
WHN customers can now provide their own Switchback assignment data for a regression based analysis in Statsig. This allows users to have maximum flexibility by controlling their own assignment method while leveraging already existing metric data in Statsig.
To get started you will need to first create a dedicated assignment source for Switchback experiments.

And select Switchback as an Experiment type in the Analyze Only experiment creation modal.

The feature is available today in beta. If you’re interested in learning more or need help getting started please reach out to us via Slack or through your account manager!
WHN users can now create Composite Metrics in Statsig. Composite Metrics let you define a single metric as the sum or difference of multiple value aggregations.

Use to:
Measure net impact (for example, gains minus losses)
Compare the differences between values (last/first or max/min)
For added flexibility, you can break out components of a composite metric in your experiment results.

The feature is available today. Visit the docs page to learn more.
We introduced a clearer, more flexible set of privacy controls for Session Replay. You can choose a baseline privacy configuration and refine it with element-level rules. This makes it easier to align replay collection with your organization’s privacy requirements while preserving useful context for analysis.
Select one of three baseline privacy options that define how text and inputs are handled by default.
Apply CSS selector rules to mask, unmask, or fully block specific elements.
Manage all replay privacy settings from a single place in the console.
You begin by choosing a baseline privacy option in the Statsig Console UI. This sets the default masking behavior for all session replays.
Baseline privacy options:
Passwords (Default): Only password inputs are replaced with asterisks (*). All other text and inputs are shown as is.
Inputs: All text inside input fields is replaced with asterisks (*). All other text is shown as is.
Maximum: All text and all inputs are replaced with asterisks (*).
After selecting a baseline, you can add CSS selector rules to override it for specific elements. Selector rules follow a strict precedence order: Block, then Mask, then Unmask. Password inputs are always masked and cannot be unmasked.
Blocking removes an element entirely from the replay and replaces it with a black placeholder of the same size. Masking replaces text with asterisks. Unmasking reveals text that would otherwise be masked by the baseline setting.
All settings are configured in the Statsig Console under Project Settings → Analytics & Session Replay and require project admin permissions.
These controls let you confidently use Session Replay in privacy-sensitive environments. You can protect PII by default while selectively revealing safe UI elements for debugging, without sacrificing visibility into user behavior.
Longer-running queries no longer block your workflow. Metrics Explorer now supports background queries, giving you a dedicated experience for queries that take longer to complete.
Kick off queries that continue running even after you close the tab or switch tasks.
Avoid timeouts for complex or large queries.
Track queries that are still running from a central, visible place.
When Metrics Explorer detects that a query will exceed typical execution time, it automatically runs it in the background. You are free to navigate away or close the browser without interrupting execution. In-progress background queries appear in Metrics Explorer under the breadcrumb menu at the top, where their status is clearly labeled.
You can confidently run heavier queries without worrying about timeouts or keeping a tab open. This makes it easier to explore larger datasets, iterate on complex funnels, and parallelize analysis with other work.
We’re excited to announce the Count Distinct metric type for Statsig Cloud. Count Distinct lets you measure unique entities—like transactions, devices, or pages views—by counting unique values across the experiment window.
The Count Distinct metric is a sketch based metric that relies on probabilistic HyperLogLog++ (HLL++). We chose this approximation method to optimize for efficiency and speed. Visit our docs page to learn more about our implementation.
It’s available starting today.
For Statsig WHN customers, you can now explore your experiment results in greater detail by grouping the results by up to three user properties.

With Multi-Dimensional Group By, you can break down results by combinations like country × device × plan all in one view.
This expands your ability to find interesting insights and data trends at a greater granularity.
Try it now in the Explore tab of your experiments.
We know writing an experimentation report is everyone’s favorite activity. For those who don’t, it just got easier with Statsig.
With AI-Powered Experiment Summaries, Statsig automatically turns your experiment data into a clear, human-readable summary. The AI summary works to understand the purpose of the experiment, make a ship recommendation, and highlight the most relevant metric observations.

This feature is available now for all Statsig customers. To enable, go to Setting → Statsig AI → AI Summary.