Product Updates

Statsig Product Updates
Arrow Left
Arrow Right
5/4/2023
Permalink ›

🧑‍🤝‍🧑Cloning Metrics

Have you ever set up a relatively complex Custom Metric and then realized you want another similar metric but with a slight tweak? Yep, we have too! To make that process easy, today we’re introducing the ability to clone Custom Metrics.

To clone a Custom Metric, go to the "…" menu in a metric page, then select “Clone.” You will have the opportunity to name your new metric, add a description and tags, and then we will auto-fill all the inputs of the metric definition from the source metric. Customize to your liking and you're good to go!

cloning metrics

4/28/2023
Permalink ›

Happy Friday, Statsig Community! To cap off a beautiful week here in Seattle ☀️, we have a number of exciting launch updates to share:

🕒 Fast(er) Pulse

Todate, when you launch a new feature roll-out or experiment, you have to wait 24 hours to start seeing your Pulse results. Today, we’re very excited to shorten that time significantly with the launch of more real-time Pulse. Now, you will see Pulse results start to flow through within 10-15 minutes of starting your roll-out or experiment.

faster pulse 2

A few things to consider-

  • For the first 24 hours, results do not include confidence intervals; early metric lifts are meant to help you ensure that things are looking roughly as expected and verify the configuration of your gate/ experiment, NOT make any launch decisions

  • The Pulse hovercard view will look a bit different; time-series and top-line impact estimates will not be available until the first 24-hour daily lift calculation

☁️ Environments in Overrides

At some companies, an user may have a different ID in different environments and hence want to specify the environment to override a given ID in. To enable this, we’ve added the ability to specify target environment for Overrides in Experiments. For Gates, you can achieve this via creating an environment-specific rule.

environments in overrides

⌛ Experiment Duration by # Target Exposures

(vs. Strictly Time Duration)

We’re introducing more flexibility into how you can measure & track experiment target duration. Now, you can choose between setting a target # of days or a target # of exposures an experiment needs to hit before a decision can be made.

experiment duration

To configure a target # of exposures, tap “Advanced Settings” in Experiment Setup tab, then under “Experiment Measured In” select “Exposures” (vs. “Days”). The progress tracker at the top of your experiment will now show progress against hitting target number of exposures.

experiment duration 2

See our docs for more details.


4/17/2023
Permalink ›

Manual assignment for Stratified Sampling

Layers in a Cake

Statsig manages randomization during experiment assignment. In some B2B (or low scale, high variance cases) the law of large numbers doesn’t work. Here it is helpful to manually assign users to test and control to ensure both groups are comparable. Statsig now lets you do this. Learn More

What is Stratified Sampling?

Stratified sampling is a sampling method that ensures specific groups of data (or users) are properly represented. You can think of this like slicing a birthday cake. If sliced recklessly, some people may get too much frosting and others will get too little. But when sliced carefully, each slice is a proper representation of the whole. In Data Science, we commonly trust random sampling. The Law of Large Numbers ensures that a sufficiently-sized sample will be representative of the entire population. However, in some cases, this may not be true, such as:

  • When the sample size is small

  • When the samples are heterogeneous


4/4/2023
Permalink ›

Warehouse ingestion tab makeover

We gave our Warehouse Ingestion tab a total makeover so that you can have better visibility into your import status! Some key improvements include:

  • A simple visual display to track your import progress, with an extended date range

  • Verify your imported data with ease and confidence using our import volume chart and data samples

  • Take actions more easily and stay in control of your imports (use the “…” menu), whether you want to trigger a backfill or edit your daily ingestion schedule

Statsig data warehouse metrics

3/31/2023
Permalink ›

Explore metrics outside just an experiment

We’ve heard from some folks that they want to explore metrics even outside an experiment’s context. We’ve just started adding capabilities to do this. Now, when you’re looking at a metric in the Metrics Catalog you can:

  • compare values to a prior period to look for anomalies

  • apply smoothing to understand trends

  • look at other metrics at the same time to see correlation (or lack thereof) group by metric dimensions

  • save this exploration as a Dashboard to revisit/share with others

  • view current experiments and feature rollouts that impact this metric (also in Insights)

This starts rolling out March 31.

Old Metrics Exploration Page
The New Metrics Exploration Page

3/31/2023
Permalink ›

Incorporating more granular target population in Power Calculator calculations

Now, you can specify which audience you want to calculate experimental power for, by selecting any existing Feature Gate via the Power Calculator.

To do this, go to the Power Calculator (either under “Advanced Settings” in Experiment creation or via the “Tools & Resources” menu) and select “Population”.

This will kick off an async power calculation based on the selected targeting gate’s historical metric value(s), and you will be notified via email and Slack once your power analysis is complete.

power analysis calculator screenshot

3/25/2023
Permalink ›

Left navigation bar auto-collapse

This gives you more free real estate to do your work in console! This will now be the default setting, but you can switch this back to manual collapse by using the “…” menu on the nav bar.

Permanent and stale gates

This can be found under Types filter in your Gates catalog. While these gates indicate helpful information about your flags, they will not change anything about the functionality of the flags.

statsig premanent and stale gates
  • Permanent Gates (set by you) are gates that are expected to stay in your codebase for a long time (e.g. user permissions, killswitches). Statsig won’t nudge you to clean up these gates.

    • You can set gates to be Permanent in the creation flow or by using the “…” menu within each gate page.

  • Stale Gates (set by Statsig) are good candidates to be cleaned up (and will be used to list out gates for email/slack nudges)

    • On Monday morning, you’ll receive your first monthly nudge (email + slack) to take action on stale gates.

    • At a high level, these gates are defined as 0%/100% rolled out or have had 0 checks in the last 30 days (but exclude newly created or Permanent gates).

statsig

Please see the permanent and stale gates documentation for more information.


3/7/2023
Permalink ›

🚨 Metric Alerts

Today we are introducing the ability to configure thresholds for your metrics that will automatically trigger an alert if breached in the context of any feature rollout or experiment.

This is especially useful for your company’s suite of core Guardrail Metrics, as you can configure thresholds once and rest assured that you’ll be notified whenever a new feature gate or experiment breaches the pre-set threshold. (As a reminder you can also hook up Statsig to Datadog monitors, read more here!)

metric alerts setup

To configure a Metric Alert-

  1. Go to the Metric Detail View page of the metric in question

  2. Tap into the “Alerts” tab and tap “+ Create Alert”

  3. Configure the threshold value and minimum # of participating units required to trigger the alert (even though this second field is optional, we highly recommend configuring it to minimize the noisiness of your alert)

When your metric alert fires, you will be notified via email (and Slack if you’ve configured the Statsig Slack bot) and directed to the “Diagnostics” tab of the offending gate or experiment.

metric alerts

Please note that these alerts are for metric values in the context of a specific gate or experiment and NOT on the top-line value of a metric. See our docs for more details on Metric Alerts, and don't hesitate to reach out if you have questions or feedback!


3/3/2023
Permalink ›

Composite Sums, Pass Rate Filter, Permanent Gates, and More

New custom metric type -

➕ Composite Sums: You can now create an aggregation (sum) metric using other metrics from your catalog, whereas previously, you were only able to sum up events. You can now do cool things such as: adding up revenue across different categories or user counts across different regions.

Iteration on the previously launched feature gate lifecycle/cleanup toolsets -

☑️ Pass Rate filter: We heard your feedback and have added a Pass Rate filter in your Gates Catalog, in addition to the existing Roll Out Rate filter, to make your launch/disable/cleanup decisions off of!

What’s the difference? Roll Out Rate is strictly based on the gate rules you’ve set, whereas Pass Rate shows what’s happening in practice at the time of evaluation. For example, if you have a set of rules that effectively pass everyone besides a small holdout group, Roll Out Rate would be < 100% but Pass Rate could still be 100% if the holdout is small enough.

♾️ Permanent Gates: Permanent feature gates are expected to live in your codebase for an extended period of time, beyond a feature release, usually for operations or infrastructure control (examples: user permissions, circuit breakers/kill switches).

permanent gates statsig screenshot

You can now mark your gates Permanent on Statsig, telling your team (and Statsig) that they should proceed with more caution if attempting to clean up these gates. This will not change anything functionally about the gate itself, but will allow us to surface and label them differently in the Statsig console for your convenience.


2/27/2023
Permalink ›

Flexible Environment Configuration and Metrics Directionality

Hi everyone! Here are a few launch announcements to kick off our week. 

🌎 Flexible Environment Configuration

Today, we’re introducing a number of updates to Environments on Statsig-

  1. Adding & Customizing Environments

    Whereas previously Statsig only supported two, pre-defined environments, we are now opening up the ability to create as many environments as you want and customize the name and relative hierarchy of these environments. To customize the environments available in your Statsig Project, go to the “Environments & Keys” tab within Project Settings.

  2. API Keys per Environment

    We are also adding support for environment-specific API keys, enabling greater launch security and privacy through more API key granularity. To configure environment-specific API keys, go to the “Environments & Keys” tab within Project Settings.

  3. Improved Environments UX

    You’ll also notice that environments look and feel a bit different in the Console. We now expose which environments a given rule is targeted at on every rule header, with the ability to toggle rule-sets by target environment via the environment filters at the top of the gate.

Statsig Specify Environments
Statsig Manage Environments

Read more about how environments work via our Docs Page. 

↗️ Metrics Directionality

Today, Pulse makes a relatively naïve assumption that the world is simple and a metric increase = good = green, and a metric decrease = bad = red. However for many metrics, it’s actually a good thing when the metric decreases and similarly bad to see an increase (think page load times or many performance metrics!)

With metric directionality, you can set the desired direction you want to see a metric move and Pulse will update the color-coding in metric lifts accordingly. To access this feature, go to any metric detail view page, and tap the “…” menu to see the “Set Metric Directionality” option.

Statsig Set Metric Directionality
Statsig Increase or Decrease Metric Directionality

We hope you enjoy this new functionality! Please get in touch with us with any feedback, and join the Statsig Slack community to receive these updates first!


Join the #1 experimentation community

Connect with like-minded product leaders, data scientists, and engineers to share the latest in product experimentation.

Try Statsig Today

Get started for free. Add your whole team!

Why the best build with us

OpenAI OpenAI
Brex Brex
Notion Notion
SoundCloud SoundCloud
Ancestry Ancestry
At OpenAI, we want to iterate as fast as possible. Statsig enables us to grow, scale, and learn efficiently. Integrating experimentation with product analytics and feature flagging has been crucial for quickly understanding and addressing our users' top priorities.
OpenAI
Dave Cummings
Engineering Manager, ChatGPT
Brex's mission is to help businesses move fast. Statsig is now helping our engineers move fast. It has been a game changer to automate the manual lift typical to running experiments and has helped product teams ship the right features to their users quickly.
Brex
Karandeep Anand
President
At Notion, we're continuously learning what our users value and want every team to run experiments to learn more. It’s also critical to maintain speed as a habit. Statsig's experimentation platform enables both this speed and learning for us.
Notion
Mengying Li
Data Science Manager
We evaluated Optimizely, LaunchDarkly, Split, and Eppo, but ultimately selected Statsig due to its comprehensive end-to-end integration. We wanted a complete solution rather than a partial one, including everything from the stats engine to data ingestion.
SoundCloud
Don Browning
SVP, Data & Platform Engineering
We only had so many analysts. Statsig provided the necessary tools to remove the bottleneck. I know that we are able to impact our key business metrics in a positive way with Statsig. We are definitely heading in the right direction with Statsig.
Ancestry
Partha Sarathi
Director of Engineering
We use cookies to ensure you get the best experience on our website.
Privacy Policy