Testing and launching an experiment

Last updated:

|Edit this page

Once you've written your code, it's a good idea to test that each variant behaves as you'd expect. If you find out your implementation had a bug after you've launched the experiment, you lose days of effort as the experiment results can no longer be trusted.

The best way to do this is adding an optional override to your release conditions. For example, you can create an override to assign a user to the test variant if their email is your own (or someone in your team). To do this:

  1. Go to your experiment feature flag.

  2. Ensure the feature flag is enabled by checking the "Enable feature flag" box.

  3. Add a new condition set with the condition to email = your_email@domain.com. Set the rollout percentage for this set to 100%.

    • In cases where email is not available (such as when your users are logged out), you can use a parameter like utm_source and append ?utm_source=your_variant_name to your URL.
  4. Set the optional override for the variant you'd like to assign these users to.

  5. Click "Save".

Once you test it works, you're ready to launch your experiment.

Note: The feature flag is activated only when you launch the experiment, or if you've manually checked the "Enable feature flag" box.

Note: While the PostHog toolbar enables you to toggle feature flags on and off, this only works for active feature flags and won't work for your experiment feature flag while it is still in draft mode.

Viewing experiment results

While the experiment is running, you can see results on the page:

Viewing experiment results

This page shows its status, whether draft, running, or complete as well as whether it has reached significance or not. It will also provide advice on how to interpret the results.

Beyond this, you can see a summary of the experiment including goals, variants, data collection, and more.

Sometimes, in the beginning of an experiment, results can be skewed to one side while enough data is still being gathered. While peeking at preliminary results is not a problem, making quick decisions based on them is problematic. We advise you to wait until your experiment has had a chance to gather enough data.

Advanced results metrics

  • Mean: For trends, the number who completed the goal divided by the number who were exposed.

  • Delta %: The percentage change between the control and test variants.

  • Credible interval: Represents a range within which we believe the true difference between the test variant and the control means lies, with 95% probability. In this context, the interval is expressed as a percentage change, indicating how much higher or lower the mean of the test variant could be compared to the control, based on the observed data and our prior beliefs.

  • Win probability: The probability that variant is the best performing one.

Ending an experiment

When you're ready to end an experiment, you can click the Ship a variant button on the experiment page to roll out a variant and stop the experiment. (Note: This button only appears when the experiment has reached statistical significance.)

Ship a variant

If you want more precise control over your release, you can also set the release condition for the feature flag and stop the experiment manually.

Beyond this, we recommend:

  1. Sharing the results with your team.

  2. Documenting conclusions and findings in the description field your experiment. This helps preserve historical context for future team members.

  3. Removing the experiment and losing variant's code.

  4. Archiving the experiment.

Further reading

Want to learn more about how to run successful experiments in PostHog? Try these tutorials:

Questions?

Was this page useful?

Next article

Troubleshooting and FAQs

How do I use an existing feature flag in an experiment? We generally don't recommend this, since experiment feature flags need to be in a specific format (see below) or otherwise they won't work. However, if you insist on doing this (for example, you don't want to make code change), you can do this for multiple variant feature flags only by doing the following: Delete the existing feature flag you'd like to use in the experiment Create a new experiment and give your feature flag the same key…

Read next article