US English (US)
ES Spanish

Submit Article Requests

Do you have a suggestion for an article you would like to see created?
Feel free to submit this form and add your suggestions to our document board.

Please fill out the contact form below and we will reply as soon as possible.

  • Integration Hub
  • Contact Us
English (US)
US English (US)
ES Spanish
  • Docs home
  • Analytics & Data
  • Experience and Event Analytics

Run a control experiment

Prove your Flow's impact by measuring conversion lift against a holdout group that never sees it.

Updated at March 26th, 2026

Submit Article Requests

Do you have a suggestion for an article you would like to see created?
Feel free to submit this form and add your suggestions to our document board.

Please fill out the contact form with the details about the help content you'd like to see.

  • Home

  • Getting Started

    • Installation & Developers

      • Web Experiences

        • Mobile Experiences

          • Workflows

            • Analytics & Data

              • Account Management

                • Best Practices

                  • Integrations

                    Table of Contents

                    Prerequisites How it works Create an experiment Configure and launch Pause and resume End an experiment Read the results Send experiment data to integrations

                    A control experiment measures whether a Flow actually drives user behavior by comparing two groups: an exposed group that sees the Flow and a control group that doesn't. The difference in Goal conversion between the groups tells you the Flow's incremental impact.

                    Prerequisites

                    • A published Flow set to Show once
                    • A Goal attached to the Flow
                    • The Flow must not be shown in a Launchpad
                    • Any Frequency Limit on the Flow should be overridden

                    How it works

                    When an experiment is running, Appcues randomly assigns each newly eligible user to one of two groups based on the Traffic Allocation slider (default 50/50):

                    • Exposed group — sees the Flow normally.
                    • Control group — is intentionally blocked from seeing the Flow for the duration of the experiment.

                    Each user gets a random number from 1–100. If that number is less than or equal to the control group percentage, they go into the control group. Because the split is random, group sizes won't be perfectly equal.

                    If a user qualifies for multiple Flows at once, the highest-priority Flow shows regardless of the experiment. This can affect when a user gets assigned to a group.

                    Create an experiment

                    1. Go to Metrics > Experiments (or Experiences > Flows > Experiment).
                    2. Click New experiment.
                    3. Enter an experiment name and select the Flow to test. Flows with an existing Draft, Running, or Paused experiment won't appear — end the previous experiment first.
                    4. Click Create.

                    Configure and launch

                    1. On the experiment settings page, enter your estimates for baseline conversion rate and minimum detectable effect. Appcues uses these to calculate a Minimum Sample Size and Estimated Duration.
                    2. Adjust the Traffic Allocation slider if you want something other than a 50/50 split.
                    3. Click Start. The experiment status changes to Running.
                    4. Make sure the Flow is also published — starting the experiment and publishing the Flow are separate actions.

                    Don't stop the experiment the first time results cross the significance threshold. Early results fluctuate. Wait until you've reached the minimum sample size before drawing conclusions. Check out this post for more information about our approach. 

                    Pause and resume

                    Click Pause in the upper-right corner of the settings page. While paused:

                    • The Flow continues running for all new qualifying users (no one is held back).
                    • No new users are added to either group.
                    • Existing control group users are still blocked from seeing the Flow.

                    Click Resume to restart sampling.

                    End an experiment

                    Click End when you're done collecting data. Once ended:

                    • The Flow continues running normally.
                    • No new users are added to either group.
                    • Control group users become eligible to see the Flow again.
                    • The experiment cannot be resumed.

                    Read the results

                    The experiment report shows:

                    • Users in each group
                    • Goal conversions and conversion rate per group (conversions must occur within 7 days of group assignment). The difference in goal conversion between the exposed and control groups will be used to determine if your Flow is having a positive or negative impact in driving users to take a specific action.
                    • Lift — the percentage difference in conversion between exposed and control groups (appears after minimum sample size is reached)
                    • Confidence — statistical significance at a 95% confidence level (appears after minimum sample size is reached)

                    Before the minimum sample size is reached, a progress bar shows how close you are and an estimated number of days remaining.

                    Export raw data (User ID, group assignment, and conversion times) from the experiments list page.

                    Send experiment data to integrations

                    Appcues emits an experiment_entered event to all front-end integrations whenever a user is sampled. Use this event to segment users in your analytics tools for deeper analysis.

                     

                    appcues experiments controlled tests control group exposed group

                    Was this article helpful?

                    Yes
                    No
                    Give feedback about this article

                    Related Articles

                    • Analyze Flow performance
                    • Export Flow Analytics Data (CSV)
                    • A/B test Experience variations
                    • Set up Goals
                    Appcues logo

                    Product

                    Why Appcues How it works Integrations Security Pricing What's new

                    Use cases

                    Appcues Integration Hub User Onboarding Software Feature Adoption Software NPS & Surveys Announcements Insights Mobile Adoption

                    Company

                    About
                    Careers

                    Support

                    Developer Docs Contact

                    Resources

                    The Appcues Blog Product Adoption Academy GoodUX Case studies Webinar Series Made with Appcues

                    Follow us

                    Facebook icon Twitter icon grey Linkedin icon Instagram icon
                    © 2022 Appcues. All rights reserved.
                    Security Terms of Service Privacy Policy

                    Knowledge Base Software powered by Helpjuice

                    Expand