Skip to content

Variant 7: Understanding the Result of an A/B Test

The Core Concept: The Data is In, Now What?

The experiment has run its course, the data has been collected, and the statistical calculations are complete. The final step in the A/B testing lifecycle is to interpret the results and make a clear, data-informed decision. Every possible outcome—win, lose, or draw—is a valuable learning opportunity that contributes to a deeper understanding of your users.

The Three Possible Outcomes: A Decision Framework

There are three primary outcomes for any A/B test. A structured framework helps in navigating the decision for each.

Outcome 1: Variant B Wins (The Challenger is Better)

This is the most straightforward and celebrated outcome.

  • Condition: The change introduced in Variant B resulted in a positive and statistically significant improvement in the primary metric (e.g., confidence level >95%), without causing significant harm to any critical guardrail metrics.
  • Decision: Implement the winning change (Variant B) for 100% of the relevant user audience. The hypothesis has been validated. The next step is to document the results and learnings in your experiment repository and move on to the next prioritized test on your roadmap.

Pro-Tip: A Win is the Start of a New Question

Don't just ship a winning variant and move on. Ask "Why did this win?" and "How can we double down on this learning?" A winning test that validates a user insight (e.g., "users prefer social proof") is a goldmine for follow-up experiments. Can you apply the same principle to other parts of the user journey? Use the win to build momentum and inform your next set of hypotheses.

Outcome 2: Variant A Wins (The Control is Better)

This outcome, often perceived as a "failure," is in fact a crucial success for the business.

  • Condition: The change in Variant B resulted in a statistically significant negative impact on the primary metric, or it caused a significant and unacceptable decline in a key guardrail metric (e.g., page load time increased dramatically or user retention dropped).
  • Decision: Discard the change proposed in Variant B and stick with the control. The hypothesis has been proven wrong.

This result is incredibly valuable. The experiment has successfully used data to prevent a change that would have harmed the user experience or business performance from being deployed to all users. It saves the company from potential lost revenue, customer frustration, or brand damage. The learning derived from the invalidated hypothesis (e.g., "Our assumption about what motivates users on this page was incorrect") is documented and becomes a critical input for forming a more informed and effective hypothesis for the next experiment.

Pro-Tip: Celebrate "Losing" Tests Publicly

The most important cultural contribution you can make is to celebrate a "losing" test as a major success. Frame it as: "This experiment just saved us three months of engineering work building something our users didn't want." When you share the results, emphasize the invalidated assumption and the resources saved. This creates psychological safety and encourages your team to take bigger, smarter risks in the future.

Outcome 3: Inconclusive Result (No Clear Winner)

This is one of the most common outcomes in A/B testing, with some industry data suggesting that 50% to 80% of all tests may end up as inconclusive. For beginners, this can be discouraging, but for seasoned practitioners, it is a rich source of information.

  • Condition: The test did not find a statistically significant difference in performance between Variant A and Variant B. The confidence level is low, the uplift is negligible, or the confidence interval for the change includes zero.

An inconclusive result does not mean the test was a waste of time. It provides a clear piece of information: the change that was tested did not have a meaningful impact on the user behavior it was intended to influence. This information is actionable and leads to a nuanced decision tree rather than a single path forward.

Pro-Tip: Inconclusive Often Means "Go Bolder"

A high percentage of inconclusive results is often a sign that your hypotheses are too timid. If you're consistently testing minor copy changes or subtle color tweaks, you're likely to get flat results. An inconclusive test is a great prompt to ask: "Was our change significant enough to even be noticed by the user?" Use it as a catalyst to test more radical, impactful changes.

  • Action A (Default): Stick with the Control. This is the most common and logical course of action. If the proposed change requires development effort to implement and maintain, and the experiment has shown that it provides no detectable benefit to the user or the business, there is no data-driven reason to proceed. Avoid introducing change simply for the sake of change.
  • Action B (Strategic Rollout): Implement the Variation. In some cases, Variant B may be preferable for reasons outside the primary metric. For example, it might align with a new brand design system, simplify the codebase, or improve accessibility. If the A/B test has shown that this strategic change is unlikely to cause harm to the primary metric, the experiment has successfully de-risked the decision. The team can proceed with the rollout, confident that it will not have a negative business impact.
  • Action C (Deeper Analysis): Segment the Results. Before making a final decision, it is always wise to analyze the inconclusive result across different user segments. An overall flat result can often hide a significant win in one segment (e.g., new users on mobile devices) being cancelled out by a loss in another (e.g., returning users on desktop). Uncovering such a pattern does not provide a final answer, but it generates a powerful new hypothesis for a more targeted follow-up experiment (e.g., "Let's test this change only for new mobile users").

Properly handling inconclusive results is a key differentiator between amateur and mature experimentation programs. It requires a shift from a simple "win/loss" mindset to a more sophisticated "learn/iterate" culture. The true value of A/B testing lies not just in finding winners, but in systematically understanding what doesn't work, which saves invaluable resources and sharpens the focus of future efforts on more impactful ideas.

Pro-Tip: Segment to Hypothesize, Not to "Find a Win"

When you slice an inconclusive result into many segments, you dramatically increase the odds of finding a "winner" by pure chance. This is a dangerous form of p-hacking. Treat post-test segmentation as a hypothesis-generation tool, not a result-finding tool. If you find a surprising lift in a specific segment (e.g., "new users on Android"), the correct action is not to ship the change to them. The correct action is to form a new hypothesis ("This change is effective for new Android users because...") and run a new, targeted experiment to validate it.

Key Takeaways

  • A Win is a Green Light: If the variant wins with statistical significance and no guardrail metrics are harmed, implement the change. But don't stop there—use the validated learning to inform your next hypothesis.
  • A Loss is a Success: When the control wins, you've successfully prevented a change that would have harmed your business. This is a valuable outcome. Document the invalidated assumption and celebrate the learning.
  • Inconclusive Means "No Impact": An inconclusive result is not a failure; it's proof that your change didn't matter to users. This saves you from investing engineering resources in a low-impact feature.
  • Default to the Control: When a test is inconclusive, the default action is to stick with the control. Don't introduce change for the sake of change if it provides no proven benefit.
  • Segment to Learn, Not to Win: Use post-test segmentation on inconclusive results to generate new, more targeted hypotheses for your next test. Do not use it to "find a winner" in a specific segment and ship the change.

Remember This Even If You Forget Everything Else

In a healthy experimentation culture, there are no "failed" tests, only learnings. A "losing" test saves you from a mistake, and an "inconclusive" test saves you from wasting effort. Both outcomes are valuable wins that make your next hypothesis smarter.

7 min read
Text-to-Speech is not supported in your browser.