Skip to content

Experiment Decisions FAQ

Making the right call on your experiments can be tricky. This FAQ covers the three most common scenarios you’ll encounter and how to handle each one.

Scenario 1: Low Statistical Significance After a Long Run

Section titled “Scenario 1: Low Statistical Significance After a Long Run”

My experiment has been running for weeks but still shows no statistical significance. What should I do?

Section titled “My experiment has been running for weeks but still shows no statistical significance. What should I do?”

This is one of the most common situations. If your experiment has been running for an extended period (typically 2-4 weeks) without reaching statistical significance, it usually means one of two things:

  1. There’s no meaningful difference between your variants
  2. The effect size is too small to detect with your current traffic

Recommended actions:

Traffic LevelRecommendation
High traffic (10k+ users/day)If no signal after 2 weeks, the variants likely perform similarly. Consider shipping your preferred variant based on other factors (brand voice, simplicity).
Low traffic (under 1k users/day)Extend the experiment to 4-6 weeks before deciding. Small sample sizes need more time.

Not indefinitely. Set a maximum runtime when you launch (we recommend 4 weeks for most experiments). If you haven’t reached significance by then, make a decision based on:

  • Directional trends — Is one variant consistently (even if not significantly) better?
  • Secondary metrics — Are there meaningful differences in other metrics like engagement or retention?
  • Business priorities — Do you need to move on to test other hypotheses?

Tip: A “no significant difference” result is still valuable. It tells you this particular change doesn’t meaningfully impact your key metric, freeing you to focus elsewhere.

Fluctuating results often indicate high variance in your metric or external factors (seasonality, marketing campaigns) affecting performance. Consider:

  1. Extending runtime to smooth out variance
  2. Segmenting results by attribute to see if specific audiences respond differently
  3. Checking for data quality issues — are events firing correctly?

I received an Auto-Tune notification. What does this mean?

Section titled “I received an Auto-Tune notification. What does this mean?”

An Auto-Tune notification means JustAI has detected that one or more variants are performing significantly better than others. Auto-Tune has automatically started shifting more traffic toward the winning variant(s).

What’s happening behind the scenes:

  • JustAI continuously monitors variant performance
  • When a variant shows statistically significant improvement, Auto-Tune increases its traffic allocation
  • Poor-performing variants receive less traffic, minimizing opportunity cost

Do I need to take action when I get an Auto-Tune notification?

Section titled “Do I need to take action when I get an Auto-Tune notification?”

Not immediately. Auto-Tune is designed to optimize automatically. However, you should:

  1. Review the results — Open the experiment dashboard to see which variant is winning and by how much
  2. Check the metrics — Confirm the winning variant aligns with your goals
  3. Monitor for stability — Watch for a few more days to ensure the winner remains consistent

Auto-Tune uses statistical methods to make decisions, but early signals can occasionally shift. That’s why Auto-Tune adjusts traffic gradually rather than switching 100% immediately. If the signal was a false positive, the system self-corrects.

When to intervene:

  • If the “winning” variant has unintended consequences (e.g., higher clicks but more unsubscribes)
  • If you notice data quality issues affecting results
  • If business context has changed (e.g., a variant references an expired promotion)

No. Even with Auto-Tune, you should eventually ship a winner. Use Auto-Tune to:

  1. Minimize losses while the experiment runs
  2. Gather confidence in the winning variant
  3. Learn which themes and approaches work for different segments

Once Auto-Tune has clearly identified a winner (typically 80%+ traffic allocation), consider shipping it permanently.


I received a Ship notification. What does this mean?

Section titled “I received a Ship notification. What does this mean?”

A Ship notification means JustAI has high confidence that a winning variant has been identified and recommends you ship it as the permanent version. This notification appears when:

  • A variant has shown consistent, statistically significant improvement
  • The result has been stable over time (not a temporary spike)
  • There’s sufficient sample size to trust the result

What should I do when I get a Ship notification?

Section titled “What should I do when I get a Ship notification?”
  1. Review the winning variant — Open the experiment to see performance details
  2. Check the lift — Understand how much improvement you’re getting (e.g., +12% click rate)
  3. Verify segment performance — Ensure the winner works across your key audiences
  4. Ship it — Click “Ship Winner” to make this variant the permanent version

When you ship:

  • The winning variant becomes the default for all users
  • The experiment ends and stops collecting data
  • You can no longer revert to other variants (without creating a new experiment)
  • Your template is updated automatically

Yes, but it’s not recommended for long. Reasons you might delay shipping:

ReasonRecommendation
Want more confidenceLet it run another week, but don’t wait indefinitely
Winner works for some segments but not othersConsider creating segment-specific templates instead of shipping globally
External factors (holidays, campaigns)Wait until normal conditions resume, then verify the winner still holds

Warning: Delaying shipping means you’re leaving performance gains on the table. If JustAI recommends shipping, the data strongly supports it.

Section titled “What if I disagree with the recommended winner?”

Trust the data, but consider context. If you have strong reasons to doubt the result:

  1. Check for data issues — Are events tracking correctly?
  2. Review secondary metrics — Is the winner causing problems elsewhere?
  3. Consider qualitative factors — Does the winner align with brand guidelines?

If the data is solid but you still prefer a different variant, you can manually ship your preferred choice. Just document your reasoning for future reference.


SituationSignalRecommended Action
No significance after max runtimeNoneShip based on directional trend or preference
Auto-Tune notificationEmerging winnerMonitor, let Auto-Tune optimize, plan to ship soon
Ship notificationClear winnerReview and ship the winner
Fluctuating resultsUnstableExtend runtime, check data quality, segment analysis
Winner in some segments onlyMixedConsider segment-specific templates