Google Ads Experiments Auto-Apply by Default Now. Here's the Setting to Flip
In 2026, Google Ads quietly made auto-apply the default behavior for experiments. A favorable result, judged on directional data and your two chosen metrics, can now shift full traffic from your control campaign to the experiment with no human approval. The confidence and guardrail controls exist, but only if you find them and turn them on.
I'll save you the suspense: the default is doing more harm than good on most accounts I would want to manage. The fix takes about four clicks. The reasoning behind why you'd flip it is what actually matters, because the same logic decides how you should set up every test from here forward.
What changed, and the announcement nobody really made
Before the 2026 update, experiments worked the way most practitioners still imagine: you built an experiment campaign as a draft, ran it for some window, looked at metrics, then manually decided whether to apply the variant to the original. Now eligible experiments default to auto-apply favorable results, pushing the experiment campaign live as soon as Google's evaluation logic decides the test arm is winning. Search Engine Land covered the rollout, and ALM Corp's writeup noted what mattered most: it shipped as a default, not an opt-in.
Two things make this less benign than the announcement framing suggests. First, the default evaluation method is "directional results," which is a softer signal than statistical significance. Second, you only get two success metrics. A third metric you care about, one you didn't or couldn't select, can decline unnoticed while Google waves the experiment through.
If you've been running experiments for a while, your mental model probably still has a manual "apply to base" step at the end. That step is gone now unless you go put it back. Worth checking before your next test ends, not after.
Directional results vs statistical significance, and why the default is the risky one
Google's experiments engine offers two ways to call a winner. The official methodology page describes "directional results" as observed differences without stricter confidence requirements, while the harder thresholds are 80%, 85%, or 95% confidence.
Directional results are like a coin landing on heads three times in a row. Suggestive. Not conclusive. They tell you which way the data is leaning, but "leaning" is doing a lot of work in that sentence. As PPC Epiphany has documented, significance tends to come and go as data accumulates, especially in lower-volume accounts. A directional result on day 9 can flip on day 14, and if auto-apply has already moved traffic, you've baked the false signal into your live setup.
From what I've seen across small-to-mid ecommerce accounts, directional results tend to land when one arm is genuinely better. They also tend to land when one arm just got lucky with a few high-AOV orders in the first week. Those two scenarios look identical in the experiment dashboard. Treat them differently or pay later.
Rough rule of thumb I've been using: on accounts with under 5,000 monthly conversions, expect at least one out of every four directional "wins" to reverse if you let the test run another two weeks. That number is not from a Google study, just my own pattern across the last year. Run your own audit if you want a more confident answer for your account.
If your experiment is testing something material (bid strategy change, audience swap, a new portfolio bid strategy), set the threshold to 95% confidence. Faster decisions are nice. A wrong bid strategy quietly running for two weeks is more expensive than a slower experiment.
Where to actually flip the switch
This is the part most "10 minutes of action" articles skip. The setting lives in two places, and they don't behave the same way.
For new experiments: when you build the experiment, scroll to the auto-apply section before clicking Save. You'll see a toggle for "Apply favorable results automatically" and a confidence dropdown. Uncheck the toggle to require manual review at the end. If you want to keep auto-apply on, set the confidence to 95%, not directional.
For experiments already running: open the Experiments page, click into the active experiment, and edit the auto-apply settings. Changes take effect for future evaluations, not retroactively.
For account-wide auto-apply on recommendations (separate but related): go to the Recommendations page within the Campaigns menu, then Auto-apply settings, then uncheck anything you didn't explicitly opt into. Google's help doc on managing auto-apply walks through this, and the History tab inside that panel will tell you what's already been applied without your sign-off. On inherited accounts, that history tab is usually a small horror show.
Worth noting: as of January 26, 2026, the "Add responsive search ads" recommendation no longer auto-applies new RSAs, per Google's own changelog. So if you remember turning that one off years ago, the lever has actually moved.
The two-metric trap most teams will walk into
Experiment success metrics in Google Ads are capped at two. That's a real constraint, not a UX choice. The auto-apply system evaluates only those metrics, so a test that improves CPA and conversion volume can still degrade ROAS if order value drifts down, and Google will promote it.
Order value is exactly the kind of "third metric" that bites accounts where AOV varies a lot by audience or product. Auto-apply on, plus directional results, plus two metrics, is three dice rolls in a row. Each one is fine in isolation. Stacking them is how you end up with a campaign that looks better and earns less.
This connects directly to a problem we covered in Target ROAS vs Target CPA: Order-Value Spread Decides It Before You Do: when you optimize on CPA, the algorithm doesn't owe you any particular order-value distribution. Run a test where the experiment arm is good at finding cheap conversions and you'll get a "win" on paper while AOV slides 9% sideways.
The practical fix: pick your two metrics deliberately. CPA alone is almost never enough. Pair it with revenue per click or conversion value per impression. If you genuinely care about three things, run two separate experiments rather than one broad one and let each test focus.
The selection moment matters more than people think. You're effectively telling Google's auto-apply system, "these are the only two outcomes I'm willing to bet the campaign on." Anything you leave off the list, the system treats as out of scope. Cost per acquisition rising on net-new buyers while remaining flat overall is a classic version of this. So is brand search cannibalization showing up in the experiment as "more conversions" while organic clicks fall.
Set a hard floor outside the experiment too. Pull a saved view in the main campaigns table that filters to (a) the experiment campaign and (b) any KPI you care about that isn't in the experiment metrics. Check it at the cadence you'd normally check pacing. If a non-tracked KPI breaks while the experiment "wins," you'll catch it manually before it auto-applies.
If you want to keep auto-apply on, configure it like this
I'm not going to argue auto-apply is always wrong. For high-volume accounts running low-risk tests (small bid adjustments, minor audience swaps), it genuinely saves time. The bar isn't "off entirely." The bar is configured so wins that aren't real don't ship.
A reasonable default setup, if you want one:
- Confidence threshold at 95%, not directional.
- Two well-chosen success metrics, with at least one of them value-based.
- Minimum experiment duration of 4 weeks. Google itself recommends 4 to 6 weeks when results are still inconclusive.
- 50/50 traffic split unless you have a specific reason to skew it.
- A saved-view dashboard tracking the metrics that aren't in the experiment.
For B2B or longer-cycle accounts, push duration to 6 to 8 weeks and require at least 500 conversions per arm before treating the result as real. That's roughly the threshold GrowthSpree highlights for detecting meaningful improvements in low-conversion-rate environments.
One more lever worth pulling. Google defaults experiment traffic to 50/50, which is the right call if you trust the test. If the change is risky and you have over 100 conversions in the test window, dropping to 70/30 or even 80/20 in favor of the control protects revenue while still feeding the test. Google's setup documentation mentions this explicitly. It's a small lever and most accounts I see never touch it.
The five-minute audit worth doing today
If you do nothing else after reading this:
- Open the Experiments page on each active account.
- For every running experiment, set confidence to 95% or uncheck auto-apply entirely.
- Open the Auto-apply settings under Recommendations and uncheck anything you don't actively want.
- Click the History tab in that same panel and skim the last 30 days. Anything you didn't authorize is a conversation worth having internally.
The broader rule I keep coming back to: Google Ads now ships defaults that assume you trust the platform's judgment more than your own. Sometimes that is fine. For experiments, where the entire point is collecting evidence before changing live campaigns, the default is working against the design of the feature.
If you're auditing an account end-to-end and want a fuller checklist, our 30-minute Google Ads audit walks through the rest. And the broader Google Ads strategy guide goes deeper on staying in control of automation rather than fighting it after the fact.
Go change the setting. Then check what's already been applied without you asking.
Notice Me Senpai Editorial