Geo Testing for Marketing: How to Run Location-Based Experiments
Most marketing tests are straightforward. You show version A to half your audience, version B to the other half, measure which wins, and move on. That works fine for testing creatives, landing pages, or offer variations.
But there's a whole category of questions this approach can't answer. What would actually happen if you turned off paid search tomorrow? How much revenue does your TV budget really drive? Would ditching promotional emails actually hurt repeat purchases?
These questions need geo testing. You run marketing normally in some geographic markets while deliberately changing or pausing it in others, then measure how customer behavior actually shifts. It's the closest thing marketing gets to a controlled lab experiment.
This article covers what geo testing is, how to design one that produces real answers instead of noise, how to interpret the results, and whether it's the right tool for what you're trying to measure.
What Geo Testing Is and Why It Actually Works
Geo testing is a form of causal measurement. Instead of trying to reverse-engineer impact from historical data, you deliberately change something in specific places and observe what happens.
The logic is dead simple. Pause paid search in Chicago but keep it running everywhere else. If Chicago's sales drop while other markets stay flat, that drop almost certainly came from pausing search.
What makes geo testing powerful is that it measures incremental impact without needing to solve attribution. You don't have to figure out which touchpoint caused which conversion. You just measure whether your overall business outcome changes when you change your marketing.
Why Geo Testing Beats Attribution for This Kind of Question
Attribution models assign credit to touchpoints based on their order in a customer's journey. The problem is ambiguity. Did that touchpoint actually cause the conversion, or was it just sitting there when the conversion happened?
Geo testing removes that ambiguity entirely. You make an intervention (pause a channel in test markets) and observe the real business impact. That's about as close to scientific causal proof as marketing gets.
When Geo Testing Actually Works
Geo testing depends on a few conditions:
- Clear geographic boundaries where you can actually control marketing spend
- Test and control markets that are reasonably similar (so you're comparing apples to apples)
- Enough customer volume that meaningful differences won't get buried in noise
- Enough time for the effect to show up
- Low customer overlap between test and control areas (people aren't living in Denver but shopping from Phoenix)
Geo testing shines for:
- Channel-level impact (what happens if Facebook just disappears?)
- Budget-level impact (what if we cut Search budget in half?)
- Campaign-type impact (do awareness campaigns actually move the needle on sales?)
- Seasonal strategies (do holiday promotions actually drive incremental sales, or are people just buying anyway?)
- New channel experiments (does TikTok drive actual sales or just vanity metrics?)
Where geo testing falls apart:
- Individual creatives or small tactical changes (not enough volume concentrated enough)
- One-day flash sales or limited-time offers (effect expires before you can measure)
- New markets where you're still building brand awareness (you can't separate the channel test from normal market development)
- Online-only products with global shipping (customers move around too much)
Designing a Geo Test That Actually Produces Answers
A good geo test produces clear results. A bad one produces a lot of confusing noise that could mean anything.
Step 1: Define What You're Actually Testing
Before you launch anything, write down exactly what you want to measure. "What's the incremental impact of paid search?" beats "Is search worthwhile?" The first one is measurable. The second one is too fuzzy.
A solid hypothesis looks like this: "If we pause all paid search campaigns in the Denver metro area for eight weeks while keeping search running everywhere else, Denver's online sales will drop by 15 to 20 percent. That drop represents the incremental impact of our search spending."
That hypothesis needs to include:
- The intervention (pause search)
- The test market (Denver)
- How long (eight weeks)
- What you expect to happen (15-20% sales drop)
- What you're measuring (online sales)
Step 2: Pick Test and Control Markets Smartly
Your test market is where you change something. Your control market is where everything stays the same. The whole point is making them so similar that any differences you see are actually caused by your change.
What makes a good market pair:
Size: Needs enough volume that you can spot effects within 4-12 weeks. Too small and random variation drowns out real signals. Test markets should be at least 5-10% of your overall revenue.
Similarity: Test and control should have comparable customer demographics, buying patterns, and seasonal trends. If you're testing search impact in Denver, test against Phoenix, not New York.
Independence: Minimal customer overlap. If people live in Denver but shop from Phoenix addresses, they'll muddy your results. The less people travel between your test and control areas, the cleaner your test is.
Isolation: You need to actually control marketing spend by geography. If your paid search platform doesn't let you segment spending geographically, you can't run this test.
Step 3: Use Multiple Control Markets
A single control market can get hit by random local stuff you didn't see coming (a competitor promotion, freak weather, local news coverage). Use 2 to 3 control markets to smooth out those local anomalies.
The bigger your test and control footprints, the more random noise averages out, and the more confident you can be that differences come from your intervention.
Step 4: Measure Your Baseline First (Before Anything Changes)
Spend 4 to 8 weeks measuring test and control markets before you change a single thing. Track:
- Average daily revenue per market
- Week-over-week growth rate
- Customer acquisition cost
- Conversion rate
- Average order value
This baseline matters a lot. It tells you whether test and control are actually performing similarly. If they're already wildly different, pick different control markets and try again.
While you're at it, note seasonality patterns. Are there specific days (weekends, paydays) when one market outperforms? Any seasonal trends? Write this down so you can account for it when your test runs.
Step 5: Run the Test Long Enough
How long depends on how often customers buy:
- Fast-moving products (fast fashion, groceries): 4 weeks is usually enough
- Medium-frequency products (electronics, furniture): 6 to 8 weeks
- Low-frequency products (cars, luxury goods): 12 to 16 weeks
Longer tests are better (more data, less noise), but they cost more and delay decisions. Run long enough to catch real effects without dragging it out unnecessarily.
Also make sure you capture full weekly cycles. If you start on a Wednesday, run the test for at least 4 weeks so you see weekday and weekend patterns in both markets.
Step 6: Monitor Throughout and Track What Could Mess Things Up
Week by week, watch:
- Revenue in test versus control
- Customer acquisition in test versus control
- Repeat purchase rate in test versus control
- Average order value in test versus control
Also watch for things that could skew results:
- Competitor activity happening in one market and not the other
- Weather or local events affecting buying patterns
- Supply chain issues affecting inventory in one market but not others
- Press coverage or PR in either market
- Website changes, checkout flow changes, or product offering changes (these need to be identical across both markets)
- Random technical problems (site outages, payment failures) in one market
If something major happens, document it thoroughly. It might explain part of what you're seeing.
Making Sense of Your Results
A solid geo test gives you numbers. Test market sales were $500K, control was $600K. Test dropped 17%, control dropped 2%. What do you actually do with that?
Isolating What Your Change Actually Caused
The number you care about isn't the absolute difference in revenue. It's the difference in how test and control markets changed during your test period.
Use this formula: Incremental Impact = (Control Growth - Test Growth)
If control markets grew 10% week over week and test markets declined 5%, your causal impact is roughly 15%. You've shown that your intervention (pausing search) reduced growth by 15 percentage points.
Why this formula? Both markets are affected by overall business trends, seasonality, and macro conditions. By comparing how much each one changed, you isolate the effect of your specific intervention.
Is This Result Real or Just Noise?
That depends on sample size and volatility.
With big customer volumes (over 10,000 purchases in your test market during the test period), even small percentage differences are usually real. With small volumes, even large percentage differences might just be random variation.
Use an online calculator to check whether your difference is statistically significant. Rough rule of thumb: if your test market had 5,000 or more purchases during the test period and you see a 5% or larger difference in conversion rate or revenue between test and control, that result is probably real.
Short-Term Results versus Long-Term Reality
Your geo test shows what happens in the immediate 4 to 12 weeks. But some effects take longer to kick in.
If you pause brand-building (awareness, consideration), the immediate impact might be tiny because brand awareness doesn't directly drive today's sales. But 6 to 12 months later, organic search volume declines, fewer people remember your brand, and sales gradually erode.
Geo tests are good at measuring immediate channel effects but can underestimate the long-term value of brand activities. Combine them with media mix modeling for a more complete picture.
Synthetic Control Methods: A More Sophisticated Approach
In a perfect world, you have similar markets and can randomly pick one as test and one as control. Reality is messier. Sometimes your markets are genuinely different.
Synthetic control methods solve this by creating a virtual control market from a weighted combination of multiple real control markets. The goal is building a control that mirrors your test market's pre-test behavior more closely.
Example: You're testing Denver. You could build a synthetic control from 40% Phoenix data, 30% Austin data, and 30% Denver's own pre-test trend. That synthetic control probably looks more like Denver than Phoenix or Austin alone would, making your comparison stronger.
Synthetic controls need more statistical sophistication but pay off when your markets are naturally dissimilar.
Geo Testing versus Other Measurement Methods
Geo Testing versus Attribution
Attribution answers: "Which touchpoints appeared in this customer's journey?" Geo testing answers: "What's the actual incremental impact of this channel?"
Use geo testing when you care about incremental impact. Use attribution when you're trying to understand how customers moved through their journey.
Geo Testing versus Incrementality Testing
Incrementality testing (holdout groups) randomly prevents some users from seeing a channel. Geo testing prevents users in specific locations from seeing a channel. Both answer similar questions.
Go with geo testing when:
- Individual-level test would be too granular
- You want business-level outcomes (sales) instead of event-level metrics (conversions)
- You're testing entire channels at scale, not small tactical changes
Go with incrementality testing when:
- You need results faster (4 to 7 days instead of 4 to 8 weeks)
- You want channel-level holdouts (block 1% of users)
- You're testing tactical tweaks, not strategic shifts
Geo Testing versus Media Mix Modeling
MMM answers: "What's the historical elasticity of each channel?" Geo testing answers: "What actually happens if I change this channel right now?"
Both are useful. Use MMM to understand long-term trends and seasonal patterns. Use geo testing to validate what MMM tells you and measure specific strategic changes.
How to Actually Run Your First Geo Test
1. Start Small
Don't test "pause all paid search" on your first go. That's too risky. Test something smaller like "cut search budget by 25%" or "pause one campaign type."
This limits downside and lets you learn the process without betting the business.
2. Choose Markets That Actually Match
Find test and control markets that genuinely perform similarly. Use demographic data and purchase history to confirm. If you can't find legitimately similar markets, your test will be too noisy to trust.
3. Plan Around Seasonality
Run your test during a period where you have solid baseline data already. Stay away from peak seasons (Black Friday, Cyber Monday) where there's too much noise and effects are temporary anyway.
4. Decide Your Stopping Points in Advance
Define what results would make you extend, pause, or reverse the change before you start. Something like "If revenue drops more than 10%, we revert immediately." This keeps emotions out of decision-making.
5. Bring Your Team Along
Tell people this is an experiment. Track results weekly. Share what you're seeing. If the test proves your hypothesis wrong, that's valuable information, not a failure.
6. Keep Records
Document the test design, the markets, the baseline numbers, the dates, what you changed, and what happened. Future tests will run faster if you have a template and a history to reference.
Platforms That Support Geo Testing
Most major ad platforms let you control and report spending by geography:
- Google Ads: Spend control by location, revenue reporting by market
- Facebook/Meta: Campaign budget rules by country and region
- Amazon Ads: Region-level controls
- Custom platforms: Tools like ORCA combine sales data, customer data, and marketing spend across geographies in one view, making it easier to measure incremental impact
ORCA's measurement approach simplifies geo testing by pulling sales data, customer data, and marketing spend across all your geographies into a single interface, so isolating causal effects happens faster.
The Mistakes You'll See (And How to Avoid Them)
Mistake 1: Testing in markets that are too small. A 1-2% revenue market has too much noise relative to signal. Use markets representing at least 5-10% of your total revenue.
Mistake 2: Ignoring confounds. A competitor runs a sale in your test market or there's weird weather. That skews results. Document every possible confound and measure its impact.
Mistake 3: Testing during peak seasons. The holidays are noisy, and effects are temporary. Run tests during stable periods.
Mistake 4: Cutting the test too short. Two or three weeks isn't enough for most products. Run for at least 4 to 8 weeks so you capture full customer cycles.
Mistake 5: Treating test results as permanent. Geo tests show 4 to 12 week effects. Long-term effects can differ (especially for brand-building). Pair geo tests with media mix modeling.
Building Your Testing Program
Aim for one geo test per quarter. Pick your top 3 to 4 channels or campaign types and measure their incremental impact. Over time, you'll build a real understanding of what actually drives incremental revenue for your business.
This experimental evidence, combined with attribution and media mix modeling, creates a measurement framework that purely historical analysis can never match. You'll allocate budgets with actual confidence instead of just guessing.
Ready to actually measure incremental channel impact? ORCA unifies your sales data, customer data, and marketing spend across geographies so you can run cleaner geo tests and move faster. See how brands are using geo testing to validate real marketing impact and make budget decisions that stick.
Related Reading
Tagged in: