Hi there,
Thanks for reaching out!
Happy to give you some initial thoughts on your question about testing on Facebook. It's a really common point of confusion, and tbh, Meta doesn't make it as straightforward as it could be. You're right to be asking how to do this properly, as getting your testing methodology right from the start will save you a lot of wasted ad spend down the line. Let's get into it.
TLDR;
- For testing creatives, don't use the formal A/B test feature. Just put your multiple ads into a single ad set and let Facebook's algorithm (Advantage+ Creative) figure out the winner.
- The formal A/B test tool splits your audience, which is inefficient for creative testing. It's better for testing bigger variables like audiences or landing pages.
- Your focus should be less on the tool and more on what you're testing. The biggest wins come from testing your core offer and messaging, not just changing button colours.
- Structure your testing with a clear hypothesis. For example: "I believe a UGC-style video will outperform a static image because it builds more trust."
- This letter includes an interactive calculator to help you determine if your test results are statistically significant, which is a massive step up from just guessing.
We'll need to look at the A/B Test Tool vs. Just Using Multiple Ads...
Okay, so let's clear this up straight away. The reason you see two options is because they do fundamentally different things, even though they sound similar. Understanding the difference is probably the most important first step.
When you create multiple ads within one ad set, you're essentially telling Facebook, "Here are three different creatives. Here is one bucket of people (your audience). Please use your algorithm to spend my money as efficiently as possible to get me the best results, showing the best performing ad more often." The ads all compete against each other in real-time for the same audience. Over time, the algorithm will naturally start to favour the creative that's getting better results (lower CPA, higher CTR, whatever you're optimising for) and allocate more of the budget to it. This is now largely automated with a feature called 'Advantage+ Creative', which you should see toggled on by default. For your goal – finding the best creative – this is almost always the method you should use. It's faster, more efficient, and leverages the power of the machine learning you're paying for.
The formal "A/B Test" feature, on the other hand, is a much more rigid, scientific tool. When you set up an A/B test to compare, say, Creative A vs. Creative B, Facebook does something different. It takes your total audience and splits it into two distinct, non-overlapping groups. Group 1 will *only* ever see Creative A. Group 2 will *only* ever see Creative B. The system then runs the ads for a set period until it can declare a statistically significant winner based on your chosen metric. It's a true experiment.
So why wouldn't you use this? Well, for testing creatives, it's often overkill and inefficient. By splitting your audience, you're preventing the algorithm from optimising budget in real-time. You're forcing it to spend equally (or proportionally) on both versions, even if one is clearly a dud from day one. This means your overall results during the test period will almost certanly be worse than if you just let the algorithm optimise within a single ad set. The A/B test tool is much more useful for testing bigger, more strategic variables where you need a clean read. For example:
- -> Audience A vs. Audience B
- -> Landing Page 1 vs. Landing Page 2
- -> Campaign Objective: Conversions vs. Traffic
- -> Offer 1 (e.g., 10% off) vs. Offer 2 (e.g., Free Shipping)
For just checking three different creatives? Stick them in one ad set and let the algorithm do the heavy lifting. It's what it's designed for.
I'd say you're asking the wrong question, though...
Here's the bit of brutally honest advice. Focusing on the technical difference between these two tools is fine, but it's a bit like asking what's the best type of hammer to use when the blueprint for the house is wrong. The tool you use to test is far less important than what you are actually testing.
Most advertisers waste thousands of pounds testing meaningless variations. They'll test a blue button vs. a green button, or a slightly different headline, and then wonder why their results don't improve. That's because they're testing tactics, not strategy. The biggest performance leaps don't come from minor creative tweaks; they come from nailing the fundamentals: your audience and your offer.
Before you even think about creating an ad, you need to have a crystal clear picture of your Ideal Customer Profile (ICP). And I don't mean a vague demographic like "women aged 25-40 who like yoga". That tells you nothing. You need to understand their nightmare. What is the specific, urgent, expensive problem that keeps them awake at night? What are they secretly afraid of? What is the deep desire they have that your product or service fulfills?
I remember one of our B2B SaaS clients wasn't selling 'workflow automation software'. They were selling a solution to the Head of Engineering's fear of his best developers quitting because they were bogged down in manual, repetitive tasks. Their ads didn't talk about features; they talked about retaining top talent. That's a message that cuts through the noise. Your advertising fails or succeeds long before you open Ads Manager. It succeeds when you understand the customer's pain so deeply that your ad feels less like an interruption and more like a solution they've been searching for.
Once you have that 'nightmare' defined, your offer must be the perfect antidote. The biggest mistake I see is a mismatch between the ad's promise and the landing page's offer. If your ad promises a simple solution, but the landing page is a high-friction "Request a Demo" form, you've already lost. The offer needs to provide immediate, undeniable value. For a SaaS product, this is a free trial. For an agency, it could be a free, automated audit tool. You must solve a small piece of their problem for free to earn the right to solve the whole thing for them.
You probably should treat your ad like a hypothesis...
So, you want to test three creatives. Great. But don't just throw three random images into an ad set. That's not testing; it's gambling. Proper testing is scientific. You need a hypothesis for each creative.
A hypothesis isn't just a guess. It's a statement that connects a specific change to an expected outcome, with a reason why. For example:
- Bad approach: "Let's test this picture, this video, and this other picture."
- Good approach (Hypothesis-driven):
- -> Creative A (Static Image): "I believe a high-quality product shot will perform best because our audience values professional aesthetics."
- -> Creative B (UGC Video): "I believe a user-generated-style video testimonial will perform best because it builds social proof and feels more authentic, which should increase trust and lower CPA."
- -> Creative C (GIF/Short Animation): "I believe a short, attention-grabbing animation highlighting the key benefit will perform best because it will stop the scroll more effectively than a static image."
See the difference? Now when you get the results, you're not just looking at a winning ad. You're learning something fundamental about your audience. If the UGC video wins, you've learned that authenticity and social proof are powerful drivers for your customers. That's an insight you can apply to your website, your emails, and your next ten ad campaigns. If you just find that "image_3_final.jpg" won, you've learned nothing of value.
Your tests should always be designed to answer a business question, not just to find a temporary winning ad. This is how you build a long-term, scalable advertising strategy instead of just lurching from one campaign to the next.
You'll need a simple structure for reliable testing...
To make sure your tests are actually telling you something useful, you need to structure your campaigns correctly to isolate variables. If you're testing creatives in one ad set, and audiences in another ad set, and headlines in a third... you'll end up with a mess of data that tells you nothing.
Here’s a simple, robust structure I'd recommend starting with. We use a variation of this for almost all our clients, from small e-commerce stores to large B2B SaaS companies.
Campaign Level:
Your campaign should have one, and only one, objective. If you want sales, choose the "Sales" objective. If you want leads, choose "Leads". Don't mix and match. I also strongly recommend turning on "Advantage Campaign Budget" (formerly CBO). This lets Facebook's algorithm distribute your budget across your ad sets, automatically sending more money to the best-performing audience. This is crucial for scaling.
Ad Set Level:
This is where you define your audience. Each ad set should target one specific, distinct audience. Don't lump interests, lookalikes, and retargeting audiences together. You could have:
- -> Ad Set 1: Retargeting - Website Visitors (30 Days)
- -> Ad Set 2: Lookalike Audience (1% of Purchasers)
- -> Ad Set 3: Interest Targeting (e.g., Competitor A + Competitor B)
By separating them, you can clearly see which *audience* is performing best. The campaign budget optimisation will do the work of allocating spend.
Ad Level:
This is where you run your creative test. Inside *each* ad set, you should place the same set of ads. So if you're testing Creative A, B, and C, then all three of those ads should be inside Ad Set 1, Ad Set 2, and Ad Set 3. This ensures you're testing the creatives fairly across all your different audiences.
Here's what that looks like visually:
Campaign
Objective: Sales | Advantage Campaign Budget: ON
Ad Set 1
Audience: Retargeting
Ad Set 2
Audience: Lookalikes
Ad Set 3
Audience: Interests
Ad A
Creative 1
Ad B
Creative 2
Ad C
Creative 3
Ad A
Creative 1
Ad B
Creative 2
Ad C
Creative 3
Ad A
Creative 1
Ad B
Creative 2
Ad C
Creative 3
This structure gives you clean data. You can look at the ad set level to see which audience is working. And you can look at the ad level (across all ad sets) to see which creative is the overall winner. It's simple, scalable, and it works.
You'll need to know if you've actually found a winner...
One final point. Just because an ad has a slightly lower cost-per-result after a day or two, doesn't mean it's a true winner. Results can fluctuate due to random chance, especially with small amounts of data. To be confident in your results, you need to consider statistical significance.
I wont go into the complex maths behind it, but the basic idea is this: is the difference in performance between your two ads large enough that it's unlikely to be random? A test might show Creative A has a 5% conversion rate and Creative B has a 5.5% conversion rate. Is that a real difference, or just noise? If you've only had 100 visitors to each, it's probably just noise. If you've had 10,000 visitors, that difference is much more likely to be real.
There are plenty of online calculators for this, but to make it easier, I've built a simple one for you below. You can plug in the numbers from two of your ads (e.g., number of clicks and number of conversions) to see how confident you can be that one is genuinely better than the other. A confidence level of 95% or higher is generally considered statistically significant.
Creative A (Control)
Creative B (Variation)
--
--% ConfidenceDon't stop a test too early. A general rule of thumb is to let each ad get at least 1,000 impressions and wait for at least a few days to a week to let the algorithm stabilise before you make any decisions. For conversion campaigns, you often need to wait until you have a decent number of conversions (e.g. 50 per variation) to make a reliable call.
This is the main advice I have for you:
To wrap this all up, here’s a table with my main recommendations. This is an actionable plan you can implement for your next campaign to test your three creatives effectively.
| Step | Action | Why You Should Do This |
|---|---|---|
| 1. Choose the Right Method | Create one ad set and place your three different creative ads inside it. Do NOT use the formal A/B Test tool for this. | This is the most efficient use of your budget. It allows Meta's algorithm (Advantage+ Creative) to automatically optimise and show the winning creative more often, improving your overall results. |
| 2. Define Your Hypothesis | For each of your three creatives, write down a simple hypothesis. E.g., "I believe this video will win because it shows the product in action, which should increase conversions." | This turns your test from a guess into a learning opportunity. You'll understand why an ad won, giving you insights for future campaigns. |
| 3. Use a Proper Campaign Structure | Set up one campaign with Advantage Campaign Budget (CBO) on. If you have multiple audiences, create a separate ad set for each one. Place the same three test ads in every ad set. | This isolates your variables. It lets you test creatives fairly across different audiences and allows the budget to flow to your best-performing audience automatically. |
| 4. Be Patient & Check Significance | Let the campaign run for at least 3-7 days and get a significant number of conversions before declaring a winner. Use the calculator above to check for statistical significance. | Early results can be misleading. Waiting ensures the algorithm has had time to learn and that your results aren't just down to random chance. |
As you can see, effective testing is much more than just clicking the 'A/B Test' button. It's a systematic process of forming a hypothesis, structuring your account correctly, and being disciplined in how you analyse the results. It takes a bit more effort up front, but the payoff in terms of improved performance and valuable customer insights is enormous.
Getting this right can be the difference between a campaign that breaks even and one that becomes a predictable engine for growth. If you find this all a bit overwhelming or simply don't have the time to manage this process yourself, it might be worth considering some expert help. We spend all day, every day, inside ad accounts, running these kinds of tests and scaling campaigns for our clients.
We offer a completely free, no-obligation initial consultation where we can take a look at your account and strategy together. We can often spot opportunities for improvement in just 20 minutes that could save you a significant amount on your ad spend. Feel free to book a call if you'd like to have a chat.
Hope this helps!
Regards,
Team @ Lukas Holschuh