Blog
Theoretically, A/B testing is fantastic: B2B marketers compare two versions of an ad, landing page, email or other copy, changing one variable. Then they look to see which version may better resonate with customers.But in practice, a lot of outside variables in today’s digital landscape make it hard to do a clean A/B test.Marketing channels can input all sorts of factors into A/B tests that likely have little to do with what you want to measure and learn about. Rather than let that slow you down, embrace it. The machine learning (ML) and artificial intelligence (AI) available in many of today’s channels and tools can help your campaigns perform far beyond completely manual management. The channels in which B2B marketers place ads are highly dynamic, i.e., too dynamic to maintain strict environments for A/B testing. Rather than trying to make an A/B test perfect, lean into the dynamic nature of the channel and test broader messaging or creative ideas. Here’s how to do it.
The digital advertising space offers increased levels and options for ML and AI automations every day. This includes the dynamic placement of ad assets (formerly extensions), headline rotation, dynamic keyword insertion and so much more. The problem with leaning into ML and automated tools within platforms is they strip away our narrow ad testing controls. Let’s take search ads as an example. Say we want to test “Ad A” vs. “Ad B” to determine which version may generate a higher click-through rate (CTR). Consider the following scenarios:
Each of these scenarios raises new, more difficult questions about how to evaluate the A/B test results. Comparing the CTR of Ad A to Ad B without considering the impact of each of those different scenarios could leave you missing extremely relevant variables. And collecting enough data for a statistically significant pool of data to be able to fairly evaluate Ads A and B while considering all of these extra factors will take extensive time and deplete budgetary allocations. Even though ad assets — like sitelinks, callouts and images — compound the variability and meaning of A/B test results, I would never encourage marketers to shy away from them. I’d much rather they appear with our ads than not. Not only do assets allow you to present your brand in a more elevated way, but studies show assets greatly increase CTR and conversion rates.Consider another scenario: Ad A may always appear in position 1 while Ad B may always appear in position 2. In this scenario, our extra variable is not related to one of Google’s automated features or ad assets. So even when we don’t employ these features, variables based on the highly dynamic nature of Google search engine results pages (SERPs) are always going to be present. All this complexity makes strict A/B testing very difficult. But B2B marketers should embrace this. Let’s face it: the search algorithms engineered within Google and Microsoft for rapidly collecting, computing and analyzing data are faster than any of us could ever be. Google knows far more about our audience in that auction moment than we could predict. And holding back from these dynamic elements in order to protect your testing may very well limit your search ad potential.
Many of the same challenges exist with social media.Conventional wisdom suggests that B2B marketers A/B test social posts by, for example, changing the color of images to determine which may best resonate with audiences. But today, you can’t control where a user sees your ad. It might be displayed next to two unrelated posts with content or design that doesn’t interest the user, prompting them to scroll past.And A/B testing doesn’t always explain an audience’s reaction to your social posts. If one color appears more often for managers while a second color appears more often for directors, is it color 1 or is it managers? (Answer: It’s not clear.)Placement can also vary by platform. A video ad on Meta could appear in stories, reels, a newsfeed or the video section on Facebook. If one ad variation serves in one of these placements more often than other variations, you could have skewed A/B test results. This lack of certainty can be frustrating. However, even if you could more tightly control where ads appear on social platforms, new challenges would crop up. Once again, you’d find yourself needing a large budget and/or a lot of time to collect a statistically significant number of clicks.
None of this is meant to suggest that B2B marketers should stop A/B testing. It remains a valuable methodology for learning about business audiences by comparing what messages, designs and/or formats they prefer. Let’s just be more efficient and refresh our approach to digital platforms by following these three steps:1. Determine what you want to learn before testing.
There’s no right answer here. So many elements of an ad or message can be tested. But it may help to consider what you plan to do with your learnings, so you can take actionable steps once your test is complete.2. Build campaigns and ads to answer these bigger questions.
3. Don’t forget the outside variables that could impact your test results.
Completing these steps can provide the insights that may help you make better decisions based on real data about and from your target audience. Even with all the new complexities, A/B testing is far better than guessing.
Remember, the goal of testing is to ultimately find the ways in which your ads can achieve the most success. It’s OK to lean into what works, even if it’s not the most scientific approach.
Five Tips for Better B2B Brand Storytelling
World’s Best Pizza? If You’re Claiming to Be #1, You Need to Back It Up
CMOs Need to “Budget Bold” for 2024
2023 Budgeting: Invest in Digital, Content to Win B2B Business