Why I Die a Little on the Inside Every Time You Ask Me What Button Color Converts Best
Also, intelligent people are more likely to binge drink (source). But, don’t expect it to get you any brownie points when police find you at 1:00pm belligerently yelling “I’M SMARTER THAN YOU!” at a fire hydrant because it won’t give you extra ketchup for your French fries…
The point is: it is POSSIBLE for two seemingly random things to be correlated and if you don't understand why, all your A/B testing is doomed to fail.
These bizarre correlations are what I think about every time I get the question: “what button color gets the highest conversions?”
When I hear this question, I know the person asking it doesn’t fully understand A/B testing and will likely spend a lot of time testing the wrong things.
Sure, it is POSSIBLE that changing a button from green to blue increases conversions. But I’m not going to be convinced until I can see the raw data. And even then, I still wouldn’t change all the buttons on my website to blue.
In this post I’m going to show you why you shouldn’t be worrying about things like button color and what you should be focusing on instead. This will allow you to quickly take advantage of A/B testing for your landing pages and A/B testing for your opt-in forms to grow your audience and business.
Data Is Your Master (Obey Your Master)
If I told you to find out if people like french fries or onion rings better, you might start by walking down the street and asking the next 5 people you see. Three of them say “onion rings.”
That’s a majority, but have you really learned anything?
Would you be confident to say that, in general, people prefer onion rings over french fries? I hope not.
Every experiment runs into the issue of imperfect measurement. This is because there's randomness in tracking, randomness in people's behavior and randomness in uncontrolled (and perhaps unknown) external factors.
Maybe one of the people you asked just finished eating onion rings. That sweet and salty goodness is still fresh on their mind, so they answer "onion rings", even though, on average, their eating habits would show they definitely prefer (the obviously far inferior) french fries.
Or maybe the Worldwide Onion Ring Convention is being held just around the corner. As a result, there are many more die-hard onion ring fans walking around than on most days.
(Or perhaps the person you asked just finished reading this blog post, which has an obvious agenda to promote onion rings—the food of gods—and this subconsciously affects their decision.)
The solution to this problem?
Ask more people (gather more data). The more people you ask, the more confident you can be about where people in general stand on the issue of french fries vs. onion rings.
But how many more people do you need to ask to overcome the imperfect measurement issue? Well, this depends on how closely related the two things you’re testing are.
This brings us to the next problem:
How Much Data Do You Need?
Let's do another thought experiment. This time, imagine asking people to judge the weight of two different objects.
First up: a paper clip and a bowling ball. The hypothesis we propose is: the bowling ball is heavier.
The results from asking 10 people to guess the weights might look like this:
We didn’t ask many people, and everyone answered slightly differently on what they thought each item weighed. But, because they are so drastically different, we can tell from these results that people generally find a bowling ball to be heavier than a paper clip.
Next, we repeat this same test, but with two new items:
The next are a quart of milk and liter of soda. The hypothesis we propose is: the liter of soda is heavier.
In this experiment, it’s a lot more difficult to come to a conclusion with the results, since the answers are overlapping.
We could ask more people, to add more data to this test. Maybe we can come to a conclusion about whether people in general guess that a liter of soda is heavier than a quart of milk, but even if we ask 100 people or more, we won't get a highly confident answer.
Milk, Soda & Button Colors
Testing different button colors is an example of testing a small change, where the difference in conversions you can measure is probably going to be very small. It's like our soda vs. milk example above.
This means you’re going to need a lot more data and time to get significant results.
When you’re just starting out, it’s a much better use of time to perform bowling ball vs. paper clip type tests. They will give you better insights into your customers and are a much more effective use of your time to grow your business in the early stages.
We'll look at some examples of effective tests you can use immediately, but first, I want to address another issue I see with people new to A/B testing: following case studies...
Why A/B Testing Case Studies Are (Almost) Useless
You read about a case study where a change to a landing page made a HUGE difference in conversions, so you go ahead and apply the exact same change to your own site, right away.
Makes sense, right?
This approach doesn't work because A/B testing is context dependent.
This means results are extremely specific to the audience and business.
For example, this case study found that 88% of the clicks on the website they were testing happened below the fold.
So let's say you read this result and promptly move all the call to action buttons on your site below the fold. What happens?
If your product is, say, a 99 cent iPhone app, your sales might plummet dramatically.
The business in the case study is a university, where students are trying to decide if they should enroll.
One of the major differences here is the price point.
If you’re looking to spend thousands of dollars (in the case of getting a college education), most people are going to need A LOT of information before they decide to buy. The visitors in the case study were scrolling below the fold to read testimonials and gather more info, so it makes sense to offer a call-to-action button at the bottom.
The same almost certainly won't apply for a low-priced product like a $0.99 iPhone app.
I get it, when you’re new, it’s easy to get caught up in the hype of A/B testing when you see results like:
But remember, the results are context dependent, so use case studies for ideas on WHAT you can test, but don’t copy them expecting to get the same results.
Conversion Optimization Without the Time Wasting Mistakes
People often think A/B testing is too complicated and intimidating and as you've just seen, there are some nuances to consider. But let's look at some methods you can apply today to make A/B testing quick and easy.
Step 1: Do You Have Enough Data?
Use this calculator to get a general idea of how long you need to test based on how many conversions you get per day.
As a general guideline, you should test for at least 14 days. Use the calculator (or our automatic winner settings), and if you’re not anywhere close to being able to get significant results in 2 weeks, go to step 2. Otherwise, skip to step 4.
Step 2: Get More Traffic
If you have a low traffic website there are 3 ways you can quickly optimize it:
- Only fix what’s broken.
- Run usability tests.
- Create 1 funnel and drive traffic to it.
Here's a quick guide explaining this approach in more detail: How To Optimize a Website with Low Traffic
Step 3: Deliver Something Good Enough. Quickly.
Initially, the goal shouldn't be to create the perfect landing page or website.
The goal should be to build something effective and get it published as soon as possible. Even if you don't have any traffic to run tests on, you can follow some basic guidelines and best practices to ensure what you publish is good enough to start with.
Here is a great FREE course that will teach you how to build highly effective landing pages to increase your conversion rate for any market. The course will show you how to use a principle called rapid implementation - one of the most important principles you can learn as an online entrepreneur.
Your chances of ever creating something great can be predicted by your ability to DELIVER something simple.
Step 4: Test Paper Clips vs. Bowling Balls
Focus on testing high-impact, big changes rather than small tweaks like button colors.
Big changes are more likely to lead to a big difference in conversion rates and thus lead to significant results (and significant insights into your customers) faster.
Here are ideas of high-impact items you can test:
- Does [Free Report A] or [Free Report B] convert better?
- Does an extremely long sales page convert better than a short sales message?
- What value proposition or overall sales message performs better?
Set up some A/B tests of some high-impact items. Even if you don't have much traffic and don't get an answer quickly, if you test big changes, you'll get a clear answer eventually.
Again, this free course is great for getting A/B test results by learning how to deploy landing pages in record time.
Step 5: Keep Testing But Realize...
It’s important to have an idea what A/B testing is actually like as you continue.
This could be an entire course by itself, so I’ll keep it to a broad overview, with a few examples to give you a taste.
A/B Testing is a Marathon, Not a Sprint
It can be an extremely effective tactic to grow your business, but it's a long-term tactic, not an overnight fix.
For example, Fiverr saw a 457% increase in landing page registrations, but this took 10 months, and over 400 tests, with at least 500,000 unique visitors to each one of those tests.
Huge Wins Aren’t The Norm
When you start to get enough traffic and enough data, you can switch to more focused testing. Meaning, start taking what you have learned works and start testing small adjustments, like headlines, to see if you can improve your conversion rates.
You can even take the testing to the extreme and test small details.
For example, the lingerie company Adore Me who is obsessively diligent in their A/B testing, even testing how different hand positions their models use affect sales.
It’s important to realize when you get to this point, you’re not going to be seeing HUGE wins and drastic improvements.
So, realize the small increments can add up over time, but sometimes you might spend a lot of time and resources, not increasing your conversions at all.
Try to avoid chasing after small details and tweaks. No matter how much traffic you get, testing big changes is always more valuable. And even the mighty Google, famously data driven and analytical as they are, doesn't have enough traffic for some tests.
Your Testing Advantage
You're now armed with more insight about how to perform A/B tests and optimize your site than most business owners will ever be.
Keep in mind that in the end, A/B testing doesn't have to be hard if you're using the right tools (Thrive Leads for your opt-in forms & Thrive Optimize for your landing pages) and it isn't so much about technical details as it is a tool to help you understand your visitors and users.
One way of looking at it is by asking yourself: will this test give me new information about who my customers are? If the answer is "no", there's probably a better test you can run. Always ask yourself: will this A/B test give me new information about my customers? No? Then there's a better test to run.
If you have any questions or feedback about this post or want to chime in on the onion rings vs. french fry war, please let me know by leaving a comment!
LONG LIVE TEAM ONION RINGS!