Typically, if you run 4 CRO tests each month
(that’s a test/week), and if 10{c87e2df4b343d0515d304e127afe4653a549475791ab451641a18e09bd64e760} of your tests win, you’re running a good
optimization program. That’s a decent testing capacity and a nice win rate.
Even better, if you manage a good uplift for
your winning tests and your program’s performance keeps improving over time.
But most optimization programs don’t run so
well.
In fact, only 22{c87e2df4b343d0515d304e127afe4653a549475791ab451641a18e09bd64e760} of companies are happy with
their CRO efforts.
Which means, a whopping 78{c87e2df4b343d0515d304e127afe4653a549475791ab451641a18e09bd64e760} of companies could
optimize their optimization programs.
But how…
What Kills Most Conversion Optimization Programs
The problem with most optimization programs is
that they aren’t designed for long-term success. Instead, they thrive on a
test-by-test basis.
Such programs are mostly only as effective (or
not) as the last test they ran.
And the companies running them look at the
implementation of the experiments as the win. Their myopic view stops them from
ever developing the infrastructure to support a consistent, quality testing
program.
While it’s true that good execution is a must
for any experiment, but even a bad experiment can be executed really well.
But nobody wins when this happens.
Focusing solely on execution and not spending enough time and effort on steps such as ideating, hypothesizing, and documenting/learning — that actually determine the quality of the experiments — usually results in only short-term success, if at all.
So let’s see how you can increase testing velocity and run a good optimization program. If you’re already running one, you can use these tips to further improve your win rate and overall program performance.
Here goes.
How To Generate More Ideas to Test
To run (at least) 4 tests each month, you need
a pipeline full of testing ideas. Without an “idea bank,” you can’t
support a good and consistent testing velocity.
Despite that, in most CRO programs, tests are
planned when someone on the team has a CRO test epiphany of some sort.
Ideally, you should have a constant influx of quality testing ideas into your experimentation program. These testing ideas can come from:
- Digging into the mountains of data your CRO tools generate. The best way of spotting testing ideas is to dig into your data. Your analytics solutions such as Google Analytics, Kissmetrics, Mixpanel, etc. are excellent sources for finding the pages where you lose most people or those that have low engagement rates. Tools like Hotjar, Clicktale, and Decibel show you what your users do on your website and can help identify your real conversion hotspots. Then there are solutions like UserTesting, UsabilityHub, and Usabilla among others that let you collect heaps of qualitative feedback that can translate to some crucial opportunities for testing. While it’s challenging to review so many data silos, these are the places where the real winning testing ideas come from.
- Running a manual CRO audit. Auditing your
website for CRO uncovers some of the most valuable optimization gaps for
testing. Running a CRO audit forces you to systematically look at each and
every aspect of your website (and beyond) and see where you could be losing
money.
- Using assessments like Stuck Score™ to spot the “conversion barriers” on your website. You can also use assessments like Stuck Score™ that uncover the conversion issues on your website and offer excellent ideas to test. These tools are intelligent and can precisely spot testing opportunities across your entire website.
Once you start tapping into these sources, you
should be able to generate a consistent stream of ideas.
But generating quality testing ideas is just
one aspect of this issue. The other is the lack of communication and
collaboration on the ideas under consideration. This might seem trivial
(because, after all, you only need data, right?), but these issues deeply
affect your people and can skew your culture of experimentation.
For instance, take the famous 41 shades of
blue experiment Google tried. Google’s experiment — data-informed as it was — was still
criticized for taking an engineer-led approach. Here’s what Douglas Bowman, who
worked as Google’s in-house designer, felt on how Google handled its
experimentation: “Yes, it’s true
that a team at Google couldn’t decide between two blues, so they’re testing 41
shades between each blue to see which one performs better. I had a recent
debate over whether a border should be 3, 4 or 5 pixels wide, and was asked to
prove my case. I can’t operate in an environment like that. I’ve grown tired of
debating such minuscule design decisions.”
Without sharing the ideas you’re considering
and engaging your team, you can’t build an all-inclusive culture of
experimentation that everyone wants to be part of.
With a CRO tool like Compass (from our Convert Suite), you can effortlessly facilitate such data-backed, collaborative ideation. Compass lets you come up with data-backed testing ideas by bringing together your different data sources and also suggests ideas to test based on insights from Stuck Score™. Compass even lets you invite your team members and engage them with options for feedback and more.
Forming Data-Backed Hypotheses and Laser-Focused Prioritization
Once you have your testing ideas, you’ll find that a few of them are just plain obvious. For example, if you get some user feedback that your content isn’t readable (and your target demographic are, say, people over forty years of age), then maybe you can implement the idea of increasing your font size or changing its color right away. After all, it’s a one-minute fix with a small CSS code change.
Some of your ideas would look promising and
seem totally worthy of testing, but you’ll still need to look for “enough”
data points to support them [more on this in a minute…].
And some ideas you’d simply have to discard because they’d be vague and you’d have no way to validate them. For example, if your CRO audit shows that you have a low NPS score, and you find it to be the reason for poor conversions, then you can’t possibly use a simple experiment to fix it.
Of these, the ideas that can actually
translate to strong hypotheses are your real testing opportunities.
But you need a LOT OF data to support each hypothesis you make. So for example, if you hypothesize that optimizing your mobile landing page experience will result in higher conversions, you’d need a bunch of data points to support it. In this case, here’s some of the data you could use:
- Low mobile conversions — data via your web analytics solution like Google Analytics.
- An unusually high dropoff for mobile traffic — again, data via your web analytics solution like Google Analytics.
- Poor feedback from customers — data via your user testing solution.
As you can see, the data to form this
hypothesis is quite balanced as you’ve inputs from multiple data sources. Also,
you’ve both quantitative and qualitative data. Ideally, you should find such
balanced data to support all your “test-worthy” ideas.
But you’re not done yet.
Because after you’re ready with all your good
hypotheses, you need a way of scoring or prioritizing them. Doing so tells you
which hypothesis to try first — or at all. HINT: “Let’s test a new website design!!! It’s going to shoot up our
sales.” is usually a VERY BAD hypothesis.
Many factors go into deciding how practical
testing a hypothesis is. Its implementation time and difficulty and the
potential impact it can have on the conversions need to be considered here.
But most companies lack a prioritization model
for this. This often results in the launching of an ambitious test like, say, a
major design overhaul that uses up the entire month’s worth of CRO bandwidth.
Which means you can’t plan or run anymore tests, at least for the month. The
worst part is that even such ambitious tests don’t guarantee any significant
results.
To avoid this, you could use the PXL prioritization framework from CXL. This framework forces you to think at a very granular level such as understanding the proposed change(s) better, evaluating how it addresses the issues discovered during the research, its potential impact, and the effort of implementation.
You could also check out the PIE and ICE Score
framework for prioritizing your hypotheses.
An even smarter way of prioritizing your hypotheses is to use a CRO tool that can tell you how resource- and time-intensive an experiment can be. For example, Compass gives you good estimates for all your hypotheses.
Learn From Your A/B Tests
Tests can be inconclusive.
For most conversion rate optimization programs, you get a low 20{c87e2df4b343d0515d304e127afe4653a549475791ab451641a18e09bd64e760} of tests reaching statistical significance.
And so all learning needs to go back into the testing mix and be used to come up with better and more refined ideas and hypotheses.
Not just that, winning experiments can also be
losers actually, when the challenger version wins but the revenue tanks.
Besides, if a hypothesis is a really strong
and data-backed one, it’s common to create about 3-4 follow up experiments for
it (even if the initial experiment won!).
Which means, just interpreting and recording your experiment’s results isn’t enough. To plan meaningful iterative testing, you need to document your entire experimentation process each time you run one.
By documenting its observations and learnings,
LinkedIn was able to follow up on a failed experiment, which was actually a
winner on the key feature being tested. Here’s the full scoop:
In 2013, LinkedIn Search kicked off a major experiment where it released its upgraded unified
search functionality. Basically, LinkedIn Search got
“smart enough” to figure out the query intent automatically without
needed qualifiers like “People,” or “Jobs,” or
“Companies.” The search
landing page was totally revamped for this release — everything right from the
navigation bar to the buttons and snippets were redone, so the users saw many,
many changes.
But the experiment failed and LinkedIn was
surprised to see its key metrics tank.
The team now decided to roll back to the
original design by backtracking one change at a time, so that it could identify
the one that didn’t go down well the users. During this time-consuming
rollback, LinkedIn found that it wasn’t the unified search that people didn’t
like, but it was a group of several small changes that had brought down the
clicks and revenue. Once LinkedIn fixed these, unified search showed to have a
positive user experience and was released for everyone.
So whether it’s a simple A/B test or a complex multivariate one, any experiment you launch should be documented in detail. Its learnings need to be documented as well. By doing so, you can ensure that your future (or follow up) experiments are actually better than your earlier ones.
With a CRO tool like Convert Compass, you can build a knowledgebase of your ideas, observations, hypotheses, and learnings so your entire team can learn and grow together. Not just that, Compass can even use your learnings to suggest the hypothesis you could try next.
Wrapping It Up…
By optimizing the ideating, hypothesizing, and
learning parts of your CRO program, you can dramatically improve the quality of
your experiments. And by collaborating and engaging with all your people over
these, you can build and boost an all-inclusive culture of experimentation.
You might find it difficult to bring together all your data when ideating or suffer from data overwhelm when hypothesizing (and prioritization), or even struggle with documenting or using learnings for your follow up experiments, but these are the things that will help you increase testing velocity and set the groundwork for your long-term CRO success.
If you’d rather use a CRO tool that does all this heavy-lifting for you, sign up below ⤵.
Compass helps you with data-backed ideation (by bringing together all the data from your different data silos and with inputs from Stuck Score™ suggesting ideas to try first), meaningful prioritization (by telling you how difficult, easy, or impactful an experiment could be), and the documentation of your learnings (by bringing together all your ideas, data research, observations, results, learnings and more in one place!).
Originally published May 21, 2019 – Updated July 17, 2024
Mobile reading?
Authors