/ A/B TestingHow A Single A/B Test Increased Conversions by 336% [Case Study] By Dustin Sparks on November 11th, 2013 in A/B TestingTotal618 296 110 212336% ? No Way! (Image Source) In the post below I am going to share a step by step case study in which we increased conversions 336% (that’s not a typo or stuck key – it’s a legit 336%) for a client who owns and operates a career college. For this client leads are their life-blood.By walking you through each step, from the hypotheses, to the strategy, to the goals, the setup, the test, the results and the next steps, I hope to give you a very clean very easy to follow strategy for you to start applying the same sort of approach to your site/pages or your clients sites and pages – immediately after reading this.At the end I have a very ‘neat analytics nugget‘ that showed up as an ‘Accidental Results‘ discovery in this case study.Let’s get startedAs mentioned above, this post walks you through a case study in which we increased conversions by 336%…(can you tell I love that percentage?). The client runs a college, whose business is graduating very talented students. But in order to graduate talented student, they need to start by enrolling students – and that’s where I came in.The TestI start all initial testing with an ‘Apples to Oranges‘ split test.An “Apples to Oranges” split test, at its simplest: if you have PPC budget – you of course want to make the most efficient use as possible of each click per dollar. If you test apples to oranges and find you visitors prefer apples, then you can stop wasting money on keywords that get you clicks for oranges, and start spending your money on tests that help you identify if the apple lovers like red apples or green apples. Then if your tests reveal that your visitors like red apples, you can test further and spend your PPC budget more efficiently finding out if your visitors like a Red Honeycrisp apple, or a Red Jonathan…and so on and so forth.The A/B Test (or Apples to Oranges) I ran for this client can be seen below:Click for full-size imageThe original (Version A) was converting at 3.12%. That conversion statistic in itself, is not great, but also not too bad considering it was less that 2% when we took it over and immediately applied as many best practices as we could (as bandaids) in order to “stop the bleeding” from the original site.The HypothesisThe hypothesis for this test was simple:Removing primary navigation and modifying the form layout placement above the fold will increase (improve) form submissionsThe Test SetupMobile traffic was high – mobile traffic has been increasing in general, but for this target group ‘potential college students’ it was really high – 59.84% during the last 90 days, including this test timeframe, was Mobile.Both A & B Versions were mobile responsiveWe noticed with this client that mobile traffic seems to support the middle of the funnel – but the conversions happen (99.9% of the time) from a web browserThe B Version removed Primary navigation, had a dramatically different theme with the same colors, and had a single image with a larger “human”The A Version had a Quality score of 8, while the B Versions had a Quality Score of 10All traffic for this test was from Google Adwords – Pay-Per-Click marketing.By making URL’s for Adwords only available from Google Experiments via Adwords you can ensure no organic traffic comes through and your referral traffic will also be almost non-existent.For Data and Page analytics, we used Google AnalyticsIt’s important to note that on all tests, we set filters to filter out all Internal and development traffic in order to get as accurate traffic and test results as possibleFor rotating the A & B Versions of the pages we used Google Experiments**Please note, while Google Experiments is a very decent tool, at times they ‘take too much control’ of your testing, often stopping a test and declaring a winner automatically. This is why you will see a lower number of visitors than what we recommend (100 minimum) on the test below.The test ran for 23 daysIn 23 days it had exactly 120 experiments – Pure PPC visitsThis is nice because some tests can take a LOT more visits and experiments to determine a clear winner.The ResultsThe winner by a landslide was the Challenger – the B Version.Click for full-size imageThe probability or confidence percentage that the champion variant would outperform the original was 95.2% and conversions increased from 3.12% to 13.64%.*Editors Note: Unbounce recommends sending at least 100 visitors to a landing page; However, due to the tool used this was not possible (Click for full-size image)So with this hypothesis, test and results, we clearly had a winner.Additional Learning & ResultsBut wait, there’s more!While having a winner, and validating a hypothesis is rewarding – the real magic happens sometimes when you dive deeper into the metrics and analytics “around the perimeter”.In this case the “bonus” findings were:Bounce Rate Improved from 32.50% to 28.70% – pretty cool, but not earth shattering88% of the clicks were below the fold. Now this is worth a look, right? Especially since we didn’t have any navigation on this page (other than social links opening in a new tab/window)Click for full-size imageClick for full-size imageVisitors were scrolling ALL the way to the bottom and clicking through the testimonials! This is very likely the Middle of the Funnel people looking for that coveted “social proof – “What are other people saying?” “What are other people doing?” “What are other people buying?” (pst – don’t know what the Middle of the Funnel (MOF) is? Read this post.Now this is great data to work with because if we have 88% of our visitors clicking around at the bottom, the next test would place a CTA at the bottom of the page near the scrolling button*Please note: While this test was a success and we learned a ton, the information in this case study is very fresh – only a few weeks old. We are currently in the process of rolling out these changes.Your TurnThe reason I wanted to share this case study so soon after getting the results was that it’s a great example of how to:Create a very simple hypothesis.Remember every hypothesis should be geared to “IMPROVE” a situation. SO when you write your hypothesis – literary spell it out…the test is to try to improve, click thru, conversions forms filled out, phone numbers called etc.Start A/B split testing to compare to very different versions “apples to oranges”.If you still aren’t sure about this – contact me in the comments below.Use a tool to serve and rotate the Control and Challenger pages.Review the results against the hypothesisThis will be an easy review – did your B Version test/challenger page improve, weaken or not change your baseline of results before you started testing?Dive deeper into the Analytics to see “what else can I learn about this test”Get a cup of coffee and literally wade through allreports and metrics for the test and the timeframe the test ran.Make notes and observations – these will help build your NEXT round of testingUse the Champion page as the new Control page and create a new Challenger page.Now get on with juicing more conversions out of your pages or the pages for your clientsI want to quickly thank two people I couldn’t have done this case study without. A big shout out to my associate at COX Media, Alfonso Montemayor and developer, Kyle Sanders.Keep testing and never hesitate to reach out via email or the comments below – I reply to everything.— Dustin SparksAbout Dustin Sparks Dustin Sparks, Landing Page, A/B Conversion Optimization Expert. He provides his clients with optimized landing page strategies for improving conversion rates. His comprehensive landing page optimization process includes the usage of advanced A/B testing to determine which landing page tactics will convert best for your specific campaign goals.» More blog posts by Dustin Sparks