Split Tests

Split testing is a method for doing controlled experiments with for instance content or layout, to see which versions performs the best. In other words, it is a way to test ideas and optimize value generation, without taking unnecessary risks.

With split test, you can:

  • Test different versions of your content and design
  • Improve conversion rates by testing different calls to action
  • Test on 10 different conversion metrics
  • Compare conversions of original and variation
  • Pick the winner manually – or based on statistical significance.

In short, split tests are a useful way for you to make more money, with a minimum of risk.

Here's a short video introduction to Split Tests:

You can conduct split tests on newsletters and marketing emails as well - read more about how to do that under the documentation for Email Marketing.

manual

With split tests, you create variation on a content or layout level, direct a portion of your visitors through the variation, and measure the conversion achieved for each version over time.

Before setting up a split test, you must:

  • Identify an area of your website where a split test could be useful
  • Define a specific conversion goal (such as ‘more visitors must place an order’ or ‘more customers should sign up for the newsletter’)
  • Create a variant of your original content containing a change in content or layout.

In this section, you will learn about:

  • Setting up a split test
  • Starting and stopping split tests
  • Monitoring split tests

In short – everything you need to know to get started with split tests.

Split tests are set up on a page level:

  • Navigate to the page where you want to do a split test.
  • Click the Marketing tab (Figure 3.1):
Figure 3.1 The Marketing tab
  • Click Setup split test

You will then be presented with the split test wizard (Figure 3.2)

Figure 3.2 The split test wizard

Following the wizard will take you through the basic setup of a split test .

Here's a rundown of your options, and what they mean.

You can choose between content-based and a page-based split test.

A content based split test:

  • Tests paragraphs against each other
  • Is useful for testing variation on design elements, such as buttons
  • Is useful for testing variation on text and images

A page based split test:

  • Tests pages against each other
  • Is useful for testing different layouts

In summary, choose content based split testing if you want to test content like buttons, images or texts, and page based split testing if you want to test a new layout or design.

To be effective, a split test needs a clearly defined goal.

The conversion goals available to you are:

Name of goal

Tests which version..

Choose another page as conversion page

..will make more visitors go to a page you specify – for instance the cart

Create an item

..will make more people create a specified item

Submitting a form from the forms module

..will make more visitors submit a form

Adding products to cart

..will make more visitors add products to the cart

Placing an order

..will make more visitors place an order

Downloading a file

..will make more visitors to download a specific file

Signing up for newsletter

..will make more visitors sign up for your newsletter

Maximizing time spent on page

..will make visitors stay in your website longer

Highest average order value (only content based)

..will increase the average value of placed order

Highest average order markup (only content based)

..will increase the average markup of placed orders. (Markup is the difference between cost and selling price of a product)

Minimize bounce rate

..will lower the amount of people who only view one page before leaving your website

Maximize page view

..will make visitors view more pages

You should design your split test variant around a single goal, to minimize variables and get better result. For the same reason, you can only choose one type of conversion goal for each split test.

In order to use the 'Highest average order markup'-goal, you must set the Global setting ShowProductCostField to True.

To do so:

  • Go to Management Center > System > Global Settings
  • Navigate to Ecom > Product > ShowProductCostField
  • Change setting to True
  • Save

Don't forget to set the Price and Cost field of your products to the correct values.

After choosing your conversion goal, you will be presented with the settings window (Figure 6.1).

Figure 6.1 Specifying the settings for you split test

From here you must:

  • Name the split test
  • Decide if the split test is active immediately or not
  • Choose between registering conversion metrics only on the next step, or through the entire visit
  • Specify the amount of traffic directed through the experiment
    • Note: If you choose 10%, 5% of visitors will be shown the variant and 5% the original, and conversion metrics measured based on those 10%. The remaining 90% of the traffic will not be part of the test.
  • Decide whether you want both the variant and the original to be available to search engine bots.

After having made your selection, you will be asked to select a setting for ending the split test.

Your options are:

  • Manually
  • At given time
  • After x views
  • When result is significant

If you select one of the last three automatic options, you can define an action to take when the split test has run its course (Figure 6.2)

Figure 6.2 Selecting an automatic action after a split test ends

With these settings, you can:

  • Decide if you want to delete visitor and conversion data or not
  • Keep all versions or only keep the best performing version
  • Notify people of the split test results (and specify an email notification template or create your own)

You can also end a split test manually.

Starting and stopping split tests can be done in a number of ways, depending on the type of split test, and what your settings are.

Here's a rundown of your options.

Page based split tests start automatically when you finish the split test wizard, if you checked the This split test is active checkbox under the split test settings.

If you did not, simply

  • Navigate to the page you are testing
  • Click the Marketing tab
  • Click start (Figure 8.1)
Figure 8.1 Starting a page based split test

Before starting a content based split test, you must select one or more paragraphs on a page and create variants.

This will create duplicates, called variations, of the selected paragraphs (Figure 9.1):

Figure 9.1 Select yourparagraph, click create variant, make changes to your variant paragraph

Next, open the variant paragraph and make the change you want to test.

Half the visitors sent through the split test will see the variation paragraph, the other half the original.

Finally, click the start button to activate the test on your site and begin collecting conversion data.

You can preview the frontend differences between the variant and the original by pressing the Preview button (Figure 10.1)

Figure 10.1 Click Preview to preview your content variation

You can stop a split test automatically when certain conditions are fulfilled.

You can stop split tests manually at any time by clicking the Stop under the Marketing tab (Figure 11.2).

Figure 11.2 Stop a split test manually by clicking the stop button

When stopping a split test manually, you will see the following window (Figure 11.3).

Figure 11.3 The stop split test window

Here you must:

  • Confirm your choice by checking the Yes, stop this experiment box
  • Select what to do with test variations:
    • Keep all versions, with the original published
    • Keep the original and delete the variation
    • Keep the variation and delete the original
    • Keep the best performing version and delete the other

Stopping a split test will delete the collected data on visitors and conversions. If you don’t want that to happen, Pause the split test instead (Figure 12.2).

Figure 12.2 Click pause to stop a split test, without deleting all collected data

If you have many split tests running, you may find the split test overview useful.

To access the split test overview:

  • Click Marketing on your accordion
  • Click on Split tests

This takes you to the split test overview (Figure 13.1).

Figure 13.1 The split test overview

From the split test overview, you can:

  • See the name, page, type, goal, and state of each split test
  • Access the report of each split test
  • By right clicking a split test you can:
    • Pause a split test
    • Stop a split test
    • View the split test report
  • Clicking the page icon to the left of the split test name to go to the page where the split test is running

The split test report is a visual representation of the data collected by your split test.

The split test report is a valuable tool for optimizing value generation, because it:

  • Gives you access to the data collected
  • Helps you determine if the data is good enough to base a decision on.

 

The report compares the conversion rate of the variant and original content (Figure 14.1).

Access the split test report from either the page you are split testing, or or from the split test overview.

Figure 14.1 The split test report

Let’s break the report down:

  • At the top, you can see the type of test and the conversion goal.
  • You then have a graph of the data collected during the split test. The data you can see will depend on the conversion goal. In the screenshot above, the goal is to maximize the time spent on your site, and consequently you can see the maximum time on site per view per variant.
  • The table at the bottom provides an overview of the data gathered so far - depending on the conversion goal.
    • The number of views per variant will always be shown
    • The Conversion Rate (CR) improvement calculation measures how a variant performs compared to the original, and will always be shown.

Other data columns may or may not be present, depending on the conversion goal of the split test:

  • Conversion goal specific parameters (e.g. time spent on site per view) - or average/total order value.
  • For conversion goals which are of the do/donot type, you will see a probability calculation, which indicates whether the calculated improvement is statistically significant or not. Values close to 100% indicate a high level of statistical significance, whereas values close to 0% indicates that the improvement can be ascribed to randomness. (Gathering more data will result in higher statistical significance.)

With these data at your disposal, it is easy to determine the best variant for your split tests.

implementing

In this section, you will learn about:

  • Modifying your template to create variants for a split test directly in the template

You can modify your template to create variants dynamically, depending on the session.

Use the template tag Global:OMC.Experiment.Variation.SOMEVARIATION to check if it’s an original or variant session, and change what happens on the frontend.

SOMEVARIATION will be replaced with ‘Original’ for the original and ‘2’ for the variant.

For example:

<!--@If Defined(Global:OMC.Experiment.Variation.SOMEVARIATION)--> <!--@Else--> <!--@EndIf-->

You still need to setup and start a split test on the page or paragraph – the tag merely enables you to create your variations directly in the template.

development

There is no developer-specific information for this subject

database

There is no database specific information for this subject

integration

There is no integrator-specific information for this subject