Campaign performance - product badging

In this step, you’ll learn about what happens when you launch a campaign, what metrics you’ll see when the Coveo Experience Hub has enough data, and how you can use those metrics to gauge the impact of campaigns.

Some useful information

When you launch a campaign, your customers will start to generate relevant data. Once this process is underway and there’s enough data to get started, the Experience Hub will begin reporting against some key metrics.

These are the metrics you’ll use in your team to determine your campaign’s impact, and you’ll likely use these metrics in wider merchandising and marketing discussions.

You can review this data by opening your campaign and navigating to Campaign metrics:

metrics
Note

For campaigns launched with audience splits of 50% and 95%, you’ll need to see 500 converters in the control and the campaign (variation) you can learn anything meaningful about campaign performance. You won’t see any metrics until this threshold has been passed.

It’s also recommended to refer to the glossary where you can find explanations for Experience Hub terms.

Post-launch

As mentioned above, for campaigns running at 50% and 95%, there’s a minimum threshold that needs to be passed before showing metrics. The magic number is 500!

So when you first launch your campaign and until this number is reached, you’ll see Not enough data for the metrics.

Once this threshold is passed, you can start reporting the changes for each metric. You can see this process in play in the following example:

cr-left-to-right
Some words about…​ 100% campaigns

For a start, for campaigns launched at 100%, you won’t need to wait for the 500 converters threshold. This is because with 100% visibility you’re not comparing your campaign to your default website content–there’s no A/B testing. All your visitors will see your campaign.

You’ll still have to wait for your campaign to generate data before you can show metrics.

You’ll also notice that the metrics you use for 100% campaigns are different.

Getting updates

Sitewide impact

There are three levels of reporting. The first is sitewide–this is an aggregation of ALL the campaigns on your site. Sitewide metrics are a great way of proving the value of the Experience Hub and your returns for your hard work and investment.

To get the sitewide impact, select Campaigns from the side menu.

Here’s an example:

sitewide-example

Campaign impact

At the next level, the campaign, you’ll find out how things are going for each of your campaigns.

To get the campaign impact, open one of your live or paused campaigns

You’ll use these metrics to understand the impact of the campaign:

campaign-metrics
Note

The metrics the Experience Hub reports for campaigns with a 50:50 and 95:5 splits are different from the metrics reported for campaigns launched at 100.

To dive into some additional campaign metrics, such as clickthrough rate and Revenue per converter (RPC), click Show campaign results.

In this example, you can see that the Experience Hub is confident in the data reported. As a result, the campaign is a winner:

campaign-metrics

You can find an explanation of each of the metrics reported against at the bottom of this article.

Experience impact

Switching to Experience metrics, you’ll also find experience-level metrics that will prove especially useful when trying to understand how each experience stacks up against your default website content:

attribution

Or, if you chose to replace your control, how experiences performed head-to-head.

These experience-level metrics align exactly with your personalized content, recommendations, and badging campaign configurations. This means that any experience, recommendation, or badge you create will get its own experience-level metrics.

Historical metrics

Each time a campaign is edited (see below) and re-launched, the Experience Hub stores a new revision and resets experience metrics to zero. The Experience Hub does this because metrics need to come from the same underlying experiences to be statistically meaningful. By starting with a new set of metrics, it’s ensured that the data reported is based on a fair comparison.

Note

Returning visitors who saw the initial and new revisions will be counted in each revision’s experience metrics. It’s therefore normal for the sum of visitors across revisions to not add up precisely to the total number of visitors in the campaign metrics.

When navigating to the "Experience metrics" tab, the data displayed is for the most recent revision by default. You can review the metrics of a previous revision by selecting it from the dropdown shown below. The date ranges in the dropdown represent the first and last day on which we calculated metrics for a revision. The dates may differ slightly from when the campaign was published or paused and will appear on newly launched revisions once we’ve computed the first batch of metrics.

attribution

Editing a campaign refers to any action that changes the campaign’s logic, including:

  • Adding, removing, or re-ordering experiences

  • Changing the audience, strategy, content, or rules of an experience

  • Adjusting the campaign from single to multi-variant, etc

Under the hood

Let’s focus now on a metric card and discuss how to interpret the results.

Data confidence

If you look at the example below, you can see what happens to data confidence as the campaign progresses and gathers more data, moving from 24% to 96%. Data confidence is an expression of how confident the Experience Hub is in the reported change for a metric–as more people see your campaign and the longer it runs for, the more confident Experience Hub can be in the result.

conv-rate-uplift

Finding winners and losers

Whether your campaign is a winner, loser, or something in between, the Experience Hub will suggest what next steps you can take. For example, you possibly get a suggestion to increase the number of people that see the campaign, pausing it, or even editing it. This will be covered in detail in the next article Taking action.

You’ll be kept updated about the progress your campaign is making, so you’ll always have the metrics to back up whatever decision you take.

Winner

The Experience Hub will declare the campaign a winner when:

  • Enough people have seen your campaign

  • It’s been running for long enough

  • The probability of an uplift in RPV is higher than 95%

Here’s an example:

winner

Loser

The Experience Hub will declare the campaign a loser when:

  • Enough people have seen your campaign

  • It’s been running for long enough

  • The probability of an uplift in RPV is less than 5%

Here’s an example:

loser

Somewhere in the middle

The Experience Hub will declare the campaign neither a winner nor a loser when:

  • Enough people have seen your campaign

  • It’s been running for long enough

  • The probability of an uplift in RPV is more than 5% but less than 95%

Before getting to the point where a campaign can be declared a winner, loser, or neither, you’ll be given a heads up about the trend the Experience Hub is seeing:

positive

The difference between a trending positive and winning campaign is only really the amount of data. When a campaign is trending positive or negative, the Experience Hub doesn’t have the right amount of data to be confident in the uplift reported.

You should pay particular attention to campaigns that are trending negative because at this stage, you still have time to intervene and make changes by changing the imagery or message.

Hold ya horses

Of course, the Coveo team understands that what a winning campaign looks like to you and your team is a little more complicated than just looking at a single metric, such as RPV.

You and your team will know the reasons for launching a campaign and precisely what you are trying to achieve. To help you build a more balanced picture of each of your campaigns, the Experience Hub provides a host of over metrics, including view and click-based attribution–these are the metrics merchandisers look for as the strongest indicators of customer relationships.

Using click-based attribution, for example, is a great way to demonstrate your campaign’s impact. Amongst other things, it shows:

  • How well your brand resonates with your customers

  • How engaging your content is

  • The effectiveness of your funnel, from the initial engagement at the top to that order at the bottom

Note

You’ll find a definition for each metric in our glossary.

Your next steps

Whatever the result, the Experience Hub will suggest what actions you can take. This could include:

  • Moving from a 50% visibility to 95% to engage more customers, where the Experience Hub has found a winner.

  • Waiting a bit longer for the data, where it’s trending positive, but the Experience Hub is not sure of the result yet.

  • Pausing it or making changes, where the Experience Hub has found a loser (and this will happen) or if it’s trending negative.

The next section covers how you take action following test results, Taking action.

A quick word about losing campaigns

The Experience Hub is a tool that allows you to get personalized content, product recommendations, and badging on your website without having to rely on a dev team and in a highly repeatable way; getting a campaign up and running can take minutes. This gives you a unique opportunity to discover what works with your customers and, equally, what doesn’t.

A losing campaign and finding out what doesn’t work is just as important as finding out what does. You’ll likely have many losing campaigns, especially at the outset. Take away the lessons, use the metrics the Experience Hub provides to understand why, and build those lessons into your next campaign.

Exporting campaign data

We understand that you sometimes need to perform deeper analyses of your campaigns, typically in a spreadsheet or BI tool. To achieve this, you can export a CSV containing all the metrics you see in the UI for every day a campaign has been live. To download this data, click Export CSV.

export-csv

Here’s more information on the CSV’s contents:

Field Definition Metric source Example Suggested formatting

string_date

The campaign metric date

n/a

2022-01-01

Text

campaign_id

The campaign’s unique identifier

n/a

abc123-def456

Text

allocation

The maximum allocation towards the campaign’s variant

n/a

0.5

Percentage

visitors_variant

The cumulative count of unique visitors in the campaign’s variant

Collected

100000

Integer

visitors_control

The cumulative count of unique visitors in the campaign’s control

Collected

100000

Integer

converters_variant

The cumulative count of unique converters in the campaign’s variant

Collected

5100

Integer

converters_control

The cumulative count of unique converters in the campaign’s control

Collected

5000

Integer

clickers_variant

The cumulative count of unique clickers in the campaign’s variant

Collected

10500

Integer

clickers_control

The cumulative count of unique clickers in the campaign’s control

Collected

10000

Integer

cvr_variant

The cumulative ratio of converters to visitors for the variant

Analyzed by the stats engine

0.051

Percentage

cvr_control

The cumulative ratio of converters to visitors for the control

Analyzed by the stats engine

0.050

Percentage

cvr_uplift

The uplift in conversion rate

Analyzed by the stats engine

0.020

Percentage

ctr_variant

The cumulative ratio of clickers to visitors for the variant

Analyzed by the stats engine

0.105

Percentage

ctr_control

The cumulative ratio of clickers to visitors for the control

Analyzed by the stats engine

0.100

Percentage

ctr_uplift

The uplift in clickthrough rate

Analyzed by the stats engine

0.050

Percentage

rpv_variant

The cumulative ratio of attributed revenue to visitors for the variant

Analyzed by the stats engine

5.050

Your property’s currency

rpv_control

The cumulative ratio of attributed revenue to visitors for the control

Analyzed by the stats engine

5.000

Your property’s currency

rpv_uplift

The uplift in revenue per visitor

Analyzed by the stats engine

0.010

Percentage

rpc_variant

The cumulative ratio of attributed revenue to converters for the variant

Analyzed by the stats engine

50.500

Your property’s currency

rpc_control

The cumulative ratio of attributed revenue to converters for the control

Analyzed by the stats engine

50.000

Your property’s currency

rpc_uplift

The uplift in revenue per converter

Analyzed by the stats engine

0.010

Percentage

incremental_revenue

Incremental revenue attributed the variant computed as the multiplication of variant visitors, control RPV, and RPV uplift

Analyzed by the stats engine

5000

Your property’s currency

impression_converters_variant

The cumulative count of unique converters that previously saw the campaign’s variant

Collected

5100

Integer

impression_converters_control

The cumulative count of unique converters that previously saw the campaign’s control

Collected

5000

Integer

impression_revenue_variant

The cumulative sum of revenue attributed to visitors that previously saw the campaign’s variant

Collected

505000

Your property’s currency

impression_revenue_control

The cumulative sum of revenue attributed to visitors that previously saw the campaign’s control

Collected

500000

Your property’s currency

clickthrough_converters_variant

The cumulative count of unique converters that previously clicked the campaign’s variant

Collected

2050

Integer

clickthrough_converters_control

The cumulative count of unique converters that previously clicked the campaign’s control

Collected

2000

Integer

clickthrough_revenue_variant

The cumulative sum of revenue attributed to visitors that previously clicked the campaign’s variant

Collected

202000

Your property’s currency

clickthrough_revenue_control

The cumulative sum of revenue attributed to visitors that previously clicked the campaign’s control

Collected

200000

Your property’s currency

Campaign metrics

Campaigns with a standard control

Audience split Metric Explanation

50% and 95%

Incremental Revenue

A prediction of the additional revenue generated by your campaign, based on the current Revenue Per Visitor uplift

Conversion Rate

The percentage of visitors that saw the campaign and went on to convert

Revenue Per Visitor

The total revenue divided by the total number of unique campaign visitors

Visitors

The number of unique campaign visitors

Unique clicks

The number of unique clicks an experience link

Clickthrough Rate

The total number of unique visitors that clicked an experience link at least once divided by the total number of unique visitors in the campaign

Conversion Rate

The total number of unique converters divided by the total number of unique visitors in the campaign

Revenue Per Customer

The total revenue divided by the total number of unique campaign converters

100%

Impression Revenue

The total amount of revenue from orders placed by visitors that saw the campaign

Conversion Rate

The total number of unique converters divided by the total number of unique visitors in the campaign

Revenue Per Visitor

The total revenue divided by the total number of unique campaign visitors

Campaigns with a replaced control

Audience split Metric Explanation

50% and 95%

Incremental Revenue

A prediction of the additional revenue generated by your campaign and is based on the current Revenue Per Visitor uplift

Conversion Rate

The percentage of visitors that saw the campaign and went on to convert

Revenue Per Visitor

The total revenue divided by the total number of unique campaign visitors

Visitors

The number of unique campaign visitors

Unique clicks

The number of unique clicks an experience link

Clickthrough Rate

The total number of unique visitors that clicked an experience link at least once divided by the total number of unique visitors in the campaign

Conversion Rate

The total number of unique converters divided by the total number of unique visitors in the campaign

Revenue Per Customer

The total revenue divided by the total number of unique campaign converters