Article

13min read

Better Understand (And Optimize) Your Average Basket Size

When it comes to using A/B testing to improve the user experience, the end goal is about increasing revenue. However, we more often hear about improving conversion rates (in other words, changing a visitor into a buyer). 

If you increase the number of conversions, you’ll automatically increase revenue and increase your number of transactions. But this is just one method among many…another tactic is based on increasing the ‘average basket size’. This approach is, however, much less often used. Why? Because it’s rather difficult to measure the associated change.

A Measurement and Statistical Issue

When we talk about statistical tests associated with average basket size, what do we mean? Usually, we’re referring to the Mann-Whitney-U test (also called the Wilcoxon), used in certain A/B testing software, including AB Tasty.  A ‘must have’ for anyone who wants to improve their conversion rates. This test shows the probability that variation B will bring in more gain than the original. However, it’s impossible to tell the magnitude of that gain – and keep in mind that the strategies used to increase the average basket size most likely have associated costs.  It’s therefore crucial to be sure that the gains outweigh the costs. 

For example, if you’re using a product recommendation tool to try and increase your average basket size, it’s imperative to ensure that the associated revenue lift is higher than the cost of the tool used….

Unfortunately, you’ve probably already realized that this issue is tricky and counterintuitive…

Let’s look at a concrete example: the beginner’s approach is to calculate the average basket size directly. It’s just the sum of all the basket values divided by the number of baskets. And this isn’t wrong, since the math makes sense. However, it’s not very precise! The real mistake is comparing apples and oranges, and thinking that this comparison is valid. Let’s do it the right way, using accurate average basket data, and simulate the average basket gain. 

Here’s the process:

  • Take P, a list of basket values (this is real data collected on an e-commerce site, not during a test). 
  • We mix up this data, and split them into two groups, A and B.
  • We leave group A as is: it’s our reference group, that we’ll call the ‘original’.
  • Let’s add 3 euros to all the values in group B, the group we’ll call the ‘variation’, and which we’ve run an optimization campaign on (for example, using a system of product recommendations to website visitors). 
  • Now, we can run a Mann-Whitney test to be sure that the added gain is significant enough. 

With this, we’re going to calculate the average values of lists A and B, and work out the difference. We might naively hope to get a value near 3 euros (equal to the gain we ‘injected’ into the variation). But the result doesn’t fit. We’ll see why below. 

How to Calculate Average Basket Size

The graph below shows the values we talked about: 10,000 average basket size values. The X (horizontal) axis represents basket size, and the Y (vertical) axis, the number of times this value was observed in the data.

It seems that the most frequent value is around 50 euros, and that there’s another spike at around 100 euros, though we don’t see many values over 600 euros. 

After mixing the list of amounts, we split it into two different groups (5,000 values for group A, and 5,000 for group B).

Then, we add 3 euros to each value in group B, and we redo the graph for the two groups, A (in blue) and B (in orange): 

We already notice from looking at the chart that we don’t see the effect of having added the 3 euros to group B: the orange and blue lines look very similar. Even when we zoom in, the difference is barely noticeable: 

However, the Mann-Whitney-U test ‘sees’ this gain:

More precisely, we can calculate pValue = 0.01, which translates into a confidence interval of 99%, which means we’re very confident there’s a gain from group B in relation to group A. We can now say that this gain is ‘statistically visible.’ 

We now just need to estimate the size of this gain (which we know has a value of 3 euros).

Unfortunately, the calculation doesn’t reveal the hoped for result! The average of group A is 130 euros and 12 cents, and for version B, it’s 129 euros and 26 cents. Yes, you read that correctly: calculating the average means that average value of B is smaller than the value of A, which is the opposite of what we created in the protocol and what the statistical test indicates. This means that, instead of gaining 3 euros, we lose 0.86 cents!  

So where’s the problem? And what’s real? A > B or B > A?

The Notion of Extreme Values

The fact is, B > A! How is this possible? It would appear that the distribution of average basket values is subject to ‘extreme values’. We do notice on the graph that the majority of the values is < 500 euros.

But if we zoom in, we can see a sort of ‘long tail’ that shows that sometimes, just sometimes, there are values much higher than 500 euros. Now, calculating averages is very sensitive to these extreme values. A few very large basket size values can have a notable impact on the calculation of the average. 

What’s happening then? When we split up the data into groups A and B, these ‘extreme’ values weren’t evenly distributed in the two groups (neither in terms of the number of them, nor their value). This is even more likely since they’re infrequent, and they have high values (with a strong variance). 

NB: when running an A/B test, website visitors are randomly assigned into groups A and B as soon as they arrive on a site. Our situation is therefore mimicking the real-life conditions of a test. 

Can this happen often? Unfortunately, we’re going to see that yes it can. 

A/A Tests

To give a more complete answer to this question, we’d need to use a program that automates creating A/A tests, i.e. a test in which no change is made to the second group (that we usually call group B). The goal is to check the accuracy of the test procedure. Here’s the process:

  1. Mix up the initial data
  2. Split it into two even groups
  3. Calculate the average value of each group
  4. Calculate the difference of the averages

By doing this 10,000 times and by creating a graph of the differences measured, here’s what we get:

X axis: the difference measured (in euros) between the average from groups A and B. 

Y axis: the number of times this difference in size was noticed.

We see that the distribution is centered around zero, which makes sense since we didn’t insert any gain with the data from group B.  The problem here is how this curve is spread out: gaps over 3 euros are quite frequent. We could even wager a guess that it’s around 20%. What can we conclude? Based only on this difference in averages, we can observe a gain higher than 3 euros in about 20% of cases – even when groups A and B are treated the same!

Similarly, we also see that in about 20% of cases, we think we’ll note a loss of 3 euros per basket….which is also false! This is actually what happened in the previous scenario: splitting the data ‘artificially’ increased the average for group A. The gain of 3 euros to all the values in group B wasn’t enough to cancel this out. The result is that the increase of 3 euros per basked is ‘invisible’ when we calculate the average. If we look only at the simple calculation of the difference, and decide our threshold is 1 euro, we have about an 80% chance of believing in a gain or loss…that doesn’t exist!

Why Not Remove These ‘Extreme’ Values?

If these ‘extreme’ values are problematic, we might be tempted to simply delete them and solve our problem. To do this, we’d need to formally define what we call an extreme value. A classic way of doing this is to use the hypothesis that the data follow ‘Gaussian distribution’. In this scenario, we would consider ‘extreme’ any data that differ from the average by more than three times the standard deviation. With our dataset, this threshold comes out to about 600 euros, which would seem to make sense to cancel out the long tail. However, the result is disappointing. If we apply the A/A testing process to this ‘filtered’ data, we see the following result: 

The distribution of the values of the difference in averages is just as big, the curve has barely changed. 

If we were to do an A/B test now (still with an increase of 3 euros for version B), here’s what we get (see the graph below). We can see that the the difference is being shown as negative (completely the opposite of the reality), in about 17% of cases! And this is discounting the extreme values. And in about 18% of cases, we would be led to believe that the gain of group B would be > 6 euros, which is two times more than in reality!

Why Doesn’t This Work?

The reason this doesn’t work is because the data for the basket values doesn’t follow Gaussian distribution. 

Here’s a visual representation of the approximation mistake that happens:

The X (horizontal) axis shows basket values, and the Y (vertical) axis shows the number of times this value was observed in this data. 

The blue line represents the actual basket values, the orange line shows the Gaussian model. We can clearly see that the model is quite poor: the orange curve doesn’t align with the blue one. This is why simply removing the extreme values doesn’t solve the problem. 

Even if we were able to initially do some kind of transformation to make the data ‘Gaussian’, (this would mean taking the log of the basket values), to significantly increase the similarity between the model and the data, this wouldn’t entirely solve the problem. The variance of the different averages is just as great.

During an A/B test, the estimation of the size of the gain is very important if you want to make the right decision. This is especially true if the winning variation has associated costs. It remains difficult today to accurately calculate the average basket size. The choice comes down soley to your confidence index, which only indicates the existence of gain (but not its size). This is certainly not ideal practice, but in scenarios where the conversion rate and average basket are moving in the same direction, the gain (or loss) will be obvious. Where it becomes difficult or even impossible to make a relevant decision is when they aren’t moving in the same direction. 

This is why A/B testing is focused mainly on ergonomic or aesthetic tests on websites, with less of an impact on the average basket size, but more of an impact on conversions. This is why we mainly talk about ‘conversion rate optimization’ (CRO) and not ‘business optimization’. Any experiment that affects both conversion and average basket size will be very difficult to analyze. This is where it makes complete sense to involve a technical conversion optimization specialist: to help you put in place specific tracking methods aligned with your upsell tool.

To understand everything about A/B testing, check out our article: The Problem is Choice.

Subscribe to
our Newsletter

bloc Newsletter EN

We will process and store your personal data to respond to send you communications as described in our  Privacy Policy.

Article

8min read

What Exactly Does Customer Experience (CX) Mean? And How Can We Measure It?

It’s a competitive world online. In order to beat your competition, you need world-class customer experience (CX). 

But what exactly does customer experience mean, anyway? 

This is a non-technical guide to customer experience; we’ll cover what CX is, why it’s important, how it benefits your business, and how you can measure it. 

By the end of the article, you’ll have everything you need to start measuring, improving, and delivering an exceptional experience for your customers.

The Definition of Customer Experience

Customer experience is a big deal. 

So much so that a study by Econsultancy found that the number one priority for more than 2,000 marketers was customer experience when asked the question:

“Over the next five years, what is the primary way your organization will seek to differentiate itself from competitors?”

So what is customer experience? 

Customer experience (CX for short) is the perception of your customers’ experience with your business or brand. 

Every time a customer interacts with your brand and everything your business does, it has an impact on how your customers perceive your business. 

Customer experience is the culmination of all of these experiences and interactions, impacting your decision whether to return to the business or make a purchase. 

Simply put, the key to your success is epic customer experience. 

The Importance of Customer Experience for Your Business

The core benefits of improving your customer experience include the following:

  1. Increase the loyalty of your customers
  2. Increase the happiness and satisfaction of your customers
  3. Increase positive reviews and recommendations
  4. Reduce returns 
  5. Reduce complaints
  6. Reduce churn 

There is no doubt that to be a successful business, you need to put your customers first. If you do, it’s not only good for them, but it’s also good for your business. 

The Customer Journey and Customer Experience

As we’ve seen, customer experience is concerned with the holistic experience of your customers with your brand. 

This means that every step of the customer journey has the potential to impact on your customer’s experience. 

In terms of customer experience, the stages of a customer’s journey include discovery, research, purchase, and post-purchase customer support. 

It should go without saying, but in terms of customer experience, there’s one extremely important caveat to the customer journey.

Instead of considering each stage of the journey from the business’s perspective, it’s imperative to consider these touchpoints from the perspective of the customer. 

To improve customer experience at each stage, we need to understand how and why they are different, and what we can do to improve the experiences of your customers.

To better understand customer experience, let’s break down the customer journey into stages.  

Stage One: Consideration

When a customer changes from an observer to a potential customer, they enter the “Consideration” phase. 

It’s at this stage that a customer is making a choice whether to consider your brand or exclude it from their buying journey. 

nike brand

At this point, branding is hugely important. Well-known brands have the upper hand at this point, but if you own a smaller company, the best way you can make an impact is by having a clear message that solves a customer’s problem. 

Stage Two: Evaluation

There are two ways in which a potential customer can enter the evaluation stage. 

First, they may have “selected” your brand during the consideration stage or, interestingly, they may enter the cycle fresh in the evaluation stage. 

While this may seem odd, it’s common sense. As a potential customer researches products similar to yours, it’s likely that they stumble across new brands, as well as the brands they’ve previously considered. 

The way you advertise to customers at this stage of the cycle should, therefore, reflect both of these types of customers and you should take into account the fact that many people have not yet interacted with your brand at all. 

Stage Three: Buying

Hooray! This is the stage we’ve all been waiting for. And it’s more than likely the stage that most marketers focus on. 

Everything from the text on the ‘Buy’ button to the copy on the product page is A/B tested, iterated, and scrutinized. 

Of course, this is highly important, but by placing too much emphasis on this one stage of the cycle, it could be detrimental to the customer experience overall. 

Stage Four: Post-Purchase 

Often overlooked, the post-purchase stage is an equally important part of the customer experience journey. 

How you treat your customers after they have bought from you will affect whether they choose to repeat the process…or not.

The effect of customer satisfaction after a purchase can have a huge impact on your bottom line. Consider these statistics:

  • It can cost up to 25 times as much to acquire a new customer than to retain an existing one
  • If you increase retention rates by 5%, then you can increase profits by 25% to 95%

And here’s the punch line: only 18% of businesses focus on customer retention. 

This represents a huge opportunity for you to beat your competition, increase your profits, and improve your customer experience. Win, win, win. 

The Importance of Website Personalization

“Website personalization is the real-time individualization of a website to suit each visitor’s unique needs and guide them through a custom conversion funnel.” 

And it goes hand in hand with customer experience. 

Personalization can be done at any stage of the funnel, from personalized email marketing to exit-intent popups and discount offers. 

A well-targeted personalized message evokes an emotional response and creates a stronger, more memorable connection with your brand. 

How to Measure Customer Experience

Instead of measuring the ROI of customer experience, which will likely take the focus off customer satisfaction, you should use one or more of the following industry-specific metrics: 

NPS (Net Promoter Score)

The Net Promoter Score, or NPS, measures customer experience and can also predict business growth. 

It’s a simple measurement, but it’s been adopted around the world and is seen as one of the best ways to measure your customer’s experience. 

To run the test, ask your customers on a scale of one to ten whether or not they would recommend your brand or business to a friend. 

Respondents are then grouped into “Promoters” (score 9-10), “Passives” (score 7-8), or “Dectractors” (0-6). 

Once you’ve collected your responses, subtract the percentage of Detractors from the percentage of Promoters, and you have your Net Promoter Score. 

It’s simple, straightforward, and it works. 

CES (Customer Effort Score)

Like the NPS, CES is another way to measure customer experience. 

Also using a survey, it asks customers to rank their experience with the business ranging from “Very Difficult” to “Very Easy.” 

The theory is, if you make it easy for your customers to make a purchase or solve their problem, then it’s more likely that they will continue to pay for your product or service. 

The easiest way to increase customer loyalty is not through “wowing your customers,” but rather through making it easier to get their job done.

CSAT (Customer Satisfaction Score) 

The final customer service metric is the Customer Satisfaction Score. 

It’s a basic test and, due to the difference in opinion over what is classed as “satisfactory,” its validity can be questioned. 

Nevertheless, it’s simple to do and can offer insight into how well your customers experience your brand.

The survey asks the question “How satisfied were you with your experience?” in order to gauge customer satisfaction with their purchase or interaction with your business. 

The response should be on a scale from 1–3, 1–5, or 1–10.

TTR (Time to Resolution)

Time to Resolution, or TTR, is the average time it takes for your customer service team to resolve an issue raised by a customer. 

One of the best ways to reduce customer frustration is to ensure the timely resolution of their issues. 

Unfortunately, if you don’t measure your response times, it’s impossible to know whether or not you’re hitting the mark. 

To test your TTR, simply add up all of the time it takes to resolve issues and divide the total by the number of tickets addressed. 

TTR is the average length of time it takes for customer service teams to resolve an issue. 

Conclusion

In a digital world, customer expectations are higher than ever, and word of good (or bad) service travels at the speed of light. Customer experience requires constant care and attention. Focus on a long-term customer experience strategy and always offer a customer-centric strategy. 

By understanding your customer experience journey and testing your customers’ satisfaction levels, you can improve your customers’ perceptions of your brand and, in turn, improve brand loyalty, increase retention, and achieve higher profits.