Statistical hypothesis testing implies that no test is ever 100% certain: that’s because we rely on probabilities to experiment.

When online marketers and scientists run hypothesis tests, both seek out **statistically relevant results. **This means that the results of their test have to be true within a range of probabilities (typically 95%).

Even though hypothesis tests are meant to be reliable, **there are two types of errors that can occur. **

**These errors are known as type 1 and type 2 errors.**

## Understanding Type 1 errors

Type 1 errors – often assimilated with false positives – happen in hypothesis testing when the null hypothesis is true but rejected. *The null hypothesis is a general statement or default position that there is no relationship between two measured phenomena.*

Simply put, type 1 errors are “false positives” – they happen when the tester validates a statistically significant difference even though there isn’t one.

Type 1 errors have a probability of “α” correlated to the level of confidence that you set. A test with a 95% confidence level means that there is a 5% chance of getting a type 1 error.

## Consequences of a type 1 Error

Type 1 errors can happen due to bad luck (the 5% chance has played against you) or because you didn’t respect the test duration and sample size initially set for your experiment.

Consequently, a type 1 error will bring in a false positive. This means that you will wrongfully assume that your hypothesis testing has worked even though it hasn’t.

In real life situations, this could potentially mean losing possible sales due to a faulty assumption caused by the test.

**Related: Sample Size Calculator for A/B Testing**

## A real-life example of a type 1 error

Let’s say that you want to increase conversions on a banner displayed on your website. For that to work out, you’ve planned on adding an image to see if it increases conversions or not.

You start your A/B test running a control version (A) against your variation (B) that contains the image. After 5 days, the variation (B) outperforms the control version by a staggering 25% increase in conversions with an 85% level of confidence.

You stop the test and implement the image in your banner. However, after a month, you noticed that your month-to-month conversions have actually decreased.

That’s because you’ve encountered a type 1 error: your variation didn’t actually beat your control version in the long run.

## Understanding type 2 errors

If type 1 errors are commonly referred to as “false positives”, **type 2 errors are referred to as “****false negatives”****.**

Type 2 errors happen when you inaccurately assume that no winner has been declared between a control version and a variation although there actually is a winner.

In more statistically accurate terms, **type 2 errors happen when the null hypothesis is false and you subsequently fail to reject it**.

If the probability of making a type 1 error is determined by “α”, the probability of a type 2 error is “β”. Beta depends on the power of the test (i.e the probability of not committing a type 2 error, which is equal to 1-β).

**There are 3 parameters that can affect the power of a test:**

- Your sample size (n)
- The significance level of your test (α)
- The “true” value of your tested parameter (read more here)

## Consequences of a type 2 error

Similarly to type 1 errors, type 2 errors can lead to false assumptions and poor decision making that can result in lost sales or decreased profits.

Moreover, getting a false negative (without realizing it) can discredit your conversion optimization efforts even though you could have proven your hypothesis. This can be a discouraging turn of events that could happen to all CRO experts and digital marketers.

## A real-life example of a type 2 error

Let’s say that you run an e-commerce store that sells high-end, complicated hardware for tech-savvy customers. In an attempt to increase conversions, you have the idea to implement an FAQ below your product page.

You launch an A/B test to see if the variation (B) could outperform your control version (A).

After a week, **you do not notice any difference** in conversions: both versions seem to convert at the same rate and you start questioning your assumption. Three days later, you stop the test and keep your product page as it is.

At this point, you assume that adding an FAQ to your store didn’t have any effect on conversions.

Two weeks later, you hear that a competitor has implemented an FAQ at the same time and observed tangible gains in conversions. You decide to re-run the test for a month in order to get more statistically relevant results based on an increased level of confidence (say 95%).

After a month – surprise – **you discover positive gains in conversions for the variation (B)**. Adding an FAQ at the bottom of your product page has indeed brought your company more sales than the control version.

That’s right – your first test encountered a type 2 error!

**Want to know more?**

Download our ebook Demystifying A/B Testing Statistics

In this ebook, you’ll learn how to understand different statistical methods with side-by-side comparisons and how to read A/B test results with confidence.