# Types of Statistical Errors and What They Mean

by in Statistics
From the Bitesize Bio channel

This column is loaded with pop quizzes for you to test yourself on. If you haven’t already done so, catch up on yesterday’s piece on hypothesis testing for a refresher.

Take a gander at the table below for a summary of the two types of error that can result from hypothesis testing.

Type I Errors occur when we reject a null hypothesis that is actually true; the probability of this occurring is denoted by alpha (a).

Type II Errors are when we accept a null hypothesis that is actually false; its probability is called beta (b). As you can see from the below table, the other two options are to accept a true null hypothesis, or to reject a false null hypothesis. Both of these are correct, though one is far more exciting than the other (and probably easier to get published).

It all looks really simple (I hope) when you put it in a table like that. The trouble is we don’t know whether the null hypothesis is true or not; that’s the whole point of statistics! So instead we are reliant on the probabilities of each type of error occurring. As in life, nothing is ever easy, so in statistics we cannot minimise the probability of both errors simultaneously. By reducing the probability of Type I Errors, we automatically increase the probability of a Type II Error occurring, and vice versa.

Pop Quiz: Given the conundrum, which type of error do we focus on minimising? You’ve got the next three paragraphs to come up with your answer (no looking at your neighbours, no texting, papers face-down on your desk when you’re finished).

For a working example I’ll depart from biology for a moment and move to medicine. Pharmaceutical Company Delta-Theta has manufactured a new pill they claim relieves headaches. As all good pharmaceutical companies do they have conducted a double-blind study* to test the effects of their pill.

Pop Quiz: Following on from last week, you should be able to tell me the null and alternate hypotheses (Go!).

The null hypothesis is that the medicine does not relieve headaches (or, that it does no better than a placebo pill). So the alternate states that the pill does relieve headaches, at least in comparison to a placebo.

Pop Quiz:What then, would constitute a Type I and Type II error?

A Type I Error refers to rejecting a true null hypothesis; so this would mean that in truth the pill does not relieve headaches, but the pharmaceutical company concludes that it does. Whereas a Type II Error occurs when we accept a false null hypothesis; the pill actually does relieve headaches, but Delta-Theta concludes that it doesn’t. [If you’re still getting your head around which error is which, don’t worry – I still needed to check my table to make sure I had them the right way around].

So, finally we can return to the question I posed at the start of this article: which type of error do we focus on minimising? The reason I’ve used a medical example is because of the hefty weight attached to right or wrong decisions. A Type II Error would falsely conclude that the pill does no good, and so it wouldn’t be put into the market, resulting in a loss for the company. However, a Type I Error concludes that the pill does work, when actually it doesn’t. The pill gets put into production and enters the market; patients take it and are not relieved from headaches. Loss for the consumer.

But what if the medicine wasn’t for headaches? What if it was for HIV, or cancer, or diabetes? What if a patient stopped taking their regular medication in order to take this new pill? Hence the hefty cost of a wrong decision.

Thus, we aim to minimise the probability of a Type I Error occurring, at the cost of Type II errors. The same argument stands for publishing biological results. Which is more embarrassing and career damaging, publishing incorrect results (Type I) or failing to recognise and publish significant results?

Biologists have reached a consensus on what probability of a Type I error we are prepared to accept; the magical 5 % threshold.  This leads to a discussion of p-values, our next subject.

Until then, you are very welcome to leave your comments and feedback on the statistics series thus far.

*A double-blind study is where neither the patient nor the doctor knows whether the actual drug or a placebo has been administered. The purpose being to negate any change in the doctor’s behaviour if he or she knows which the patient has been allocated, e.g. by giving less attention to patients on the placebo. I highly recommend pretty much anything Ben Goldacre has written if you wish to delve further into this fascinating subject of human psychology and medical testing.

Enter your email to be informed when we publish more articles like this on BsB, and also get access to all of these goodies:

• Free ebooks and audiobooks on the topics that matter to you
• Advance notice of new webinars and eBooks

# Statistics: A Good P-value is Not Enough

Like many scientists, I don’t consider myself a statistics expert. But I am determined to do things right in my science, and that includes statistics. In my experience, a lot of scientists who are “scared” of statistics fall into the trap of ignoring the existence of anything beyond a t-test. But using the right method [...]

# Don’t Be Another P-value Victim

In previous articles, I’ve primed you on hypothesis testing and how we are forced to choose between minimising either Type I or Type II errors. In the world of the null hypothesis fetish, the p-value (p) is the most revered number. It may also be the least understood. The p-value is the probability, assuming the [...]

# Pseudoreplication: Don’t Fall For This Simple Statistical Mistake

Now we come to the third part of our trifecta; in the last two posts I have gone over p-values and how they determine significance in null hypothesis testing, and we talked about degrees of freedom and their effect on the p-value. Finally, we come to pseudoreplication: where it can all go terribly wrong. Replication [...]

# A Primer on Statistical Hypotheses

Hypothesis testing is the foundation around which we prove our science is worth funding, publishing and sitting through a conference presentation for. I can’t overstate the  importance of understanding hypothesis testing, such is the integral part it plays in biological analyses. The Null Hypothesis Fundamental to statistics is the concept of a null hypothesis, and [...]

# Sarah-Jane O'Connor

Sarah-Jane is an ecologist who sometimes masquerades as a geneticist. Her statistical knowledge is embarassing in some social circles, but revered in others. Which probably just makes it neutral. She has a PhD in ecology from the University of...

## One comment

1. from on

Thank you for the informative reading

## Subscribe to Channels

To receive information about any of our new channels click on the button below.
subscribe to the channel newsletter »

## Write for us

Have a short tip, a written
article or a video you'd like
to see published?
write for us »