Skip to main content
What is Precision vs. Recall?
  1. Glossary/

What is Precision vs. Recall?

7 mins·
Ben Schmidt
Author
I am going to help you build the impossible.

You have just launched a new feature. Maybe it is a recommendation engine for an e-commerce site or a fraud detection system for a fintech app. The engineering team tells you the model is ninety percent accurate. That sounds incredible.

But then users start complaining.

Customers are seeing irrelevant product suggestions. Or worse, legitimate users are getting flagged as fraudsters and locked out of their accounts. You are left wondering how a model with such high accuracy can result in such a poor user experience.

The answer usually lies in the nuance of how you measure success. It comes down to two specific metrics: precision and recall.

Understanding the difference between these two concepts is not just a job for your data scientists. It is a strategic imperative for founders. The balance you strike between precision and recall defines the personality of your product and the risks your business is willing to accept.

Defining the Terms

#

At a high level, these metrics help you understand the quality of the results your system retrieves or classifies. To understand them, you first have to understand that your system makes predictions. Sometimes it predicts correctly, and sometimes it predicts incorrectly.

Precision answers a specific question. Of all the items the system identified as relevant, how many actually were relevant?

Think of this as a measure of exactness or quality. If your system flags ten transactions as fraudulent, and nine of them actually are fraud, you have high precision. You are not wasting time on false alarms.

Recall answers a different question. Of all the relevant items that exist in the entire dataset, how many did the system manage to find?

Think of this as a measure of completeness or quantity. If there were one hundred total fraudulent transactions occurred yesterday, and your system only caught those nine we mentioned earlier, you have very low recall. You missed ninety-one cases.

Precision focuses on the results you see. Recall focuses on the results you missed.

The Necessary Trade-Off

#

Founders often ask why they cannot have one hundred percent of both. In a perfect world with perfect data, perhaps you could. In the messy reality of business and human behavior, these two metrics are almost always at odds.

Imagine you are casting a net to catch a specific type of fish.

To ensure you catch every single one of those specific fish (high recall), you have to use a very large net with a fine mesh. You will catch all the fish you want. However, you will also catch old boots, tires, and other fish you did not want. Your precision goes down because the ratio of target fish to total debris is low.

Now imagine you want to ensure you catch only that specific fish with zero garbage (high precision). You might switch to using a spear. You will only strike when you are absolutely certain. You will not catch any boots. But you will likely miss a vast number of the target fish swimming by because you were too hesitant to strike.

This push and pull exists in every algorithm. When you tune a model to be more sensitive to ensure you don’t miss anything, you introduce noise. When you filter out the noise, you inevitably accidentally filter out some valid signals.

The Cost of Being Wrong

#

The decision to prioritize precision over recall, or vice versa, is not a mathematical one. It is a business decision based on the cost of making a mistake.

You have to look at the consequences of False Positives versus False Negatives.

A False Positive occurs when the system says something is true, but it is not. A False Negative occurs when the system says something is false, but it is actually true.

Let’s look at a spam filter.

Precision is about trust.
Precision is about trust.
If a spam filter has low precision, it flags legitimate emails as spam. This is a False Positive. If an investor emails you a term sheet and it goes to your spam folder, that is a catastrophe. Therefore, most email providers prioritize precision. They would rather let a few spam emails into your inbox (lower recall) than risk deleting a crucial email.

Now consider a medical screening tool for a serious disease.

If the tool misses a positive case, the patient does not get treatment. This is a False Negative. It is a life-or-death scenario. In this case, doctors prioritize recall. They would rather flag a healthy person for further testing (a False Positive) than miss a sick person. The cost of a False Positive is just anxiety and the cost of a follow-up test. The cost of a False Negative is much higher.

Applying This to Your Startup

#

As you build your internal tools or customer-facing products, you need to ask your team which error is more expensive to your business model.

Consider a hiring algorithm that scans resumes.

If you are a small startup overwhelmed with thousands of applications, you might value precision. You only want to see candidates who are a perfect match. You are okay with missing out on a few hidden gems (low recall) because you simply do not have the time to interview everyone. You need efficiency.

However, if you are a specialized recruiter looking for a very rare executive skillset, you value recall. You want to see every resume that even remotely matches the description. You cannot afford to let the algorithm filter out a potential candidate just because their formatting was weird. You are willing to sift through some bad matches to find the right person.

Ask yourself these questions regarding your data products:

What happens to the user if we show them the wrong thing? Does it annoy them, or does it cause them harm?

What happens to the business if we miss a valid opportunity? Do we lose a few dollars, or do we lose our reputation?

The Unknowns in the Data

#

There is a danger in relying too heavily on these metrics without looking at the context. Precision and recall are calculated based on labeled data. This assumes that we know the absolute truth about what is right and wrong in our historical data.

But for many startups, truth is subjective.

If you are building a content moderation bot, what counts as “offensive”? If your human moderators have been inconsistent in the past, your metrics will be flawed. A model might have high precision based on training data, but in the real world, it fails because the definitions have shifted.

We also do not always know the full scope of what we are missing. Measuring recall is difficult because, by definition, it requires you to know how many relevant items you failed to find. In a live production environment, you often do not know what you do not know.

How do you measure the customers who churned because they couldn’t find what they were looking for?

How do you account for the bias inherent in how you collected your initial data?

These are the areas where a founder’s intuition must overlay the data science. You cannot optimize a metric in a vacuum. You must look at the user behavior that results from the metric.

Finding the Balance

#

There is a metric called the F1 Score, which is the harmonic mean of precision and recall. It tries to provide a single number that balances both. While useful for data scientists comparing models, it can sometimes obscure the business reality.

Sometimes a balanced score is not what you want. Sometimes you need to be extreme.

Do not let your team just report “accuracy.” Push them to break it down. Ask them to show you the confusion matrix. Ask them to explain a False Positive in the context of a user story.

Build your product with the understanding that the machine will be wrong. When you optimize for precision, build mechanisms to catch what falls through the cracks. When you optimize for recall, build workflows to help users filter the noise.

Great products are not just about great algorithms. They are about how the business handles the inevitable errors those algorithms make.