Have you too been struggling to understand what precision and recall ACTUALLY mean? Not in terms of formulae, but, actually how to interpret it? Well, I struggled with this too.
So, let me try and explain you, without using any formulae, but just intuition and a real, practical example, what they actually mean.
For this, we'll take 2 different binary classification scenarios:
Scenario 1. Classifying whether a person have committed murder(=1) or not(=0).
Scenario 2. Classifying whether a person has diabetes(=1) or not(=0).
Now, what if I were to ask you, which of these have more dire consequences:
- Falsely marking 0 as 1, or
- Falsely marking 1 as 0?
Let's take the first case: the murder one:
-
Falsely marking 0 as 1: You falsely classify an innocent person of murder.Consequence: That person dies.
-
Falsely marking 1 as 0: You falsely classify the murderer as innocent.Consequence: The murderer goes free
Which of these 2 cases is the one having more dire consequences?
Don't answer yet.
Let's do the same for second case now: the diabetes one:
-
Falsely marking 0 as 1: You falsely classify a person as having diabetes, even though that person doesn't have it.Consequence: Starts to take medicines, a few side effects, maybe, maybe, death because of those medicines
-
Falsely marking 1 as 0: You falsely classify a person as NOT having diabetes, even though that person actually has.Consequence: A definite death
Now let's answer that question about which of these scenarios is more dire:
- Wrongly convicting an innocent person of murder, and subsequently letting that person die of death punishment is far, far more dire than letting a murderer go free.
- Wrongly diagnosing a diabetic person as NOT having diabetes, and letting that person go without any medication, which ultimately will lead to the person's death is far, far more dire than wrongly diagnosing a non-diabetic person as diabetic.
That means:
-
For case 1, the consequence of
Falsely marking 0 as 1(False Positive) if greater thanFalsely marking 1 as 0(False Negative) -
For case 2, the consequence of
Falsely marking 1 as 0(False Negative) if greater thanFalsely marking 0 as 1(False Positive)
And this is where our precision and recall comes into play.
- Wherever the consequences of
False Positiveis greater than that ofFalse Negative(case 1 here), go with higherprecision. - Wherever the consequences of
False Negativeis greater than that ofFalse Positive(case 2 here), go with higherrecall.
Let me expand on this with another example, but, this time a little less intuitive, but, more close to numbers:
- Consider a certain binary classification problem in which true
1cases are 8, but, your model gave 10. In this case, those 2 extra1's arefalse positives. This is whatprecisioncan mathematically tell you. This here has aprecisionof: 8/(8+2) => 0.8 => 80%. - Now consider that true no. of
1outcomes are actually 10, but, your model gave only 8 (rest 2 were classified as 0). In this case, those 2 less1's, which were classified as0's arefalse negatives. This is whatrecallcan mathematically tell you. This here has arecallof: 10/(10+2) => 0.8333 => 83.33%.
I hope it helps...
Do point it out in case you find issues in this. I'm still learning all of this..

Top comments (0)