Precision, Recall and F1-Score

Precision, Recall, and F1-Score are important evaluation metrics used for classification problems in Machine Learning. These metrics provide deeper insight into model performance than accuracy, especially when dealing with imbalanced datasets.

They are commonly calculated using values from the Confusion Matrix:

  • True Positive (TP)
  • True Negative (TN)
  • False Positive (FP)
  • False Negative (FN)

Why These Metrics are Important

These metrics help:

  • Evaluate classification performance
  • Analyze prediction quality
  • Handle imbalanced datasets
  • Understand false predictions
  • Compare classification models

Note: Accuracy alone may be misleading for imbalanced datasets, so Precision, Recall, and F1-Score are widely used.

1. Precision

Precision measures how many predicted positive values are actually correct.

Formula

Precision= TP / (TP + FP)

It means: 

Out of all positive predictions:

  • How many were truly positive?

Suppose:

  • Model predicts 100 emails as spam
  • Only 80 are actually spam

Then:

Precision = 80 / 100 = 0.8

Precision = 80%

High Precision Means

  • Fewer False Positives
  • More reliable positive predictions

When Precision is Important ?

Precision is important when:

  • False Positives are costly

Example

Email Spam Detection

Incorrectly marking important emails as spam is problematic.

2. Recall

Recall measures how many actual positive values are correctly identified by the model.

Formula

Recall=TP/ (TP+FN) 

It means: 

Out of all actual positive cases:

  • How many did the model correctly detect?

Suppose:

  • 100 patients actually have a disease
  • Model correctly detects 90 patients

Then:

Recall = 90 / 100 = 0.9

Recall = 90%

High Recall Means

  • Fewer False Negatives
  • Better detection of positive cases

When Recall is Important

Recall is important when:

  • Missing positive cases is dangerous

Example

Cancer Detection

Failing to detect cancer can be life-threatening.

3. F1-Score

F1-Score is the harmonic mean of Precision and Recall.

It balances both metrics into a single value.

Formula

F1=2× (Precision×Recall) / (Precision+Recall)

Why F1-Score is Useful

Sometimes:

  • Precision is high
  • Recall is low

or vice versa.

F1-Score provides a balanced evaluation.

Suppose:

Precision = 0.8
Recall = 0.6

Then:

F1-Score ≈ 0.69

Relationship Between Precision and Recall

Metric Focus
Precision Reduce False Positives
Recall Reduce False Negatives

Trade-Off Between Precision and Recall

Improving Precision may reduce Recall.

Improving Recall may reduce Precision.

Balancing both is important.

Python Example

from sklearn.metrics import (
    precision_score,
    recall_score,
    f1_score
)

# Actual values
y_true = [1, 1, 1, 0, 0]

# Predicted values
y_pred = [1, 1, 0, 0, 1]

# Precision
print("Precision:",
      precision_score(y_true, y_pred))

# Recall
print("Recall:",
      recall_score(y_true, y_pred))

# F1 Score
print("F1-Score:",
      f1_score(y_true, y_pred))

Expected Output 

Precision: 0.67
Recall: 0.67
F1-Score: 0.67

Understanding with Confusion Matrix

Suppose:

Actual / Predicted Positive Negative
Positive TP = 80 FN = 20
Negative FP = 10 TN = 90

Precision

Precision=80 / (80+10) =0.89

Recall

Recall= 80 / (80+20) =0.80

F1-Score

F1=2 × (0.89×0.80) / (0.89+0.80) ≈0.84

Real-World Applications

Application Important Metric
Spam Detection Precision
Disease Detection Recall
Fraud Detection F1-Score

Benefits of the metrics

  • Better model evaluation
  • Useful for imbalanced datasets
  • Deeper understanding of predictions
  • Improved classification analysis

Important Points

1. Precision measures correctness of positive predictions.

2. Recall measures ability to detect actual positive cases.

3. F1-Score balances Precision and Recall.

4. Precision focuses on reducing False Positives.

5. Recall focuses on reducing False Negatives.

Summary

Precision, Recall, and F1-Score are important classification evaluation metrics used to analyze machine learning model performance beyond accuracy. They help measure prediction quality, detect false predictions, and evaluate models effectively, especially for imbalanced datasets.

Keywords

Precision, Recall, F1-Score, Precision Recall F1 Score, Classification Metrics, Machine Learning Evaluation Metrics, Confusion Matrix Metrics, True Positive, False Positive, False Negative, Precision Formula, Recall Formula, F1 Score Formula, Binary Classification Metrics, Imbalanced Dataset Evaluation, Classification Performance Metrics, Precision vs Recall

Previous Topic Confusion Matrix