You made a machine learning or deep learning model. Amazing! π₯³

But how do you check its performance and robustness? Simply building a predictive model is not enough. You have to create a model which gives high accuracy on **out-of-sample data**.

That's where evaluation metrics come into the picture. They are used to measure the quality of the statistical or machine learning model, and this article will introduce you to the most common and yet important metrics to evaluate your machine learning algorithms.

If you're someone who just started out in the field of Data Science and Machine Learning, this is for you. We will discuss the following terms:

- Confusion matrix
- Accuracy
- Precision
- Recall
- F1 score
- Mean Absolute Error
- R Squared Score
- Root Mean Squared Error
- ROC (Receiver Operating Characteristics) curve

So letβs dive into it to learn about them :

## Confusion Matrix

A confusion matrix is one of the simplest concepts in the field of machine learning, and yet so important when it comes to statistical classification problems.

A confusion matrix is a N X N matrix, where N is the number of classes being predicted. It is usually used to **describe the performance of a classification model based on the true values known for a data set**.

In simple words, it is a summary of the prediction result made by the classification model.

### Terms Associated with Confusion Matrix

**True Positive (TP):** True positive means their actual class value is 1 and the predicted value is also 1.

For example, The case where the woman is actually pregnant and the model also classifies that she is pregnant.

**False Positive (FP):** False positive means their actual class value is 0 and the predicted value is 1.

For example, The case where the woman is not pregnant and the model classifies that she is pregnant.

**True Negative (TN):** True negative means their actual class value is 0 and the predicted value is also 0.

For example, The case where the woman is actually not pregnant and the model also classifies that she is not pregnant.

**False Negative (FN):** False-negative means their actual class value is 1 and the predicted value is 0.

For example, The case where the woman is actually pregnant and the model also classifies that she is not pregnant.

The confusion matrix in itself is not a performance measure but all the performance metrics are dependent on the confusion matrix.

## Accuracy

Accuracy is the most common metric that you can see everywhere when you're evaluating your model. It is simply defined as the number of correct predictions made by your model.

In simple terms, **accuracy is the ratio of all the correct predictions to the total number of predictions**.

It can be calculated using a confusion matrix.

## Precision

Precision is defined as **the total number of correctly classified positive examples by the total number of predicted positive examples**.

In some cases, your classification model might classify based on the most frequent classes. Which in turn will bring a low accuracy because your model didnβt learn anything and just classified based on the top class.

Therefore, we need class-specific performance metrics to analyze. Precision is one of them.

It depicts how much the model is right when it says it is right.

## Recall

A recall (also known as *sensitivity*) refers to the percentage of total relevant results correctly classified by the classification model.

It is the **the number of positive samples returned by the custom-trained model**.

The relation between recall and precision -

### High recall, low precision:

This means, most of the positive examples are correctly recognized but there are a lot of false positives.

### Low recall, high precision:

This means that we missed a lot of positive examples but those we predicted as positive are indeed positive.

## F1 Score

Based on the use case of the classification model, the priority is given either to precision or recall, but in some classification models, we need both of these metrics to be combined as a single one.

**F1-Score is a metric that combines both precision and recall** and has an equal and relative contribution of both precision and recall.

It is the harmonic mean of precision and recall.

Remember that if your use case needs either recall or precision, one higher than the other then F1-score may not be the good metric for it.

## Mean Absolute Error

Mean absolute error(MAE) is defined as **the average of all errors that are calculated based on values predicted by your model**.

It is intended to measure average model bias in a set of predictions, without considering their direction

## R Squared Score

R squared is a measure of how close the data are to the fitted regression line. It is also known as the coefficient of determination.

It defines the degree to which the variance in the dependent variable (or target) can be explained by the independent variable (features).

For example if the R-squared value for our predictive model is 0.8. This means that 80% of the variation in the dependent variable is explained by the independent variables.

Therefore we can say that the **higher the r-squared value is, the better is the model**.

SSRES: the Residual sum of squared errors

SSTOT: the total sum of squared errors

## Root Mean Squared Error

RMSE is one of the most popular metrics used today for evaluating regression-based models. This is an important evaluation metric since itβs essential to find the average squared error between the predicted values.

RMSE **measures the average magnitude of the error**. Itβs the square root of the average of squared differences between prediction and actual observation.

RMSE is highly affected by outlier values. Hence, make sure youβve removed outliers from your data set prior to using this metric.

As compared to mean absolute error, RMSE gives higher weightage and punishes large errors

here 'N' is total observations.

### Correlation between MAE and RMSE

Both of these metrics express the average error of the machine learning models.

These two metrics can range from 0 to infinity and both of these metrics are negatively oriented scores, which means that a lower score defines better results.

## ROC Curve

ROC curve stands for βReceiver Operating Characteristic" Curve.

ROC curve is a graph showing the performance of a classification model at all its cut-off threshold.

In simple words, ROC Curve is the one that **tells how much your model is capable of differentiating among the different classes**.

ROC is a probability curve. It is a representation of the performance of your model in a graphical manner.

This curve represents:

- True positive rate (recall/sensitivity)
- False-positive rate (1- specificity) : FPR = FP/ FP+TN

The curve separates the space into two areas, one for good and the other for poor performance levels.

# Conclusion

In this article, you have seen some important metrics that are helpful for model evaluation and are needed to be understood by a data scientist or a machine learning engineer.

I hope you have found it useful. Also please feel free to suggest corrections and improvements.

Thanks for reading :))

### Want to start with web development? Checkout βΆ HTML To React: The Ultimate Guide

This ebook is a comprehensive guide that teaches you **everything you need to know to be a web developer through a ton of easy-to-understand examples and proven roadmaps**

It contains π

β Straight to the point explanations

β Simple code examples

β 50+ Interesting project ideas

β 3 Checklists of secret resources

β A Bonus Interview prep

You can even check out a free sample from this book

and here's the link to the complete book, if you want to buy it for yourself

#### Original price is $ 40 but price with this link - $16(which is 60% less than the original price)π

π

## Top comments (0)