Skip to main content

Why Evaluation Metrics Matters


This is a follow up article on "The Importance of F1 Score" in which we understood the technical aspects of evaluating a Machine Learning model. In this article we will understand how different evaluation metrics can help us in designing solutions based on the problem statement and domain.

I will like to distinguish evaluation metrics with respect to the following criteria:

When there is life involved:

In case of aircraft or ships, where a Machine Learning algorithm misses to detect a failure can be a costly affair, since there is life involved. Also in the case of detecting cancer, a failure to predict a true positive can lead to delay in treatment, and hence can be life threatening.

In cases such as detecting failures in aircrafts, even if we get false positives is acceptable, but not able to detect true positives can be too expensive in terms of loss of life. Thus it becomes important to use Recall as a measure, as any positive detection becomes important, even if it turns out to be false positive. It is reasonable to waste resources on analyzing a false positive in such cases.

When there is cost involved:

In cases where the cost of responding to a false positive is too high, and where there are no lives involved, it is acceptable even if few failures are not detected. Take for example, a power plant is located in a village. Responding to a false positive from a sensor can be costly, as the transportation cost to check whether the fault has really occurred is too high. In such a scenario it becomes important to consider Precision as a measure. Thus the model should be trained for high precision, as the quality of the detection by the Machine Learning algorithm matters.


When two or more models need to be compared:

In cases where we need to perform comparison between two or more models going with a single metric like Precision or Recall may not be right. In such cases it is advisable to consider the harmonic mean of both the metrics, and hence F1 score matters in such scenarios. The model with high F1 score should be considered, except when the first two criteria are involved.

Understanding evaluation metrics can help you in deciding what model to choose, which metric matters for which domain. It will also help you to interact with your client and understand different Machine Learning algorithms better. Also it gives you the ability of understanding a problem statement well, thus helping you to implement the correct Machine Learning solution.

Evaluation metrics should be the main area of focus while designing Machine Learning solutions.


By,
Siddhesh Wagle,
Research Consultant,
Cere Labs





 

Comments

  1. Slot machines at lainborough casino - Dr. McD
    The 여주 출장마사지 casino has over 30 slot machines and 20 table games that 전라북도 출장안마 include video poker, blackjack, craps, 평택 출장안마 roulette, blackjack 삼척 출장안마 and 여수 출장샵 video

    ReplyDelete

Post a Comment

Popular posts from this blog

GPU - The brain of Artificial Intelligence

Machine Learning algorithms require tens and thousands of CPU based servers to train a model, which turns out to be an expensive activity. Machine Learning researchers and engineers are often faced with the problem of running their algorithms fast. Although initially invented for processing graphics in computer games, GPUs today are used in machine learning to perform feature detection from vast amount of unlabeled data. Compared to CPUs, GPUs take far less time to train models that perform classification and prediction. Characteristics of GPUs that make them ideal for machine learning Handle large datasets Needs far less data centre infrastructure Can be specialized for specific machine learning needs Perform vector computations faster than any known processor Designed to perform data parallel computation NVIDIA CUDA GPUs today are used to build deep learning image processing tools for  Adobe Creative Cloud. According to NVIDIA blog future Adobe applicati

Understanding Projection Pursuit Regression

The following article gives an overview of the paper "Projection Pursuit Regression” published by Friedman J. H and Stuetzle W. You will need basic background of Machine Learning and Regression before understanding this article. The algorithms and images are taken from the paper. ( http://www.stat.washington.edu/courses/stat527/s13/readings/FriedmanStuetzle_JASA_1981.pdf )  What is Regression? Regression is a machine learning technology used to predict a response variable given multiple predictor variables or features. The main distinction is that the response to be predicted is any real value and not just any class or cluster name. Hence though similar to Classification in terms of making a prediction, it is largely different given what it’s predicting.  A simple to understand real world problem of regression would be predicting the sale price of a particular house based on it’s square footage, given that we have data of similar houses sold in that area in the past. T

Anomaly Detection based on Prediction - A Step Closer to General Artificial Intelligence

Anomaly detection refers to the problem of finding patterns that do not conform to expected behavior [1]. In the last article "Understanding Neocortex to Create Intelligence" , we explored how applications based on the workings of neocortex create intelligence. Pattern recognition along with prediction makes human brains the ultimate intelligent machines. Prediction help humans to detect anomalies in the environment. Before every action is taken, neocortex predicts the outcome. If there is a deviation from the expected outcome, neocortex detects anomalies, and will take necessary steps to handle them. A system which claims to be intelligent, should have anomaly detection in place. Recent findings using research on neocortex have made it possible to create applications that does anomaly detection. Numenta’s NuPIC using Hierarchical Temporal Memory (HTM) framework is able to do inference and prediction, and hence anomaly detection. HTM accurately predicts anomalies in real