The Area Under the Receiver Operating Characteristic Curve (AUC-ROC) is a performance measurement for classification problems. The ROC curve is a plot of true positive rate (recall) against the false positive rate (1 - specificity). The AUC represents the degree or measure of separability, and it tells how much the model is capable of distinguishing between classes. The higher the AUC, the better the model. It is defined as:AUC=∫01TPR(f)dfwhere:
Cross-Entropy Loss is a measure used to quantify the difference between two probability distributions for a given random variable or set of events. It is defined as:H(y,y^)=−i∑yilog(y^i)
The Explained Variance Score measures the proportion to which a mathematical model accounts for the variation (dispersion) of a given data set. It is defined as:Explained Variance=1−Var(y)Var(y−y^)where:
(\text(y - \hat)) is the variance of the errors,
(\text(y)) is the variance of the actual values.
An Explained Variance Score close to 1 indicates that the model explains a large portion of the variance in the data.
The F1 score is a harmonic mean of precision and recall, providing a balance between them. It is calculated as:F1=2⋅precision+recallprecision⋅recallwhere:
precisionrecall
Mean Average Precision (mAP) is a common metric used to evaluate object detection models. It summarizes the precision-recall curve for different classes. It is calculated as:mAP=N1i=1∑NAPiwhere:
(N) is the number of classes,
(AP_i) is the Average Precision for class (i).
Average Precision (AP) is computed as the area under the precision-recall curve for a specific class. It integrates the precision over all recall values from 0 to 1. For object detection, AP can be calculated at different IoU thresholds to provide a comprehensive evaluation.In addition to the standard mAP, specific metrics include:
Mean Absolute Error (MAE) measures the average magnitude of the errors in a set of predictions, without considering their direction. It is calculated as:MAE=n1∑(i=1)n∣yi−y^i∣where:
Mean Squared Error (MSE) measures the average of the squares of the errors—that is, the average squared difference between the estimated values and the actual value. It is calculated as:MSE=n1∑(i=1)n(yi−y^i)2where:
The Softmax function is used for multi-class classification. It converts logits to probabilities that sum to 1. It is defined for class (j) as:σ(z)j=∑(k=1)Kezkezjforj=1,...,K
Weighted Average F1 Score takes into account the F1 score of each class and the number of instances for each class. It is defined as:WeightedAverageF1Score=∑(i=1)n(TP+FNTPi+FNi⋅F1i)where:
Weighted Average Precision takes into account the precision of each class and the number of instances for each class. It is defined as:WeightedAveragePrecision=∑(i=1)n(TP+FNTPi+FNi⋅Precisioni)where:
Weighted Average Recall takes into account the recall of each class and the number of instances for each class. It is defined as:WeightedAverageRecall=∑(i=1)n(TP+FNTPi+FNi⋅Recalli)where: