Two classification indicators Precision, Recall, Accuracy, F score, TPR, FPR, MCC

  1. Ture Positive Rate (TPR): T P T P + F N \frac{TP}{TP+FN} TP+FNTP, to examine the proportion of correct predictions of the classifier in GT-positive samples.
  2. False Positive Rate (FPR): F P F P + T N \frac{FP}{FP+TN} FP+TNFP, to examine the proportion of classifiers predicting errors in GT-negative samples.
  3. AUC: The area under the ROC curve, which comprehensively examines the performance of the classifier.
  4. Percision: T P T P + F P \frac{TP}{TP+FP} TP+FPTP, to examine the proportion of correct predictions among the samples predicted by the classifier as positive.
  5. Recall: T P T P + F N \frac{TP}{TP+FN} TP+FNTP, same as TPR.
  6. Average Precision (AP): The area under the PR curve, which comprehensively examines the performance of the classifier.
  7. ACC: T P + T N T P + F P + F N + T N \frac{TP+TN}{TP+FP+FN+TN} TP+FP+FN+TNTP+TN, to examine the proportion of correct predictions of the classifier in all samples.
  8. F1 score: 2 1  Precision  + 1  Recall  \frac{2}{\frac{1}{\text { Precision }}+\frac{1}{\text { Recall }}}  Precision 1+ Recall 12, to calculate the harmonic mean of the two indicators of Precision and Recall, taking into account the performance of the classifier on these two indicators.
  9. F β \betaβ score: 1 + β 2 β 2 1  Recall  + 1  Precision  \frac{1+\beta^{2}}{\beta^{2} \frac{1}{\text { Recall }}+\frac{1}{\text { Precision }}} b2 Recall 1+ Precision 11 + b2, compared to F1 score, F β \betaThe β score can adjust the weight to make the indicator pay more attention to the performance of the classifier on Recall or Precision (for example, F2 score pays more attention to Recall).
  10. Matthews correlation coefficient (MCC): T P × T N − F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N ) \frac{TP \times TN-FP \times FN}{\sqrt{(TP+FP)(TP+FN)(TN+F P)(TN+FN)}} (TP+FP)(TP+FN)(TN+FP)(TN+FN) TP×TNFP×FN​, considering the four situations of TP, FP, FN, and TN comprehensively, this indicator can be used even when the sample content of the two categories is very different. (I didn't understand this indicator very well, so I found the meaning of this indicator from the Internet)

Guess you like

Origin blog.csdn.net/fuss1207/article/details/126841669