As I said in the last topic, I’m going to provide some topics in machine learning that all of you must know but it’s not provided in this specialization.
When dealing with logistic regression, there is a concept that is used widely that tells how much the model is capable of distinguishing between classes, which is the ROC curve, so what is it?!
An ROC curve (receiver operating characteristic curve) is a graph showing the performance of a classification model at all classification thresholds. This curve plots two parameters:
- True Positive Rate
- False Positive Rate
True Positive Rate (TPR) is a synonym for recall and is therefore defined as follows:
TPR=TPTP+FN
False Positive Rate(FPR) is defined as follows:
FPR=FPFP+TN
An ROC curve plots TPR vs. FPR at different classification thresholds. Lowering the classification threshold classifies more items as positive, thus increasing both False Positives and True Positives. The following figure shows a typical ROC curve.
AUC: Area Under the ROC Curve
AUC stands for “Area under the ROC Curve.” That is, AUC measures the entire two-dimensional area underneath the entire ROC curve (think integral calculus) from (0,0) to (1,1).
AUC provides an aggregate measure of performance across all possible classification thresholds. One way of interpreting AUC is as the probability that the model ranks a random positive example more highly than a random negative example.
For more details about it check out these links: