Binary Classification
How to log your model schema for binary classification models
Binary Classification Cases
prediction label, actual label
Accuracy, Recall, Precision, FPR, FNR, F1, Sensitivity, Specificity
prediction score, prediction label, actual label
AUC, PR-AUC, Log Loss, Accuracy, Recall, Precision, FPR, FNR, F1, Sensitivity, Specificity
Click here for all valid model types and metric combinations.
Case #1 - Supports Only Classification Metrics
Example Row
ca
True
12345
25
not_fraud
fraud
Code Example
For more details on Python Batch API Reference, visit here:
Pandas Batch LoggingCase #2 - Supports Classification & AUC/Log Loss Metrics
Example Row
ca
True
12345
25
not_fraud
fraud
0.3
Code Example
For more details on Python Batch API Reference, visit here:
Pandas Batch LoggingCase #3: Supports AUC & Log Loss Metrics
Example Row
ca
True
12345
25
fraud
0.3
Code Example
For more details on Python Pandas API Reference, visit here:
Pandas Batch LoggingDefault Actuals
For some use cases, it may be important to treat a prediction for which no corresponding actual label has been logged yet as having a default negative class actual label.
For example, consider tracking advertisement conversion rates for an ad clickthrough rate model, where the positive class is click
and the negative class is no_click
. For ad conversion purposes, a prediction without a corresponding actual label for an ad placement is equivalent to logging an explicit no_click
actual label for the prediction. In both cases, the result is the same: a user has not converted by clicking on the ad.
For AUC-ROC, PR-AUC, and Log Loss performance metrics, Arize supports treating predictions without an explicit actual label as having the negative class actual label by default. In the above example, a click
prediction without an actual would be treated as a false positive, because the missing actual for the prediction would, by default, be assigned to the no_click
negative class.
This feature can be enabled for monitors and dashboards via the model performance config section of your model's config page.
Quick Definitions
Prediction Label: The classification label of this event (Cardinality = 2)
Actual Label: The ground truth label (Cardinality = 2)
Prediction Score: The likelihood of the event (Probability between 0 to 1)
Actual Score: The ground truth score (0 or 1)
Last updated
Was this helpful?