Categorical Target Drift

TL;DR: The report explores the changes in the categorical target function (prediction).

  • Performs a suitable statistical test to compare target (prediction) distribution

  • Plots the relations between each individual feature and the target (prediction)


The Target Drift report helps detect and explore changes in the target function and/or model predictions.

The Categorical Target Drift report is suitable for problem statements with the categorical target function: binary classification, multi-class classification, etc.


To run this report, you need to have input features, and target and/or prediction columns available.

You will need two datasets. The reference dataset serves as a benchmark. We analyze the change by comparing the current production data to the reference data.

You can potentially choose any two datasets for comparison. But keep in mind that only the reference dataset will be used as a basis for comparison.

How it works

We estimate the drift for the target (actual values) and predictions in the same manner. If both columns are passed to the dashboard, we build two sets of plots.

If only one of them (either target or predictions) is provided, we build one set of plots. If neither target nor predictions column is available, you will get an error.

To estimate the categorical target (prediction) drift, we compare the distribution of the target (prediction) in the two datasets. We use the chi-squared test with 0.95 confidence level to detect if the distribution has changed significantly.

How it looks

The report includes 2 components. All plots are interactive.

1. Target (Prediction) Drift

The report first shows the comparison of target (prediction) distributions in the current and reference dataset. The result of the statistical test and P-value are displayed in the title.

For a classification problem with three classes, it can look like this (an example of the extreme target drift with the appearance of a new class):

2. Target (Prediction) Behavior By Feature

The report generates an interactive table with the visualizations of dependencies between the target and each feature.

If you click on any feature, you get a plot that shows the feature distribution for the different target labels.

These plots help analyze how feature values relate to the target labels and identify the differences between the datasets.

We recommend paying attention to the behavior of the most important features since significant changes might confuse the model and cause higher errors.

When to use the report

Here are our suggestions on when to use it—best combined with the Data Drift report. 1. Before model retraining. Before feeding fresh data into the model, you might want to verify whether it even makes sense. 2. When you are debugging the model decay. If you observe a drop in performance, this report can help see what has changed. 3. When you are flying blind, and no ground truth is available. If you do not have immediate feedback, you can use this report to explore the changes in the model output and the relationship between the features and prediction. This can help anticipate data and concept drift.

JSON Profile

If you choose to generate a JSON profile, it will contain the following information:

cat_target_drift": {
"name": "cat_target_drift",
"datetime": "datetime",
"data": {
"utility_columns": {
"date": null,
"id": null,
"target": "target",
"prediction": null
"cat_feature_names": [],
"num_feature_names": [],
"metrics": {
"target_name": "target",
"target_type": "cat",
"target_drift": p_value
"timestamp": "timestamp"


  • See Iris Data Drift, Categorical Target Drift, and Classification Performance report generation to explore the reports both inside a Jupyter notebook and as a separate HTML file: Jupyter notebook

You can also read the initial release blog.