Early Access: The content on this website is provided for informational purposes only in connection with pre-General Availability Qlik Products.
All content is subject to change and is provided without warranty.
Skip to main content Skip to complementary content

Performing quick model analysis

When models finish training for an experiment version, perform analysis of the most important model metrics and access a summary of how the training data was processed. Quick analysis is performed in the Data and Models tabs.

Use the Models tab for an overview of your training results. When you run an experiment version, you are automatically switched to the Models tab. It is also recommended that you revisit the Data tab to inspect how the data was preprocessed and handled.

Models tab for a binary classification experiment

The 'Models' tab in an ML experiment

Additional options are available for further evaluation of models. You can compare additional metrics and hyperparameters between models, or focus on a specific model for granular analysis. These options are available in the Compare and Analyze tabs. For more information, see:

Understanding the concepts

It can be helpful to have a basic understanding of the concepts behind model analysis before you start evaluating your models. For more information, see Understanding model review concepts.

Impact of optimization settings on analysis

Your analysis experience can be slightly different depending on whether or not you have used intelligent model optimization. Intelligent model optimization is turned on by default for new experiments.

Analyzing models trained with intelligent optimization

By default, new experiments run with intelligent model optimization.

Intelligent model optimization provides a more robust training process that ideally creates a model that is ready to deploy with little to no further refinement. The performance of these models when deployed for production use cases is still dependent on training them with a high-quality dataset that includes relevant features and data.

If your version was trained with intelligent model optimization, consider the following:

  • Each model in the version can have different feature selection depending on how the algorithm analyzed the data.

  • From the Models tab, read the Model training summary for the model before div ing into specific analysis. The Model training summary shows a summary of how AutoML automatically optimized the model through iterating on feature selection and applying advanced transformations.

For more information about intelligent model optimization, see Intelligent model optimization.

Analyzing models trained without intelligent optimization

Alternatively, you might have turned off intelligent model optimization for the version of the training. Manual optimization of models can be helpful if you need more control over the training process.

If you used manual optimization, all models in the version will have the same feature selection, so a Model training summary is not needed.

Inspecting the configuration

During preprocessing, features might have been excluded from being used in the training. This typically happens because more information is known about the data as training progresses than before you run the version.

After reviewing the Model training summary (only shown with intelligent optimization), you can take a closer look at the experiment configuration if you need to check for these other changes.

  1. In the experiment, switch to the Data tab.

  2. Ensure you are in Table rows Schema view.

  3. Use the drop down menu in the toolbar to select a model from the version.

  4. Analyze the model schema. You might want to focus on the Insights and Feature type columns to see if certain features are dropped or have been transformed to a different feature type.

    For example, it is possible that a feature initially marked as Possible free text has been excluded after you ran the version.

    For more information about what each of the insights means, see Interpreting dataset insights.

Note that if you ran the version with the default intelligent optimization option, each model in the version could have different feature selection due to automatic refinement. If the version was run without intelligent optimization, the feature selection will be the same for all models in the version. For more information about intelligent model optimization, see Intelligent model optimization.

Based on what you find in this configuration, you might need to return to the dataset preparation stage to improve your feature data.

With intelligent model optimization turned off, you will not see a Model training summary as part of your quick analysis experience. Additionally, all the models from an experiment version will use the same feature combinations, while models trained with intelligent optimization can have different feature combinations.

Selecting a model

When the training completes, the top-performing model is automatically selected for analysis, and the metrics for that model are displayed. The top-performing model for each experiment version is marked with a Trophy icon.

Switch between models to compare the differences in metrics and configuration. If you are on the Data tab, use the drop down menu in the toolbar to pick a model. If you are on the Models tab, click a model in the Model metrics table to pick a model.

Analyzing the training summary

For models trained with intelligent model optimization, the Model training summary outlines key information about what happened during training. This summary includes:

  • The sampling ratio for the model.

  • A list of features that were automatically dropped as a result of intelligent optimization. This list varies depending on the select model.

  • A link to the training dataset.

The Model training summary is not shown if you are optimizing your models manually.

For more information about intelligent model optimization, see Intelligent model optimization.

Model training summary chart for a model, shown in the Models tab

Training summary chart for a model trained with intelligent optimization. Features from the training data were automatically excluded from the model for reasons such as target leakage and high correlation

The model training summary does not list features that were dropped during the preprocessing stage. To check for this information, return to the Data tab of the experiment. See: Inspecting the configuration

  1. In the experiment, switch to the Models tab.

  2. Confirm you have selected the model you want to analyze.

  3. In the Model insights section on the right side of the page, look at the Model training summary.

Analyzing the model metrics table

The Model metrics table provides high-level information about the performance of each model in the experiment. This table is shown in the Models tab.

When you are in the Models tab, the Model metrics table is also where you select a model to analyze.

Model metrics table

The 'Model metrics' table in an ML experiment

The model metrics table shows:

  • Whether the model is the top performer for its version

  • The name of each model (This can be edited for convenience. See: Editing model details)

  • The algorithm used to the train the model

  • Values for the most important model performance score, given the model type defined by the experiment target:

    • Binary classification models: F1.

      F1

    • Multiclass classification models: F1 Macro.

      Macro F1

    • Regression models: R2.

      R2

    For more information about what type of models your experiment trains, see Determining the type of model created.

Filtering models in the table

The model metrics table can be filtered to show only the models you want to compare.

The following filters are available, and can be combined as needed:

  • Version: Select one or more experiment versions.

  • Algorithm: Select one or more algorithms.

  • Top performers: Show top-performing models across all versions (marked with Trophy icons). Found under More model filters.

  • 100% sampling: This filter is applied by default. It can be removed if desired. The filter shows models that used the entire training dataset to complete the training and cross-validation process. Found under More model filters.

  • Deployed: Show models that have been deployed into ML deployments. Found under More model filters.

Interpreting feature importance visualizations

In the Models tab, scroll down to view the feature importance visualizations. These will be Permutation importance and SHAP importance.

Feature importance visualization in Models tab

Permutation and SHAP importance charts for quick model analysis

Permutation importance chart

The Permutation importance chart shows the importance of each feature to the overall prediction of the model. What you see in the Permutation importance chart can help you understand how to refine your models.

Common insights you can gain by analyzing this chart include:

  • If one feature is consuming almost all of the importance, this is likely a sign of target leakage. The feature needs to be removed. If this is caused by data quality issues, those need to be addressed as well.

    Data leakage

  • Some features might have very little to no impact on the model predictions. Features with extremely low permutation importance can be considered statistical noise and should be removed.

 If intelligent model optimization was used to train the version, some of the common issues noted above might have been automatically addressed by dropping these features.

For more information about permutation importance, see Understanding permutation importance.

SHAP importance chart

The SHAP importance chart provides another way to analyze the influence that each feature in the experiment is having on the predictions from each model. This can give you early insight into what features are significant, or if you need to re-configure the training. For more information, see Understanding SHAP importance in experiment training.

Interpreting model scoring visualizations

If your experiment is a binary classification problem, a number of additional visualizations are auto-generated for quick consumption. These visualizations give you more insight into how well the model is predicting the positive and negative classes.

For more information about these visualizations, see Scoring binary classification models.

Confusion matrix

The Confusion matrix chart shows the accuracy of the predictions created by the model. The predictions are performed on the automatic holdout data.

Confusion matrix

ROC curve

The ROC curve chart describes how good the model is at predicting the positive class when the actual outcome is positive.

For an indication of what an ideal ROC curve looks like, see AUC and ROC curve.

Did this page help you?

If you find any issues with this page or its content – a typo, a missing step, or a technical error – let us know how we can improve!