Use the Responsible AI dashboard in Azure Machine Learning studio

Responsible AI dashboards are linked to your registered models. To view your Responsible AI dashboard, go into your model registry and select the registered model you've generated a Responsible AI dashboard for. Then, select the Responsible AI tab to view a list of generated dashboards.

Screenshot of the model details pane in Azure Machine Learning studio, with the 'Responsible AI' tab highlighted.

You can configure multiple dashboards and attach them to your registered model. Various combinations of components (interpretability, error analysis, causal analysis, and so on) can be attached to each Responsible AI dashboard. The following image displays a dashboard's customization and the components that were generated within it. In each dashboard, you can view or hide various components within the dashboard UI itself.

Screenshot of Responsible AI tab with a dashboard name highlighted.

Select the name of the dashboard to open it into a full view in your browser. To return to your list of dashboards, you can select Back to models details at any time.

Screenshot of a Responsible AI dashboard with the 'Back to model details' button highlighted.

Full functionality with integrated compute resource

Some features of the Responsible AI dashboard require dynamic, on-the-fly, and real-time computation (for example, what-if analysis). Unless you connect a compute resource to the dashboard, you might find some functionality missing. When you connect to a compute resource, you enable full functionality of your Responsible AI dashboard for the following components:

  • Error analysis
    • Setting your global data cohort to any cohort of interest will update the error tree instead of disabling it.
    • Selecting other error or performance metrics is supported.
    • Selecting any subset of features for training the error tree map is supported.
    • Changing the minimum number of samples required per leaf node and error tree depth is supported.
    • Dynamically updating the heat map for up to two features is supported.
  • Feature importance
    • An individual conditional expectation (ICE) plot in the individual feature importance tab is supported.
  • Counterfactual what-if
    • Generating a new what-if counterfactual data point to understand the minimum change required for a desired outcome is supported.
  • Causal analysis
    • Selecting any individual data point, perturbing its treatment features, and seeing the expected causal outcome of causal what-if is supported (only for regression machine learning scenarios).

You can also find this information on the Responsible AI dashboard page by selecting the Information icon, as shown in the following image:

Screenshot of the 'Information' icon on the Responsible AI dashboard.

Enable full functionality of the Responsible AI dashboard

  1. Select a running compute instance in the Compute dropdown list at the top of the dashboard. If you don’t have a running compute, create a new compute instance by selecting the plus sign (+) next to the dropdown. Or you can select the Start compute button to start a stopped compute instance. Creating or starting a compute instance might take few minutes.

    Screenshot of the 'Compute' dropdown box for selecting a running compute instance.

  2. When a compute is in a Running state, your Responsible AI dashboard starts to connect to the compute instance. To achieve this, a terminal process is created on the selected compute instance, and a Responsible AI endpoint is started on the terminal. Select View terminal outputs to view the current terminal process.

    Screenshot showing that the responsible AI dashboard is connecting to a compute resource.

  3. When your Responsible AI dashboard is connected to the compute instance, you'll see a green message bar, and the dashboard is now fully functional.

    Screenshot showing that the dashboard is connected to the compute instance.

  4. If the process takes a while and your Responsible AI dashboard is still not connected to the compute instance, or a red error message bar is displayed, it means there are issues with starting your Responsible AI endpoint. Select View terminal outputs and scroll down to the bottom to view the error message.

    Screenshot of an error connecting to a compute.

    If you're having difficulty figuring out how to resolve the "failed to connect to compute instance" issue, select the Smile icon at the upper right. Submit feedback to us about any error or issue you encounter. You can include a screenshot and your email address in the feedback form.

UI overview of the Responsible AI dashboard

The Responsible AI dashboard includes a robust, rich set of visualizations and functionality to help you analyze your machine learning model or make data-driven business decisions:

Global controls

At the top of the dashboard, you can create cohorts (subgroups of data points that share specified characteristics) to focus your analysis of each component. The name of the cohort that's currently applied to the dashboard is always shown at the top left of your dashboard. The default view in your dashboard is your whole dataset, titled All data (default).

Screenshot of a responsible AI dashboard showing all data.

  1. Cohort settings: Allows you to view and modify the details of each cohort in a side panel.
  2. Dashboard configuration: Allows you to view and modify the layout of the overall dashboard in a side panel.
  3. Switch cohort: Allows you to select a different cohort and view its statistics in a pop-up window.
  4. New cohort: Allows you to create and add a new cohort to your dashboard.

Select Cohort settings to open a panel with a list of your cohorts, where you can create, edit, duplicate, or delete them.

Screenshot showing the cohort settings on the dashboard.

Select New cohort at the top of the dashboard or in the Cohort settings to open a new panel with options to filter on the following:

  1. Index: Filters by the position of the data point in the full dataset.
  2. Dataset: Filters by the value of a particular feature in the dataset.
  3. Predicted Y: Filters by the prediction made by the model.
  4. True Y: Filters by the actual value of the target feature.
  5. Error (regression): Filters by error (or Classification Outcome (classification): Filters by type and accuracy of classification).
  6. Categorical Values: Filter by a list of values that should be included.
  7. Numerical Values: Filter by a Boolean operation over the values (for example, select data points where age < 64).

Screenshot of making multiple new cohorts.

You can name your new dataset cohort, select Add filter to add each filter you want to use, and then do either of the following:

  • Select Save to save the new cohort to your cohort list.
  • Select Save and switch to save and immediately switch the global cohort of the dashboard to the newly created cohort.

Screenshot of making a new cohort in the dashboard.

Select Dashboard configuration to open a panel with a list of the components you’ve configured on your dashboard. You can hide components on your dashboard by selecting the Trash icon, as shown in the following image:

Screenshot showing the dashboard configuration.

You can add components back to your dashboard via the blue circular plus sign (+) icon in the divider between each component, as shown in the following image:

Screenshot of adding a component to the dashboard.

Error analysis

The next sections cover how to interpret and use error tree maps and heat maps.

Error tree map

The first pane of the error analysis component is a tree map, which illustrates how model failure is distributed across various cohorts with a tree visualization. Select any node to see the prediction path on your features where an error was found.

Screenshot of the dashboard showing an error analysis on the tree map pane.

  1. Heat map view: Switches to heat map visualization of error distribution.
  2. Feature list: Allows you to modify the features used in the heat map using a side panel.
  3. Error coverage: Displays the percentage of all error in the dataset concentrated in the selected node.
  4. Error (regression) or Error rate (classification): Displays the error or percentage of failures of all the data points in the selected node.
  5. Node: Represents a cohort of the dataset, potentially with filters applied, and the number of errors out of the total number of data points in the cohort.
  6. Fill line: Visualizes the distribution of data points into child cohorts based on filters, with the number of data points represented through line thickness.
  7. Selection information: Contains information about the selected node in a side panel.
  8. Save as a new cohort: Creates a new cohort with the specified filters.
  9. Instances in the base cohort: Displays the total number of points in the entire dataset and the number of correctly and incorrectly predicted points.
  10. Instances in the selected cohort: Displays the total number of points in the selected node and the number of correctly and incorrectly predicted points.
  11. Prediction path (filters): Lists the filters placed over the full dataset to create this smaller cohort.

Select the Feature list button to open a side panel, from which you can retrain the error tree on specific features.

Screenshot of the dashboard side panel, which lists selectable features of an error analysis tree map.

  1. Search features: Allows you to find specific features in the dataset.
  2. Features: Lists the name of the feature in the dataset.
  3. Importances: A guideline for how related the feature might be to the error. Calculated via mutual information score between the feature and the error on the labels. You can use this score to help you decide which features to choose in the error analysis.
  4. Check mark: Allows you to add or remove the feature from the tree map.
  5. Maximum depth: The maximum depth of the surrogate tree trained on errors.
  6. Number of leaves: The number of leaves of the surrogate tree trained on errors.
  7. Minimum number of samples in one leaf: The minimum amount of data required to create one leaf.

Error heat map

Select the Heat map tab to switch to a different view of the error in the dataset. You can select one or many heat map cells and create new cohorts. You can choose up to two features to create a heat map.

Screenshot of the dashboard, showing an error analysis heat map and list of features to compare.

  1. Cells: Displays the number of cells selected.
  2. Error coverage: Displays the percentage of all errors concentrated in the selected cell(s).
  3. Error rate: Displays the percentage of failures of all data points in the selected cell(s).
  4. Axis features: Selects the intersection of features to display in the heat map.
  5. Cells: Represents a cohort of the dataset, with filters applied, and the percentage of errors out of the total number of data points in the cohort. A blue outline indicates selected cells, and the darkness of red represents the concentration of failures.
  6. Prediction path (filters): Lists the filters placed over the full dataset for each selected cohort.

Model overview and fairness metrics

The model overview component provides a comprehensive set of performance and fairness metrics for evaluating your model, along with key performance disparity metrics along specified features and dataset cohorts.

Dataset cohorts

On the Dataset cohorts pane, you can investigate your model by comparing the model performance of various user-specified dataset cohorts (accessible via the Cohort settings icon at the top right of the dashboard).

Screenshot of the 'Model overview' pane, showing the 'Dataset cohorts' tab.

  1. Help me choose metrics: Select this icon to open a panel with more information about what model performance metrics are available to be shown in the table. Easily adjust which metrics to view by using the multi-select dropdown list to select and deselect performance metrics.
  2. Show heat map: Toggle on and off to show or hide heat map visualization in the table. The gradient of the heat map corresponds to the range normalized between the lowest value and the highest value in each column.
  3. Table of metrics for each dataset cohort: View columns of dataset cohorts, the sample size of each cohort, and the selected model performance metrics for each cohort.
  4. Bar chart visualizing individual metric: View mean absolute error across the cohorts for easy comparison.
  5. Choose metric (x-axis): Select this button to choose which metrics to view in the bar chart.
  6. Choose cohorts (y-axis): Select this button to choose which cohorts to view in the bar chart. Feature cohort selection might be disabled unless you first specify the features you want on the Feature cohort tab of the component.

Select Help me choose metrics to open a panel with a list of model performance metrics and their definitions, which can help you select the right metrics to view.

Machine learning scenario Metrics
Regression Mean absolute error, Mean squared error, R-squared, Mean prediction.
Classification Accuracy, Precision, Recall, F1 score, False positive rate, False negative rate, Selection rate.

Feature cohorts

On the Feature cohorts pane, you can investigate your model by comparing model performance across user-specified sensitive and non-sensitive features (for example, performance across various gender, race, and income level cohorts).

Screenshot of the dashboard 'Model overview' pane, showing the 'Feature cohorts' tab.

  1. Help me choose metrics: Select this icon to open a panel with more information about what metrics are available to be shown in the table. Easily adjust which metrics to view by using the multi-select dropdown to select and deselect performance metrics.

  2. Help me choose features: Select this icon to open a panel with more information about what features are available to be shown in the table, with descriptors of each feature and their binning capability (see below). Easily adjust which features to view by using the multi-select dropdown to select and deselect them.

    Screenshot of the dashboard 'Model overview' pane, showing how to choose features.

  3. Show heat map: Toggle on and off to see a heat map visualization. The gradient of the heat map corresponds to the range that's normalized between the lowest value and the highest value in each column.

  4. Table of metrics for each feature cohort: A table with columns for feature cohorts (sub-cohort of your selected feature), sample size of each cohort, and the selected model performance metrics for each feature cohort.

  5. Fairness metrics/disparity metrics: A table that corresponds to the metrics table and shows the maximum difference or maximum ratio in performance scores between any two feature cohorts.

  6. Bar chart visualizing individual metric: View mean absolute error across the cohorts for easy comparison.

  7. Choose cohorts (y-axis): Select this button to choose which cohorts to view in the bar chart.

    Selecting Choose cohorts opens a panel with an option to either show a comparison of selected dataset cohorts or feature cohorts, depending on what you select in the multi-select dropdown list below it. Select Confirm to save the changes to the bar chart view.

    Screenshot of the dashboard 'Model overview' pane, showing how to choose cohorts.

  8. Choose metric (x-axis): Select this button to choose which metric to view in the bar chart.

Data analysis

With the data analysis component, the Table view pane shows you a table view of your dataset for all features and rows.

The Chart view panel shows you aggregate and individual plots of datapoints. You can analyze data statistics along the x-axis and y-axis by using filters such as predicted outcome, dataset features, and error groups. This view helps you understand overrepresentation and underrepresentation in your dataset.

Screenshot of the dashboard, showing the data analysis.

  1. Select a dataset cohort to explore: Specify which dataset cohort from your list of cohorts you want to view data statistics for.

  2. X-axis: Displays the type of value being plotted horizontally. Modify the values by selecting the button to open a side panel.

  3. Y-axis: Displays the type of value being plotted vertically. Modify the values by selecting the button to open a side panel.

  4. Chart type: Specifies the chart type. Choose between aggregate plots (bar charts) or individual data points (scatter plot).

    By selecting the Individual data points option under Chart type, you can shift to a disaggregated view of the data with the availability of a color axis.

Screenshot of the dashboard, showing the data analysis with the 'Individual data points' option selected.

Feature importances (model explanations)

By using the model explanation component, you can see which features were most important in your model’s predictions. You can view what features affected your model’s prediction overall on the Aggregate feature importance pane or view feature importances for individual data points on the Individual feature importance pane.

Aggregate feature importances (global explanations)

Screenshot of the dashboard, showing aggregate feature importances on the 'Feature importances' pane.

  1. Top k features: Lists the most important global features for a prediction and allows you to change it by using a slider bar.

  2. Aggregate feature importance: Visualizes the weight of each feature in influencing model decisions across all predictions.

  3. Sort by: Allows you to select which cohort's importances to sort the aggregate feature importance graph by.

  4. Chart type: Allows you to select between a bar plot view of average importances for each feature and a box plot of importances for all data.

    When you select one of the features in the bar plot, the dependence plot is populated, as shown in the following image. The dependence plot shows the relationship of the values of a feature to its corresponding feature importance values, which affect the model prediction.

    Screenshot of the dashboard, showing a populated dependence plot on the 'Aggregate feature importances' pane.

  5. Feature importance of [feature] (regression) or Feature importance of [feature] on [predicted class] (classification): Plots the importance of a particular feature across the predictions. For regression scenarios, the importance values are in terms of the output, so positive feature importance means it contributed positively toward the output. The opposite applies to negative feature importance. For classification scenarios, positive feature importances mean that feature value is contributing toward the predicted class denoted in the y-axis title. Negative feature importance means it's contributing against the predicted class.

  6. View dependence plot for: Selects the feature whose importances you want to plot.

  7. Select a dataset cohort: Selects the cohort whose importances you want to plot.

Individual feature importances (local explanations)

The following image illustrates how features influence the predictions that are made on specific data points. You can choose up to five data points to compare feature importances for.

Screenshot of the dashboard, showing the 'Individual feature importances' pane.

Point selection table: View your data points and select up to five points to display in the feature importance plot or the ICE plot below the table.

Screenshot of the dashboard, showing a bar plot on the 'Individual feature importances' pane.

Feature importance plot: A bar plot of the importance of each feature for the model's prediction on the selected data points.

  1. Top k features: Allows you to specify the number of features to show importances for by using a slider.
  2. Sort by: Allows you to select the point (of those checked above) whose feature importances are displayed in descending order on the feature importance plot.
  3. View absolute values: Toggle on to sort the bar plot by the absolute values. This allows you to see the most impactful features regardless of their positive or negative direction.
  4. Bar plot: Displays the importance of each feature in the dataset for the model prediction of the selected data points.

Individual conditional expectation (ICE) plot: Switches to the ICE plot, which shows model predictions across a range of values of a particular feature.

Screenshot of the dashboard, showing an ICE plot on the 'Individual feature importances' pane.

  • Min (numerical features): Specifies the lower bound of the range of predictions in the ICE plot.
  • Max (numerical features): Specifies the upper bound of the range of predictions in the ICE plot.
  • Steps (numerical features): Specifies the number of points to show predictions for within the interval.
  • Feature values (categorical features): Specifies which categorical feature values to show predictions for.
  • Feature: Specifies the feature to make predictions for.

Counterfactual what-if

Counterfactual analysis provides a diverse set of what-if examples generated by changing the values of features minimally to produce the desired prediction class (classification) or range (regression).

Screenshot of the dashboard, showing counterfactuals.

  1. Point selection: Selects the point to create a counterfactual for and display in the top-ranking features plot below it.

    Screenshot of the dashboard, showing a top ranked features plot.

    Top ranked features plot: Displays, in descending order of average frequency, the features to perturb to create a diverse set of counterfactuals of the desired class. You must generate at least 10 diverse counterfactuals per data point to enable this chart, because there's a lack of accuracy with a lesser number of counterfactuals.

  2. Selected data point: Performs the same action as the point selection in the table, except in a dropdown menu.

  3. Desired class for counterfactual(s): Specifies the class or range to generate counterfactuals for.

  4. Create what-if counterfactual: Opens a panel for counterfactual what-if data point creation.

    Select the Create what-if counterfactual button to open a full window panel.

    Screenshot of the dashboard, showing what-if counterfactuals.

  5. Search features: Finds features to observe and change values.

  6. Sort counterfactual by ranked features: Sorts counterfactual examples in order of perturbation effect. (Also see Top ranked features plot, discussed earlier.)

  7. Counterfactual examples: Lists feature values of example counterfactuals with the desired class or range. The first row is the original reference data point. Select Set value to set all the values of your own counterfactual data point in the bottom row with the values of the pre-generated counterfactual example.

  8. Predicted value or class: Lists the model prediction of a counterfactual's class given those changed features.

  9. Create your own counterfactual: Allows you to perturb your own features to modify the counterfactual. Features that have been changed from the original feature value are denoted by the title being bolded (for example, Employer and Programming language). Select See prediction delta to view the difference in the new prediction value from the original data point.

  10. What-if counterfactual name: Allows you to name the counterfactual uniquely.

  11. Save as new data point: Saves the counterfactual you've created.

Causal analysis

The next sections cover how to read the causal analysis for your dataset on select user-specified treatments.

Aggregate causal effects

Select the Aggregate causal effects tab of the causal analysis component to display the average causal effects for pre-defined treatment features (the features that you want to treat to optimize your outcome).

Note

Global cohort functionality is not supported for the causal analysis component.

Screenshot of the dashboard, showing causal analysis on the 'Aggregate causal effects' pane.

  1. Direct aggregate causal effect table: Displays the causal effect of each feature aggregated on the entire dataset and associated confidence statistics.

    • Continuous treatments: On average in this sample, increasing this feature by one unit will cause the probability of class to increase by X units, where X is the causal effect.
    • Binary treatments: On average in this sample, turning on this feature will cause the probability of class to increase by X units, where X is the causal effect.
  2. Direct aggregate causal effect whisker plot: Visualizes the causal effects and confidence intervals of the points in the table.

Individual causal effects and causal what-if

To get a granular view of causal effects on an individual data point, switch to the Individual causal what-if tab.

Screenshot of the dashboard showing causal analysis on the individual causal what-if tab.

  1. X-axis: Selects the feature to plot on the x-axis.
  2. Y-axis: Selects the feature to plot on the y-axis.
  3. Individual causal scatter plot: Visualizes points in the table as a scatter plot to select data points for analyzing causal what-if and viewing the individual causal effects below it.
  4. Set new treatment value:
    • (numerical): Shows a slider to change the value of the numerical feature as a real-world intervention.
    • (categorical): Shows a dropdown list to select the value of the categorical feature.

Treatment policy

Select the Treatment policy tab to switch to a view to help determine real-world interventions and show treatments to apply to achieve a particular outcome.

Screenshot of the dashboard, showing causal analysis on the 'Treatment policy' pane.

  1. Set treatment feature: Selects a feature to change as a real-world intervention.

  2. Recommended global treatment policy: Displays recommended interventions for data cohorts to improve the target feature value. The table can be read from left to right, where the segmentation of the dataset is first in rows and then in columns. For example, for 658 individuals whose employer isn't Snapchat and whose programming language isn't JavaScript, the recommended treatment policy is to increase the number of GitHub repos contributed to.

    Average gains of alternative policies over always applying treatment: Plots the target feature value in a bar chart of the average gain in your outcome for the above recommended treatment policy versus always applying treatment.

    Screenshot of the dashboard showing a bar chart of the average gains of alternative policies over always applying treatment on the treatment policy tab.

    Recommended individual treatment policy:

    Screenshot of the dashboard showing a recommended individual treatment policy table on the treatment policy tab.

  3. Show top k data point samples ordered by causal effects for recommended treatment feature: Selects the number of data points to show in the table.

  4. Recommended individual treatment policy table: Lists, in descending order of causal effect, the data points whose target features would be most improved by an intervention.

Next steps