Javascript must be enabled to continue!
Counterfactual Models for Fair and Adequate Explanations
View through CrossRef
Recent efforts have uncovered various methods for providing explanations that can help interpret the behavior of machine learning programs. Exact explanations with a rigorous logical foundation provide valid and complete explanations, but they have an epistemological problem: they are often too complex for humans to understand and too expensive to compute even with automated reasoning methods. Interpretability requires good explanations that humans can grasp and can compute. We take an important step toward specifying what good explanations are by analyzing the epistemically accessible and pragmatic aspects of explanations. We characterize sufficiently good, or fair and adequate, explanations in terms of counterfactuals and what we call the conundra of the explainee, the agent that requested the explanation. We provide a correspondence between logical and mathematical formulations for counterfactuals to examine the partiality of counterfactual explanations that can hide biases; we define fair and adequate explanations in such a setting. We provide formal results about the algorithmic complexity of fair and adequate explanations. We then detail two sophisticated counterfactual models, one based on causal graphs, and one based on transport theories. We show transport based models have several theoretical advantages over the competition as explanation frameworks for machine learning algorithms.
Title: Counterfactual Models for Fair and Adequate Explanations
Description:
Recent efforts have uncovered various methods for providing explanations that can help interpret the behavior of machine learning programs.
Exact explanations with a rigorous logical foundation provide valid and complete explanations, but they have an epistemological problem: they are often too complex for humans to understand and too expensive to compute even with automated reasoning methods.
Interpretability requires good explanations that humans can grasp and can compute.
We take an important step toward specifying what good explanations are by analyzing the epistemically accessible and pragmatic aspects of explanations.
We characterize sufficiently good, or fair and adequate, explanations in terms of counterfactuals and what we call the conundra of the explainee, the agent that requested the explanation.
We provide a correspondence between logical and mathematical formulations for counterfactuals to examine the partiality of counterfactual explanations that can hide biases; we define fair and adequate explanations in such a setting.
We provide formal results about the algorithmic complexity of fair and adequate explanations.
We then detail two sophisticated counterfactual models, one based on causal graphs, and one based on transport theories.
We show transport based models have several theoretical advantages over the competition as explanation frameworks for machine learning algorithms.
Related Results
Explaining Data-Driven Decisions made by AI Systems: The Counterfactual Approach
Explaining Data-Driven Decisions made by AI Systems: The Counterfactual Approach
We examine counterfactual explanations for explaining the decisions made by model-based AI systems. The counterfactual approach we consider defines an explanation as a set of the s...
Data Augmentation using Counterfactuals: Proximity vs Diversity
Data Augmentation using Counterfactuals: Proximity vs Diversity
Counterfactual explanations are gaining in popularity as a way of explaining machine learning models. Counterfactual examples are generally created to help interpret the decision o...
Downward counterfactual insights into weather extremes
Downward counterfactual insights into weather extremes
<p>There are many regions where the duration of reliable scientific observations of key weather hazard variables, such as rainfall and wind speed, is of the order of ...
Multi-Class Counterfactual Explanations using Support Vector Data Description
Multi-Class Counterfactual Explanations using Support Vector Data Description
<p>Explainability is becoming increasingly crucial in machine learning studies and, as the complexity of the model increases, so does the complexity of its explanation. Howev...
The Asymmetry of Counterfactual Dependence
The Asymmetry of Counterfactual Dependence
A certain type of counterfactual is thought to be intimately related to causation, control, and explanation. The time asymmetry of these phenomena therefore plausibly arises from a...
Counterfactual Reasoning in Children: Evidence from an Eye-Tracking Study with Turkish-Speakers
Counterfactual Reasoning in Children: Evidence from an Eye-Tracking Study with Turkish-Speakers
Previous research has produced mixed results regarding the ability of children as young as four years of age to engage in counterfactual reasoning. In this study, we employed a vis...
Diffusion Counterfactuals forĀ Image Regressors
Diffusion Counterfactuals forĀ Image Regressors
Abstract
Counterfactual explanations have been successfully applied to create human interpretable explanations for various black-box models. They are handy for tasks in t...
Curriculum Development for FAIR Data Stewardship
Curriculum Development for FAIR Data Stewardship
Abstract
The FAIR Guidelines attempts to make digital data Findable, Accessible, Interoperable, and Reusable (FAIR). To prepare FAIR data, a new data science discipl...

