Part I
Kacper Sokol
Trustworthiness
No silly mistakes
Fairness
Does not discriminate
New knowledge
Aids in scientific discovery
Legislation
Does not break the law
\[
f(\mathbf{x}) = 0.2 \;\; + \;\; 0.25 \times x_1 \;\; + \;\; 0.7 \times x_4 \;\; - \;\; 0.2 \times x_5 \;\; - \;\; 0.9 \times x_7
\]
\[
\mathbf{x} = (0.4, \ldots, 1, \frac{1}{2}, \ldots \frac{1}{3})
\]
\[ f(\mathbf{x}) = 0.2 \;\; \underbrace{+0.1}_{x_1} \;\; \underbrace{+0.7}_{x_4} \;\; \underbrace{-0.1}_{x_5} \;\; \underbrace{-0.3}_{x_7} \;\; = \;\; 0.6 \]
It requires effort
A generic eXplainable Artificial Intelligence process is beyond our reach at the moment
XAI Taxonomy spanning social and technical desiderata:
• Functional • Operational • Usability • Safety • Validation •
(Sokol and Flach, 2020. Explainability Fact Sheets: A Framework for Systematic Assessment of Explainable Approaches)
Framework for black-box explainers
(Henin and Le Métayer, 2019. Towards a generic framework for black-box explanations of algorithmic decision systems)
(Explainability Fact Sheets)
Social and technical explainability desiderata spanning five dimensions
👥 Audience
⚙️️ Operationalisation
🧰 Applicability
Had you been 10 years younger, your loan application would be accepted.
F1 Problem Supervision Level |
|
F2 Problem Type |
|
F6 Applicable Model Class |
|
F7 Relation to the Predictive System |
|
F5 Computational Complexity |
|
F8 Compatible Feature Types |
|
F9 Caveats and Assumptions |
|
F3 Explanation Target |
|
F4 Explanation Breadth/Scope |
|
U1 Soundness |
How truthful it is with respect to the black box? |
(✔) |
U2 Completeness |
How well does it generalise? |
(✗) |
U3 Contextfullness |
“It only holds for people older than 25.” |
|
U11 Parsimony |
How short is it? |
(✔) |
U6 Chronology |
More recent events first. |
|
U7 Coherence |
Comply with the natural laws (mental model). |
|
U8 Novelty |
Avoid stating obvious / being a truism. |
|
U9 Complexity |
Appropriate for the audience. |
U5 Actionability |
Actionable foil. |
(✔) |
U4 Interactiveness |
User-defined foil. |
(✔) |
U10 Personalisation |
User-defined foil. |
(✔) |
O1 Explanation Family |
|
O2 Explanatory Medium |
|
O3 System Interaction |
|
O4 Explanation Domain |
|
O5 Data and Model Transparency |
|
O6 Explanation Audience |
|
O7 Function of the Explanation |
|
O8 Causality vs. Actionability |
|
O9 Trust and Performance |
|
O10 Provenance |
|
S1 Information Leakage |
Contrastive explanation leak precise values. |
S2 Explanation Misuse |
Can be used to reverse-engineer the black box. |
S3 Explanation Invariance |
Does it always output the same explanation (stochasticity / stability)? |
S4 Explanation Quality |
Is it from the data distribution? |
V1 User Studies |
|
V2 Synthetic Experiments |
🔍 has nice theoretical properties (F9 Caveats and Assumptions)
The explanation is always [insert your favourite claim here].
(You know it when you see it!)
\[ \texttt{Explainability} \; = \] \[ \underbrace{ \texttt{Reasoning} \left( \texttt{Transparency} \; | \; \texttt{Background Knowledge} \right)}_{\textit{understanding}} \]
Explainability → explainee walking away with understanding
A continuous spectrum rather than a binary property
Humans | Task | |
---|---|---|
Application-grounded Evaluation | Real Humans | Real Tasks |
Human-grounded Evaluation | Real Humans | Simple Tasks |
Functionally-grounded Evaluation | No Real Humans | Proxy Tasks |
Each (real-life) explainability scenario is unique and requires a bespoke solution
Explainers are socio-technical constructs, hence we should strive for seamless integration with humans as well as technical correctness and soundness