Ground-Truth Faithfulness—Feature Agreement
Learn to evaluate the ground-truth faithfulness of an explanation using the Top-k feature agreement metric.
We'll cover the following
There are several algorithms to generate explanations to model predictions; let’s try to understand how to evaluate these algorithms. While explanations are inherently subjective, benchmarking these algorithms under certain aspects can be insightful to their behavior. We’ll start with our first class of metrics, known as ground-truth faithfulness.
Ground-truth faithfulness
Ground-truth faithfulness is a class of metrics that evaluate how an explanation generated by an algorithm compares to the ground-truth explanation. A human annotator can provide the ground-truth explanation, or it can be assumed to come from a state-of-the-art XAI algorithm.
In this lesson and the next, we’ll evaluate the ground-truth faithfulness of saliency map algorithms like vanilla gradients, smooth gradients, etc., using two metrics:
Top-k feature agreement
Rank correlation
Since we don't have ground-truth explanations, we’ll assume integrated gradients as our ground-truth saliency.
Top-k feature agreement
The Top-k feature agreement computes the common fraction of Top-k features between a given post-hoc explanation and the corresponding ground-truth explanation. In other words, given a post-hoc explanation
Get hands-on with 1400+ tech skills courses.