Layer-wise Relevance Propagation
Explore how Layer-wise Relevance Propagation (LRP) works to explain neural network outputs by attributing relevance scores to input features. Understand its backward propagation method, implementation on neural networks, and how it generates saliency maps highlighting important image regions that influence predictions. Gain a practical grasp of LRP's application and limitations for interpreting classifier decisions.
We'll cover the following...
Mathematical model for LRP
Layer-wise Relevance Propagation (LRP) explains the output of a neural network by attributing relevance scores to each input feature. Like saliency maps, these scores indicate how much each input feature contributes to the network output.
LRP propagates the prediction score backward in the neural network by using a set of purposely designed propagation rules. These rules backpropagate the relevance of each neuron from the target layer back to the input layer, following a conservation property.
Note: The conservation property ensures that the sum of relevances of all neurons in a particular layer equals the prediction score of the network.
In other words, given the relevance score
where