Post Hoc Explanations Feature Attributions 3 of 4
Post Hoc Explanations Feature Attributions 3 of 4
by adding noise
Output
▪ Key idea: backpropagate importance
… through the network
o Vanilla gradients
… o Layerwise relevance propagation (Bach et al.)
o Integrated gradients (Sundararajan et al.)
…
o DeepLIFT (Shrikumar et al.)
… o Deconvolution (Zeiler & Fergus, 2014)
Yellow = inputs o Guided Backpropagation (Springenberg et al,
2014)
Limitations of Sensitivity Maps
▪ Visually noisy
o Often highlight pixels that–to a human eye–seem randomly selected
o a priori, we cannot know if this noise reflects an underlying truth about
how networks perform classification, or is due to more superficial factors
■ The SmoothGrad paper answers this question - we’ll get to this soon!
Theory Behind SmoothGrad: Noisy Gradients
▪ Key idea behind SmoothGrad: noisy maps are due to noisy gradients
▪ Given these rapid fluctuations, gradient of S c at any given point will be less
meaningful than a local average of gradient values.
SmoothGrad: Intuition
▪ Recall that noisy maps are due to noisy gradients
▪ Simple solution:
o take an image of interest
o sample similar images by adding Gaussian noise to the image
o take the average of the resulting sensitivity maps for each sampled
image
■ This smoothes the gradient
SmoothGrad: Algorithm
1. Take random samples in a neighborhood of an input x with added noise
2. Average the resulting sensitivity maps.
n is the number of samples, and Ɲ(0, σ2) represents Gaussian noise with
standard deviation σ.
Experimental Setup
Definition (Visual Coherence): Highlights are only on the object of interest, not
the background
Definition (Discriminativity):
the ability to explain /
distinguish separate objects
without confusion
Qualitative Results: Discriminativity
Open Problem