Professional Documents
Culture Documents
5 March 2017
Making Deep Neural Nets Transparent
DNN transparency
Examples:
I Creating a class prototype: maxxX log p(c |x).
converged solutions x ?
initial solutions
Observations:
I AM builds typical patterns for these classes (e.g. beaks, legs).
I Unrelated background objects are not present in the image.
Idea:
I Force the interpretation x ? to match the data more closely.
Find the input pattern that Find the most likely input
maximizes class probability. pattern for a given class.
x0 x*
x0
x*
Find the input pattern that Find the most likely input
maximizes class probability. pattern for a given class.
Nguyen et al. 2016 introduced several enhancements for activation
maximization:
I Multiplying the objective by an expert p(x):
p(x|c ) p(c |x) p(x)
| {z }
old
I Optimization in code space:
max p(c | g(z)) + kzk2 x ? = g(z ? )
zZ |{z}
x
These two techniques require an unsupervised model of the data,
either a density model p(x) or a generator g(z).
log p(c|x)
interpre- 0123456789 AM + density
neural
10
tation network - optimum has
for c
clear meaning
+ log p(x|c)
784
900
1
log p(x)
discriminative model
density model
x=g(z)
z
AM + generator log p(c|x)
- more straightforward neural
784
900
100
10
network
to optimize
- not optimizing
log p(x|c) 0123456789
generative model
x0 x*
x0
x*
Observations:
I Summarizing a concept or category like motorbike into a
single image can be difficult (e.g. different views or colors).
I A good interpretation would grow as large as the diversity of
the concept to interpret.
Individual explanation:
GoogleNet
"motorbike"
GDB-7
high
Individual explanation:
GDB-7
= ...
f(x) R1 R2
x2
f(x')
R1 R2
x1
Remarks:
I Easy to implement (we only need access to the gradient
of the decision function).
I But does it really explain the prediction?
R3
R2 R4
aggregate R1
f(x) =
quantity
Ri = f (x)
P
i
decomposition
Examples:
I Economic activity (e.g. petroleum, cars, medicaments, ...)
I Energy production (e.g. coal, nuclear, hydraulic, ...)
I Evidence for object in an image (e.g. pixel 1, pixel 2, pixel 3, ...)
I Evidence for meaning in a text (e.g. word 1, word 2, word 3, ...)
Sensitivity analysis
f 2
Ri =
xi x=x
x = lim x 0.
e
0
x1
f(x)
500
500
500
500
x2
f(x)
x2
x1
f
Ri = (xi xei )
xi x=e
x
f
Ri = (xi xei )
xi x=e Naive Taylor decomposition
x
x1
f(x)
500
500
500
500
x2
decomposition
"naive" Taylor
Advantages
I Decomposes the desired
quantity f (x) in a
principled way.
Disadvantages
I Relevance functions are
highly non-smooth.
I Relevance scores are
sometimes negative.
I Inflexible w.r.t. the model.
x = 0)
Naive Taylor (e x = 0)
Naive Taylor (e
f(x)
= +
x2
x1 1. decompose
decision function
2. explain
subfunctions
releva
nce
ce of x
3. aggregate
van 1 explanations
le x2
re of
Taylor
decomposition
(TD)
deep Taylor
decomposition
(DTD)
TD TD TD
x {x : h 0 constraints}
e
f (x) = i Ri with Ri = hi ci
P
and X wij+ hj cj
ci = + >0
i hi wij
P
j
500
500
500
500
x1 x2
0
1
sensitivity analysis Taylor decomposition deep Taylor LRP
x = 0)
Naive Taylor (e x = 0)
Naive Taylor (e