Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals

Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals

Yanai Elazar1,2 Shauli Ravfogel1,2 Alon Jacovi1 Yoav Goldberg1,2
1Computer Science Department, Bar Ilan University
2Allen Institute for Artificial Intelligence
{yanaiela,shauli.ravfogel,alonjacovi,yoav.goldberg}@gmail.com

Abstrakt

A growing body of work makes use of probing
in order to investigate the working of neural
Modelle, often considered black boxes. Kürzlich,
an ongoing debate emerged surrounding the
limitations of the probing paradigm. In diesem
arbeiten, we point out the inability to infer beha-
vioral conclusions from probing results, Und
offer an alternative method that focuses on
how the information is being used, rather than
on what information is encoded. Our method,
Amnesic Probing, follows the intuition that the
utility of a property for a given task can be as-
sessed by measuring the influence of a causal
intervention that removes it from the represen-
Station. Equipped with this new analysis tool,
we can ask questions that were not possible
Vor, Zum Beispiel, is part-of-speech informa-
tion important for word prediction? We per-
form a series of analyses on BERT to answer
these types of questions. Our findings demon-
strate that conventional probing performance
is not correlated to task importance, and we
call for increased scrutiny of claims that draw
behavioral or causal conclusions from probing
results.1

1

Einführung

What drives a model to perform a specific pre-
diction? Was
information is being used for
prediction, and what would have happen if that
information went missing? Because neural repre-
sentation is opaque and hard to interpret, Antwort-
ing these questions is challenging.

The recent advancements in Language Models
(LMs) and their success in transfer learning of
many NLP tasks (z.B., Peters et al., 2018; Devlin
et al., 2019; Liu et al., 2019B) spiked interest

1The code is available at: https://github.com

/yanaiela/amnesic probing.

160

in understanding how these models work and what
is being encoded in them. One prominent meth-
odology that attempts to shed light on those ques-
tions is probing (Conneau et al., 2018) (Auch
known as auxilliary prediction [Adi et al., 2016]
and diagnostic classification [Hupkes et al., 2018]).
Under this methodology, one trains a simple model
—a probe—to predict some desired information
from the latent representations of the pre-trained
Modell. High prediction performance is interpreted
as evidence for the information being encoded
in the representation. A key drawback of such an
approach is that while it may indicate that the
information can be extracted from the represen-
Station, it provides no evidence for or against the
actual use of this information by the model.
In der Tat, Hewitt and Liang (2019) have shown that
unter bestimmten Bedingungen, above-random probing
accuracy can be achieved even when the infor-
mation that one probes for is linguistically mean-
ingless noise, which is unlikely to have any use
by the actual model. More recently, Ravichander
et al. (2020) showed that models encode linguis-
tic properties, even when not required at all for
solving the task, questioning the usefulness and
common interpretation of probing. These results
call for higher scrutiny of causal claims based on
probing results.

In diesem Papier, we propose a counterfactual ap-
proach that serves as a step towards causal at-
tribution: Amnesic Probing (siehe Abbildung 1 for a
schematic view). We build on the intuition that
if a property Z (z.B., part-of-speech) is being
used for a task T (z.B., language modeling), Dann
the removal of Z should negatively influence the
ability of the model to solve the task. Umgekehrt,
when the removal of Z has little or no influence on
the ability to solve T , one can argue that knowing
Z is not a significant contributing factor in the
strategy the model employs in solving T .

As opposed to previous work that focused on
intervention in the input space (Goyal et al., 2019;

Transactions of the Association for Computational Linguistics, Bd. 9, S. 160–175, 2021. https://doi.org/10.1162/tacl a 00359
Action Editor: Radu Florian. Submission batch: 7/2020; Revision batch: 9/2020; Published 3/2021.
C(cid:2) 2021 Verein für Computerlinguistik. Distributed under a CC-BY 4.0 Lizenz.

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

where one removes some component and mea-
sures the influence of that intervention.

We study several

linguistic properties such
as part-of-speech (POS) and dependency labels.
Gesamt, we find that as opposed to the common
belief, high probing performance does not mean
that the probed information is used for predicting
the main task (§4). This is consistent with the
recent findings of Ravichander et al. (2020). Unser
analysis also reveals that the properties we exam-
ine are often being used differently in the masked
setting (which is mostly used in LM training) Und
in the non-masked setting (which is commonly
used for probing or fine-tuning) (§5). We then
dive deeper into a more fine-grained analysis, Und
show that not all of the linguistic property labels
equally influence prediction (§6). Endlich, we re-
evaluate previous claims about the way that BERT
process the traditional NLP pipeline (Tenney et al.,
2019A) with amnesic probing and provide a novel
interpretation on the utility of different layers (§7).

2 Amnesic Probing

2.1 Setup and Formulation

Given a set of labeled data of data points X =
2 and task labels Y = y1, . . . , yn we
x1, . . . , xn
analyze a model f that predicts the labels Y
from X: ˆyi = f (xi). We assume that this model
is composed of two parts: an encoder h that
transforms input xi into a representation vector
hxi and a classifier c that is used for predicting ˆyi
based on hxi: ˆyi = c(H(xi)). We refer by model to
the component that follows the encoding function
h and is used for the classification of the task of
interest y. Each data point xi is also associated
with a property of interest zi which represents
additional information, which may or may not
affect the decision of the classifier c.

In this work, we are interested in the change in
prediction of the classifier c on the prediction ˆyi
which is caused due to the removal of the property
Z from the representation h(xi), that is h(xi)¬Z.

2.2 Amnesic Probing with INLP

Under the counterfactual approach, we aim to
evaluate the behavioral influence of a specific
type of information Z (z.B., POS) on some task

2The data points can be words, documents, Bilder, usw.,

based on the application.

Figur 1: A schematic description of the proposed
amnesic intervention: We transform the contextualized
representation of the word ‘‘ran’’ so as to remove
Information (Hier, POS), resulting in a ‘‘cleaned’’
version h¬P OS
ran . This representation is fed to the word-
prediction layer and the behavioral influence of POS
erasure is measured.

Kaushik et al., 2020; Vig et al., 2020) or in specific
Neuronen (Vig et al., 2020), our intervention is done
on the representation layers. This makes it easier
than changing the input (which is non-trivial) Und
more efficient than querying hundred of neurons
(which become combinatorial when considering
the effect of multiple neurons simultaneously).

We demonstrate that amnesic probing can func-
tion as a debugging and analysis tool for neural
Modelle. Speziell, by using amnesic probing
we show how to deduce whether a property is
used by a given model in prediction.

In order to build the counterfactual represen-
tations, we need a function that operates on a pre-
trained representation and returns a counterfactual
version which no longer encodes the property we
focus on. We use the recently proposed algorithm
for neutralizing linear information: Iterative Null-
space Projection (INLP) (Ravfogel et al., 2020).
This approach allows us to ask the counterfactual
question: ‘‘How will the prediction of a task differ
without access to some property?’’ (Pearl and
Mackenzie, 2018). This approach relies on the as-
sumption that the usefulness of some information
can be measured by neutralizing it from the repre-
sentation, and witnessing the resulting behavioral
ändern. It echoes the basic idea of ablation tests

161

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

(z.B., language modeling). To do so, we selectively
remove this information from the representation
and observe the change in the behavior of the
model on the main task.

One commonly used method for information
removal relies on adversarial training through the
gradient reversal layer technique (Ganin et al.,
2016). Jedoch, this techniques requires chang-
ing the original encoding by retraining the model,
which is not desired in our case as we wish to
study the original model’s behavior. Zusätzlich,
Elazar and Goldberg (2018)
Das
technique does not completely remove all the
information from the learned representation.

found that

Stattdessen, we make use of a recently proposed
algorithm called Iterative Nullspace Projection
(INLP) (Ravfogel et al., 2020). Given a labeled
dataset of representations H, and a property to
remove, Z, INLP neutralizes the ability to linearly
predict Z from H. It does so by training a sequence
of linear classifiers (probes) c1, . . . , ck that predict
Z, interpreting each one as conveying information
on a unique direction in the latent space that
corresponds to Z, and iteratively removing each
of these directions. Concretely, we assume that
the ith probe ci is parameterized by a matrix Wi.
In the ith iteration, ci is trained to predict Z from
H 3, and the data is projected onto its nullspace
using a projection matrix PN (Wi). This operation
guarantees WiPN (Wi)H = 0, d.h., it neutralizes
the features in the latent space which were found
by Wi to be indicative to Z. By repeating this
process until no classifier achieves above-majority
accuracy, INLP removes all such features.4

Amnesic Probing vs. Probing Note
Das
amnesic probing extends conventional probing,
as it is only relevant in cases where the property of
interest can be predicted from the representation.
If a probe gets random accuracy, the information
cannot be used by the model to begin with. Als
solch, amnesic probing can be seen as a comple-
mentary method, which inspects probe accuracy as
a first step, but then proceeds to derive behavioral
outcomes from the directions associated with the
probe, with respect to a specific task.

2.3 Controls

The usage of INLP in this setup involves some
subtleties we aim to account for: (1) Any modifi-
cation to the representation, regardless of whether
it removes information necessary to the task, may
cause a decrease in performance. Can the drop
in performance be attributed solely to the modi-
fication of the representation? (2) The removal
of any property using INLP may also cause re-
moval of correlating properties. Does the re-
moved information only pertain to the property
in question?

Control Over Information In order to control
for the information loss of the representations, Wir
make use of a baseline that removes the same
number of directions as INLP does, but randomly.
For every INLP iteration the data matrix’s rank
decreases by the number of labels of the inspected
Eigentum. This operation removes information
from the representation which might be used for
prediction. Using this control, Rand, instead of
finding the directions using a classifier that learned
some task, we generate random vectors from a
uniform distribution, that accounts for random
directions. Dann, we construct the projection ma-
trix as in INLP, by finding the intersection of
nullspaces.

If the Rand impact on performance is lower
than the impact of amnesic probing for some
Eigentum, we conclude that we removed important
directions for the main task. Ansonsten, when the
Rand control has a similar or higher impact, Wir
conclude that there is no evidence for property
usage for the main task.

Control over Selectivity5 The result of the am-
nesic probing is taken as an indication to whether
or not the model we query makes use of the
inspected property for prediction. Jedoch, Die
removed features might solely correlate with the
Eigentum (z.B., word position in the sentence has
a nonzero correlation to syntactic function). To
what extent is the information removal process we
employ selective to the property in focus?

We test that by explicitly providing the gold
information that has been removed from the

3Concretely, we use linear SVM (Pedregosa et al., 2011).
4All relevant directions are removed to the extent they
are identified by the classifiers we train. daher, we run
INLP until the last linear classifier achieves a score within
one point above majority accuracy on the development set.

5Nicht

to be confused with Hewitt and Liang (2019)
Selectivity. Although recommended to use when performing
standard probing, we argue it does not fit as a control
for amnesic probing and provide a detailed explanation in
Appendix B.

162

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

representation, and finetuning the subsequent
layers (while the rest of the network is frozen).
Restoring the original performance is taken as
evidence that the property we aimed to remove
is enough to account for the damage sustained
by the amnesic intervention (it may still be the
case that
the intervention removes unrelated
properties; but given the explicitly-provided pro-
perty information, the model can make up for the
damage). Jedoch, if the original performance is
not restored, this indicates that the intervention
removed more information than intended, Und
this cannot be accounted for by merely explicitly
providing the value of the single property we
focused on.

Concretely, we concatenate feature vectors of
the studied property to the amnesic representa-
tionen. Those vectors are 32-dimensional, and are
initialized randomly, with a unique vector for each
value of the property of interest. Those are fine-
tuned until convergence. We note that as the new
representation vectors are of a higher dimension
than the original ones, we cannot use the original
Matrix. For an easier learning process, we use the
original embedding matrix and concatenate it with
a new embedding matrix, randomly initialized, Und
treat it as the new decision function.

3.2 Studied Properties

We focus on six tasks of sequence tagging: coarse
and fine-grained part-of-speech tags (c-pos and
f-pos, jeweils); syntactic dependency labels
(dep); named-entity labels (ner); and syntactic
constituency boundaries7 that mark the beginning
and the end of a phrase (phrase start, and phrase
end, jeweils).

We use the training and dev data of the follow-
ing datasets for each task: English UD Treebank
(McDonald et al., 2013) for c-pos, f-pos, and dep;
and English OntoNotes (Weischedel et al., 2013)
for ner, phrase start, and phrase end. For train-
ing, wir gebrauchen 100,000 random tokens from those
datasets.

3.3 Metrics

We report the following metrics:
LM accuracy: Word prediction accuracy.
Kullback-Leibler Divergence (DKL): We calcu-
late the DKL between the distribution of the
model over tokens, before and after the amnesic
intervention. This measure focuses on the entire
distribution, rather than the correct token only.
Larger values implies a more significant change.

3 Studying BERT: Experimental Setup

4 To Probe or Not to Probe?

3.1 Modell

We use our proposed method to investigate BERT
(Devlin et al., 2019),6 a popular and competitive
masked language model (MLM) that has recently
been the subject of many analysis works (z.B.,
Hewitt and Manning, 2019; Liu et al., 2019A;
Tenney et al., 2019A). While most probing works
focus on the ability to decode a certain linguistic
property of the input text from the representation,
we aim to understand which information is being
used by it when predicting words from context.
Zum Beispiel, we seek to answer questions such
as the following: ‘‘Is POS information used by
the model in word prediction?’’ The following
experiments focus on language modeling, as a
basic and popular task, but our method is more
widely applicable.

By using the probing technique, different linguis-
tic phenomenon such as POS, dependency infor-
mation, and NER (Tenney et al., 2019A; Liu et al.,
2019A; Alt et al., 2020) have been found to
be ‘‘easily extractable’’ (typically using linear
probes). A naive interpretation of these results may
conclude that because information can be easily
extracted by the probing model, this information
is being used for the predictions. We show that
this is not the case. Some properties such as
syntactic structure and POS are very informative
and are being used in practice to predict words.
Jedoch, we also find some properties, wie zum Beispiel
phrase markers, which the model does not make
use of when predicting tokens, in contrast to what
one can naively deduce from probing results. Das
finding is in line with a recent work that observed
the same behavior (Ravichander et al., 2020).

For each linguistic property, we report the prob-
ing accuracy using a linear model, as well as the

6Speziell, BERT-BASE-UNCASED (Wolf et al., 2019).

7Based on the Penn Treebank syntactic definitions.

163

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Properties

Probing

LM-Acc

LM-DKL

N. dir
N. classes
Majority
Vanilla
Vanilla
Rand
Selectivity
Amnesic
Rand
Amnesic

dep

738
41
11.44
76.00
94.12
12.31
73.78
7.05
8.11
8.53

f-pos

585
45
13.22
89.50
94.12
56.47
92.68
12.31
4.61
7.63

c-pos

264
12
31.76
92.34
94.12
89.65
97.26
61.92
0.36
3.21

ner

phrase start

phrase end

133
19
86.09
93.53
94.00
92.56
96.06
83.14
0.08
1.24

36
2
59.25
85.12
94.00
93.75
96.96
94.21
0.01
0.01

22
2
58.51
83.09
94.00
93.86
96.93
94.32
0.01
0.01

Tisch 1: Property statistics, probing accuracies, and the influence of the amnesic intervention
on the model’s distribution over words. dep: dependency edge identity; f-pos and c-pos:
fine-grained and coarse POS tags; phrase start and phrase end: beginning and end of
phrases. Rand refers to replacing our INLP-based projection with removal of an equal
number of random directions from the representation. The number of iterations per task can
be inferred from: N.dir/N.classes.

word prediction accuracy after removing informa-
tion about that property. The results are summa-
rized in Table 1.8 Probing achieves substantially
higher performance over majority across all tasks.
Darüber hinaus, after neutralizing the studied prop-
erty from the representation, the performance on
that task drops to majority (not presented in the
table for brevity). Nächste, we compare the LM
performance before and after the projection and
observe a major drop for dep and f-pos information
(decrease of 87.0 Und 81.8 accuracy points,
jeweils), and a moderate drop for c-pos
and ner information (decrease of 32.2 Und 10.8
accuracy points, jeweils). For these tasks,
Rand performance on LM-Acc is lower than the
original scores, but substantially higher than the
Amnesic scores. Recall that the Rand experiment
is done with respect to the amnesic probing, daher
the number of removed dimension is the same, Aber
each task may differ in the amount of dimensions
removed. Außerdem, the DKL metric shows
the same trend (but in reverse, as a lower value
indicates on a smaller change). We also report the
selectivity results, where in most experiments the
LM performance is restored, indicating amnesic
probing works as expected. Note that the dep
performance is not fully restored,
thus some
non-related features must have been coupled
and removed with the dependency features. Wir

8Note that because we use two different datasets, Die
f-pos, c-pos, and OntoNotes for
the Vanilla LM-Acc

UD Treebank for dep,
ner, phrase-start, and phrase-end,
performance differ between these setups.

improvement

believe that this happens in part due to the large
number of removed directions.9 These results
suggests that to a large degree, the damage to
LM performance is to be attributed to the specific
information we remove, and not to rank-reduction
allein. We conclude that dependency information,
POS and NER are important for word prediction.
Interessant, for phrase start and phrase end
in accuracy
we observe a small
von 0.21 Und 0.32 points, jeweils. The per-
formance for the control on these properties is
lower, therefore not only are these properties not
important for the LM prediction at this part of the
Modell, they slightly harm it. The last observation
is rather surprising as phrase boundaries are cou-
pled to the structure of sentences, and the words
that form them. A potential explanation for this
phenomenon is that this information is simply
not being used at this part of the model, und ist
rather being processed in an earlier stage. We fur-
ther inspect this hypothesis in Section 7. Endlich,
the probe accuracy does not correlate with task
importance as measured by our method (Spearman
correlation of 8.5, with a p-value of 0.871).

These results strengthen recent works that
question the usefulness of probing as an analysis
tool (Hewitt and Liang, 2019; Ravichander et al.,

9Since this experiment involves additional fine-tuning
and is not entirely comparable to the vanilla setup (also due
to the additional explicit information), we also experiment
with concatenating the inspected features and finetuning. Das
results in an improvement of 3-4 points, above the vanilla
Experiment.

164

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

dep

f-pos

c-pos

ner

phrase start

phrase end

Properties

Probing

LM-Acc

LM-DKL

N. dir
N. classes
Majority
Vanilla
Vanilla
Rand
Selectivity
Amnesic
Rand
Amnesic

820
41
11.44
71.19
56.98
4.67
20.46
4.67
7.77
7.77

675
45
13.22
78.32
56.98
24.69
59.51
6.01
6.10
7.26

240
12
31.76
84.40
56.98
54.55
66.49
33.28
0.45
3.36

95
19
86.09
90.68
57.71
56.88
60.35
48.39
0.10
1.39

35
2
59.25
85.53
57.71
57.46
60.97
56.89
0.02
0.06

52
2
58.51
83.21
57.71
57.27
60.80
56.19
0.04
0.13

Tisch 2: Amnesic probing results for the masked representations. Properties statistics,
word-prediction accuracy and DKL results for the different properties inspected in this
arbeiten. We report the vanilla word prediction accuracy and the Amnesic scores, sowie
the Rand and 1-Hot controls which shows minimal information loss and high selectivity
(except for the dep property which all information was removed). The DKL is also reported
for all properties in the last rows which show similar trends as the accuracy performance.

2020), but measure it from the usefulness of
properties on the main task. We conclude that
high probing performance does not entail this
information is being used at a later part of the
Netzwerk.

5 What Properties are Important for the

Pre-Training Objective?

Probing studies tend to focus on representations
that are used for an end-task (usually the last
hidden layer before the classification layer). Im
case of MLM models, the words are not masked
when encoding them for downstream tasks.

Jedoch, these representations are different
from those used during the pre-training LM phase
(of interest to us), where the input words are
masked. It is therefore unclear if the conclusions
drawn from conventional probing also apply to
the way that the pre-trained model operates.

From this section on, unless mentioned other-
wise, we report our experiments on the masked
Wörter. Das
Ist, given a sequence of tokens
x1, . . . , xi, . . . , xn we encode the representa-
tion of each token xi using its context, als
x1, . . . , xi−1, [M ASK], xi+1, . . . , xn.
follows:
The rest of the tokens remain intact. We feed
these input tokens to BERT, and only use the
masked representation of each word in its context
H(x1, . . . , xi−1, [M ASK], xi+1, . . . , xn)ich.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Figur 2: LM accuracy over INLP predictions, für
the masked tokens version. We present both the vanilla
word-prediction score (straight, blue line), as well as the
Kontrolle (orange, large circles) and INLP (Grün, small
circles). Note that the number of removed dimensions
per iteration differs, based on the number of classes of
that property.

We repeat the experiments from Section 4 Und
report the results in Table 2. Wie erwartet, Die
LM accuracy drops significantly, as the model
does not have access to the original word, und es
has to infer it only based on context. Gesamt, Die

165

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

c-pos

Vanilla Rand Amnesic Δ

c-pos

Vanilla

Amnesic

verb
noun
adposition
determiner
numeral
punctuation
particle
conjunction
adverb
pronoun
adjective
andere

46.72
42.91
73.80
82.29
40.32
80.71
96.40
78.01
39.84
70.29
46.41
70.59

44.85
38.94
72.21
83.53
40.19
81.02
95.71
72.94
34.11
61.93
42.63
76.47

34.99
34.26
37.86
16.64
33.41
47.03
18.74
4.28
23.71
33.23
34.56
52.94

11.73
8.65
35.93
65.66
6.91
33.68
77.66
73.73
16.14
37.06
11.85
17.65

Tisch 3: Masked, c-pos removal, fine-grained
LM analysis. Removing c-pos information and
testing the accuracy performance of words,
accumulating by their label. Δ is the difference
in performance between the vanilla and Amnesic
scores.

trends in the masked setting are similar to the non-
masked setting. Jedoch, this is not always the
Fall, as we show in Section 7. We also report the
selectivity control. Notice that the performance
for this experiment was improved across all tasks.
In the case of dep and f-pos, where we had to
neutralize most of the dimensions the performance
does not fully recover. Note that the number of
classes in those experiments might be a factor
in the large performance gaps (expressed by the
number of removed dimensions, N. dir, im
table). While not part of this study, it would be
interesting to control for this factor in future work.
However for the rest of the properties (c-pos, ner,
and the phrase-markers) the performance is fully
recovered, showing our methods’ selectivity.

To further study the effect of INLP and inspect
how the different dimensions removal affect per-
Form, we display in Figure 2 the LM perfor-
mance after each iteration, both with the amnesic
probing and the control, and observe a consistent
gap between them. Darüber hinaus, we highlight the
difference in the slope for our method and the
random direction removal. The amnesic probing
exemplifies a much steeper slope than the random
Richtung, indicating that the studied properties are
indeed correlated with words prediction. We also
provide the main task performance after each iter-
ation in Figure 5 in the Appendix, which steadily
decreases with each iteration.

verb
noun
adposition
determiner
numeral
punctuation
particle
conjunction
adverb
pronoun
adjective

56.98
56.98
56.98
56.98
56.98
56.98
56.98
56.98
56.98
56.98
56.98

55.60
55.79
53.40
51.04
55.88
53.12
55.26
54.29
55.64
54.97
55.95

Δ

1.38
1.19
3.58
5.94
1.10
3.86
1.72
2.69
1.34
2.02
1.03

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Tisch 4: Word prediction accuracy after fine-
grained tag distinction removal, masked version.
Rand control performance are all between 56.05
Und 56.49 accuracy (with a maximum difference
from vanilla of 0.92 points).

6 Specific Labels and Word Prediction

In the previous sections we observed the impact
(or lack thereof) of different properties on word
prediction. But when a property affects words
prediction, are all words affected similarly? In
this section, we inspect a more fine-grained ver-
sion of the properties of interest, and study the
impact of those on word predictions.

Fine-Grained Analysis When we remove the
POS information from the representation, Sind
nouns affected to the same degree as conjunctions?
We repeat the masked experimental setting from
Abschnitt 5, but this time we inspect the word
prediction performance for the different labels.
We report the results for the c-pos tagging in
Tisch 3. We observe large differences in the
word prediction performance before and after the
POS removal between the labels. Nouns, Zahlen,
and verbs show a relatively small impact in per-
Form (8.64, 6.91, Und 11.73 jeweils),
while conjunctions, particles and determiners de-
monstrate large performance drops (73.73, 77.66,
Und 65.65, jeweils). We see that the infor-
mation about POS labels at the word-level pre-
diction is much more important
in closed-set
vocabularies (such as conjunctions and determin-
ers) than with open vocabularies (such as nouns
and verbs).

166

A manual inspection of predicted words after
removing the POS information reveals that many
of the changes are due to the transformation of
function words to content words. Zum Beispiel,
the words ‘and’, ‘of’, and ‘a’ become ‘rotate’,
‘say’, and ‘final’, jeweils, in the inspected
Sätze. For a more quantitative analysis, Wir
use a POS tagger in order to measure the POS
label confusion before and after the intervention.
Out of the 12,700 determiners conjuncions and
punctuations, 200 of the predicted words by
BERT were tagged as nouns and verbs before
the intervention, compared to 3,982 nach.

Removal of Specific Labels Following the
observation that classes are affected differently
when predicting words, we further investigate the
differences of specific label removal. Zu diesem Zweck,
we repeat the amnesic probing experiments, Aber
instead of removing the fine-grained information
of a linguistic property, we make a cruder re-
moval: The distinction between a specific label
and the rest. Zum Beispiel, with POS as the
general property, we now investigate whether the
information of noun vs. the rest is important for
predicting a word. We perform this experiment
for all of the pos-c labels, and report the results in
Tisch 4.10

We observe big performance gaps when re-
moving different labels. Zum Beispiel, removing
the distinctions between nouns and the rest, oder
verbs and the rest has minimal impact on per-
Form. Andererseits, determiners and
punctuations are highly affected. This is consis-
tent with the previous observation on removing
specific information. These results call for more
detailed observations and experiments when
studying a phenomenon as the fine-grained prop-
erty distinction does not behave the same across
labels.11

7 Behavior Across Layers

The results up to this section treat all of BERT’s
‘Transformer blocks’ (Vaswani et al., 2017) als
the encoding function and the embedding matrix

10In order to properly compare the different properties, Wir
run INLP for solely 60 Iterationen, for each property. Since the
‘other’ tag is not common, we omit it from this experiment.
11We repeat these experiments with the other studied

properties and observe similar trends.

as the model. But what happens when we remove
the information of some linguistic property from
earlier layers?

By using INLP to remove a property from
an intermediate layer, we prevent the subsequent
layer from using linearly present information ori-
ginally stored in that layer. Though this operation
does not erase all
the information correlative
with the studied property (as INLP only removes
linearly present information), it makes it harder
for the model to use this information. Concretely,
we begin by extracting the representation of some
text from the first k layers of BERT and then
run INLP on these representations to remove the
property of interest. Given that we wish to study
the effect of a property on layer i, we project the
representation using the corresponding projection
matrix Pi that was learned on those representation,
and then continue the encoding of the following
layers.12

7.1 Property Recovery After an Amnesic

Operation

Is the property we linearly remove from a given
layer recoverable by subsequent layers? We re-
move the information about some linguistic pro-
perty from layer i, and learn a probe classifier
on all subsequent layers i + 1, . . . , N. This tests
how much information about this property the
following layers have recovered. We experiment
that could be removed
with the properties
without reducing too many dimensions: pos-c,
ner, phrase start, and phrase end. These results are
summarized in Figure 3, both for the non-masked
Ausführung (upper row) and the masked version (lower
Reihe).

Vor allem, for the pos-c, non-masked version, Die
information is highly recoverable in subsequent
layers when applying the amnesic operation on
the first seven layers: the performance drops from
the regular probing of that layer between 5.72 Und
12.69 accuracy points. Jedoch, in the second
part of the network, the drop is substantially larger:
zwischen 16.57 Und 46.39 accuracy points. For the
masked version, we witness an opposite trend: Der
pos-c information is much less recoverable in the
lower parts of the network than the upper parts. In
besondere, the removal of pos-c from the second

12As the representations used to train INLP do not include
BERTs’ special tokens (z.B., ‘CLS’, ‘SEP’), we also don’t
use the projection matrix on those tokens.

167

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Figur 3: Layer-wise removal. Removing from layer i (the rows) and testing probing performance on layer j (Die
columns). Top row (3A) is non-masked version, bottom row (3B) is masked.

layer appears to affect the rest of the layers, welche
do not manage to recover a high score on this task,
ranging from 32.7 Zu 42.1 accuracy.

For all of the non-masked experiments the upper
layers seem to make it harder for the subsequent
layers to extract
the property. In the masked
version however, there is no consistent trend.
It is harder to extract properties after the lower
parts for pos-c and ner. For phrase start the upper
part makes it harder for further extraction and
for phrase end both the lower and upper parts
make it harder, as opposed to the middle layers.
Further research is needed in order to understand
the significance of those findings, and whether or
not they are related to information usage across
layers.

This lead us to the final experiment where we
test for the main task performance after an amnesic
operation at the intermediate layers.

7.2 Re-rediscovering the NLP Pipeline

In the previous set of experiments, we measured
how much of the signal removed in layer i is
recovered in subsequent layers. We now study
how the removal of information in layer i affects
the word prediction accuracy at the final layer,
in order to get a complementary measure for
layer importance with respect to a property. Der
results for the different properties are presented
in Abbildung 4, where we plot the difference in word
prediction performance between the control and

the amnesic probing when removing a linguistic
property from a certain layer.

These results provide a clear interpretation
on the internal function of BERT’s layers. Für
the masked version (Figur 4), we observe that
the pos-c properties are mostly important
In
layer 3 and its surrounding layers, sowie
layer 12. Jedoch, this information is accurately
extractable only towards the last layers. For ner,
we observe that the main performance loss occurs
at layer 4. For phrase-markers the middle layers
are important: layers 5 Und 7 for phrase start
(although the absolute performance loss is not
big) and layer 6 for phrase end contributes the
most for the word prediction performance.

The story with the non-masked version is quite
anders (Figur 4). Erste, notice that the amnesic
operation improves the LM performance for all
properties, in some layers.13 Second, the drop in
performance peak across all properties is different
than the masked version experiments. Insbesondere,
it seems that for pos-c, when the words are non-
masked in the input, the most important layer
for pos-c is 11 (and not
layer 3, as in the
masked version), while this information is easily
extractable (by standard probing) across all layers
(über 80% accuracy).

Interessant, the conclusions we draw on layer-
importance from amnesic probing partly differ

13Giulianelli et al. (2018) observed a similar behavior by

performing an intervention on LSTM activations.

168

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Figur 4: The influence of the different properties, from each layer on LM predictions. Top figure (4A) zeigt die
results on the regular, non-masked version, bottom figure (4B) for the masked version. Colors allow ease of layer
comparison across graphs.

the last

trends are similar:

from the ones in the ‘‘Pipeline processing’’ hy-
pothesis (Tenney et al., 2019A), which aims to
localize and attribute information processing of
linguistic properties to parts of BERT (für die
non-masked version).14 On one hand the ner
Experiment
layers
are much more important than earlier ones (In
besondere, layer 11 is the most affected in our
Fall, with a decrease of 31.09 accuracy points.
Andererseits, in contrast to their hypotheses,
we find that POS information, pos-c (which was
considered to be more important in the earlier
layers) affects the word prediction performance
much more in the upper layers (40.99 accuracy
loss in the 11th layer). Endlich, we note that our
approach performs an ablation of these properties
in the representation space, which reveals which
layers are actually responsible for processing
properties, as opposed to Tenney et al. (2019A),
who focused on where this information is easily
extractable.

clearer distinctions between the two. Endlich, Wir
stress that the different experiments should not be
compared between one setting to the other, Und
thus the different y-scales in the figures. Das ist
due to confounding variables (z.B., the number
of removed dimensions from the representations),
which we do not control for in this work.

8 Related Work

With the established impressive performance of
large pre-trained language models (Devlin et al.,
2019; Liu et al., 2019B), based on the Transformer
architecture (Vaswani et al., 2017), a large body of
work started studying and gaining insight into how
these models work and what do they encode.15 For
a thorough summary of these advancements we
refer the reader to a recent primer on the subject
(Rogers et al., 2020).

We note the big differences in behavior when
analyzing the masked vs. the non-masked version
of BERT, and call for future work to make a

14We note that this work analyzes BERT-base, in contrast

to Tenney et al. (2019A) who analyzed BERT-Large.

15These works cover a wide variety of topics, einschließlich
grammatical generalization (Goldberg, 2019; Warstadt et al.,
2019), syntax (Tenney et al., 2019B; Lin et al., 2019; Reif
et al., 2019; Hewitt and Manning, 2019; Liu et al., 2019A),
world knowledge (Petroni et al., 2019; Jiang et al., 2020),
reasoning (Talmor et al., 2019), and common sense (Forbes
et al., 2019; Zhou et al., 2019; Weir et al., 2020).

169

A particularly popular and easy-to-use inter-
pretation method is probing (Conneau et al.,
2018). Despite its popularity, recent works have
questioned the use of probing as an interpretation
tool. Hewitt and Liang (2019) haben betont
the need to distinguish between decoding and
learning the probing tasks. They introduced con-
trol tasks, a consistent but linguistically mea-
ningless attribution of labels to tokens, and have
shown that probes trained on the control tasks
often perform well, due to the strong lexical infor-
mation held in the representations and learned by
the probe. This leads them to propose a selectivity
measure that aims to choose probes which achieve
high accuracy only on linguistically-meaningful
tasks. Tamkin et al. (2020) claim that probing
cannot serve as an explanation of downstream
task success. They observe that the probing scores
do not correlate with the transfer scores achieved
by fine-tuning.

Endlich, Ravichander et al. (2020) show that
probing can achieve non-trivial results for linguis-
tic properties that were not needed for the task the
model was trained on. In this work, we observe a
similar phenomenon, but from a different angle.
We actively remove some property of interest
from the queried representation, and measure
the impact of the amnesic representation of the
property on the main task.

Two recent works study the probing paradigm
from an information-theory perspective. Pimentel
et al. (2020) emphasize that under a mutual-
information maximization objective, ‘‘better’’ probes
are increasingly more accurate, regardless of
complexity. They use the data-processing in-
equality to question the rationale behind methods
that focus on encoding, and propose ease of ex-
tractability as an alternative criterion. Voita and
Titov (2020) follow this direction, using the
concept of minimum description length (MDL,
Rissanen, 1978) to quantify the total information
needed to transmit both the probing model and
the labels it predicts. Our discussion here is
somewhat orthogonal to those on the meaning
of encoding and probe complexity, as we focus on
the information influence on the model’s behavior,
rather than on the ability to extract it from the
representation.

Finally and concurrent to this work, Feder et al.
(2020) have studied a similar question of a causal
attribution of concepts to representations, verwenden
adversarial training guided by causal graphs.

9 Diskussion

(selectivity). Das

Intuitively, we would like to completely neutralize
the abstract property we are interested in—e.g.,
POS information (completeness), as represented
by the model—while keeping the rest of the
representation intact
ist ein
nontrivial goal, as it is not clear whether neural
models actually have abstract and disentangled
representations of properties such as POS, welche
are independent of other properties of the text.
It may be the case that the representation of
many properties is intertwined. In der Tat, there is
an ongoing debate on the assertion that certain
information is ‘‘encoded’’ in the representation
(Voita and Titov, 2020; Pimentel et al., 2020).
Jedoch, even if a disentangled representation of
the information we focus on exists, it is not clear
how to detect it.

We implement the information removal oper-
ation with INLP, which gives a first order approx-
imation using linear classifiers; we note, Jedoch,
that one can in principle use other approaches to
achieve the same goal. While we show that we do
remove the linear ability to predict the properties
and provide some evidence to the selectivity of
this method (§2), one has to bear in mind that we
remove only linearly-present information, und das
the classifiers can rely on arbitrary features that
happen to correlate with the gold label, be it a result
of spurious correlations or inherent encoding of the
direct property. In der Tat, we observe this behavior
in Section 7.1 (Figur 3), where we neutralize the
information from certain layers, but occasionally
observe higher probing accuracy in following
layers. We thus stress that the information we
remove in practice should be seen only as an
approximation for the abstract information we are
interested in, and that one has to be cautious of
causal interpretations of the results. Although in
this paper we use the INLP algorithm in order to
remove linear information, amnesic probing is not
restricted to removing linear information. Wann
non-linear removal methods become available,
they can be swapped instead of INLP. Das
stresses the importance of creating algorithms for
non-linear information removal.

Another unanswered question is how to quan-
tify the relative importance of different prop-
erties encoded in the representation for the word
prediction task. The different erasure portion
for different properties makes it hard to draw

170

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

conclusions on which property is more important
for the task of interest. Although we do not
make claims such as ‘‘dependency information is
more important than POS’’, these are interesting
questions that should be further discussed and
researched.

10 Conclusions

In this work, we propose a new method, Amnesic
Probing, which aims to quantify the influence of
specific properties on a model that is trained on a
task of interest. We demonstrate that conventional
probing falls short in answering such behavioral
Fragen, and perform a series of experiments
on different linguistic phenomenon, quantifying
their influence on the masked language modeling
Aufgabe. Außerdem, we inspect both unmasked
and masked BERT’s representation and detail
the differences between them, which we find to
be substantial. We also highlight the different
influence of specific fine-grained properties (z.B.,
nouns and determiners) on the final task. Endlich,
we use our proposed method on the different
layers of BERT, and study which parts of the
model make use of
the different properties.
Taken together, we argue that compared with
probing, counterfactual intervention—such as the
one we present here—can provide a richer and
more refined view of the way symbolic linguistic
information is encoded and used by neural models
with distributed representations.16

Danksagungen

Zu

wie

thank Hila Gonen,
We would
Amit Moryossef, Divyansh Kaushik, Abhilasha
Ravichander, Uri Shalit, Felix Kreuk, Jurica ˇSeva,
and Yonatan Belinkov for their helpful comments
and discussions. We also thank the anonymous
reviewers and the action editor, Radu Florian, für
their valuable suggestions.

This project has received funding from the
Europoean Research Council (ERC) under the
Europoean Union’s Horizon 2020 research and
innovation programme, grant agreement no.
802774 (iEXTRACT). Yanai Elazar is grateful
to be partially supported by the PBC fellowship
for outstanding PhD candidates in Data Science.

16All of the experiments were logged and tracked using

Weights and Biases (Biewald, 2020).

171

Verweise

Yossi Adi, Einat Kermany, Yonatan Belinkov,
Ofer Lavi, and Yoav Goldberg. 2016. Fine-
grained analysis of sentence embeddings using
auxiliary prediction tasks. CoRR, abs/1608
.04207.

Christoph Alt, Aleksandra Gabryszak,

Und
Leonhard Hennig. 2020. Probing linguistic fea-
tures of sentence-level representations in rela-
tion extraction. In Proceedings of
the 58th
Annual Meeting of the Association for Compu-
tational Linguistics, pages 1534–1545, Online.
Verein für Computerlinguistik.

Lukas Biewald. 2020. Experiment tracking with
weights and biases. Software available from
wandb.com.

Alexis Conneau, Germ´an Kruszewski, Guillaume
Lample, Lo¨ıc Barrault, and Marco Baroni.
2018. What you can cram into a single $&!#*
vector: Probing sentence embeddings for lin-
guistic properties. In Proceedings of the 56th
Annual Meeting of the Association for Compu-
tational Linguistics (Volumen 1: Long Papers),
pages 2126–2136. DOI: https://doi.org
/10.18653/v1/P18-1198

Jacob Devlin, Ming-Wei Chang, Kenton Lee,
and Kristina Toutanova. 2019. BERT: Pre-
training of deep bidirectional transformers for
In Proceedings of
language understanding.
Die 2019 Conference of the North American
Chapter of the Association for Computational
Linguistik: Human Language Technologies,
NAACL-HLT 2019, Minneapolis, MN, USA,
June 2–7, 2019, Volumen 1 (Long and Short
Papers), pages 4171–4186.

Yanai Elazar and Yoav Goldberg. 2018. Adversar-
ial removal of demographic attributes from text
Daten. In Proceedings of the 2018 Conference
on Empirical Methods in Natural Language
Processing, pages 11–21. Association for Com-
putational Linguistics. DOI: https://doi
.org/10.18653/v1/D18-1002

Amir Feder, Nadav Oved, Uri Shalit, Und
Roi Reichart. 2020. Causalm: Causal model
explanation through counterfactual
Sprache
Modelle.

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Maxwell Forbes, Ari Holtzman, and Yejin Choi.
2019. Do neural language representations learn
physical commonsense? Proceedings of
Die
41st Annual Conference of
the Cognitive
Science Society.

Yaroslav Ganin, Evgeniya Ustinova, Hana
Ajakan, Pascal Germain, Hugo Larochelle,
Franc¸ois Laviolette, Mario Marchand, Und
Victor Lempitsky. 2016. Domain-adversarial
training of neural networks. Journal of Machine
Learning Research, 17(1):2096–2030.

Mario Giulianelli, Jack Harding, Florian Mohnert,
Dieuwke Hupkes, and Willem Zuidema. 2018.
Under the hood: Using diagnostic classifiers
to investigate and improve how language mo-
dels track agreement information. In Proceed-
ings of the 2018 EMNLP Workshop Blackbox
NLP: Analyzing and Interpreting Neural Net-
works for NLP, pages 240–248. DOI: https://
doi.org/10.18653/v1/W18-5426

Yoav Goldberg. 2019. Assessing bert’s syntactic
abilities. arXiv preprint arXiv:1901.05287.

Yash Goyal, Uri Shalit, and Been Kim. 2019.
Explaining classifiers with causal concept effect
(cace). arXiv preprint arXiv:1907.07165.

John Hewitt and Percy Liang. 2019. Designing and
interpreting probes with control tasks. In Em-
pirical Methods in Natural Language Pro-
Abschließen (EMNLP). DOI: https://doi.org
/10.18653/v1/D19-1275

John Hewitt and Christopher D. Manning. 2019.
A structural probe for finding syntax in word
Darstellungen. In Proceedings of the Con-
ference of
the North American Chapter of
the Association for Computational Linguistics:
Human Language Technologies, NAACL-HLT,
pages 4129–4138.

Dieuwke Hupkes, Sara Veldhoen, and Willem
Zuidema. 2018. Visualisation and’diagnostic
classifiers’ reveal how recurrent and recursive
neural networks process hierarchical structure.
Journal of Artificial Intelligence Research,
61:907–926. DOI: https://doi.org/10
.1613/jair.1.11196

Zhengbao Jiang, Frank F Xu, Jun Araki, Und
Graham Neubig. 2020. How can we know what

language models know? Transactions of the
Verein für Computerlinguistik,
8:423–438. DOI: https://doi.org/10
.1162/tacl a 00324

Divyansh Kaushik, Eduard Hovy, and Zachary
Lipton. 2020. Learning the difference that
makes a difference with counterfactually-
augmented data. In International Conference
on Learning Representations.

Yongjie Lin, Yi Chern Tan, and Robert Frank.
2019. Open sesame: Getting inside berts lin-
Die
guistic knowledge.
2019 ACL Workshop BlackboxNLP: Analyzing
and Interpreting Neural Networks for NLP,
pages 241–253.

In Proceedings of

Nelson F. Liu, Matt Gardner, Yonatan Belinkov,
Matthew E. Peters, and Noah A. Schmied. 2019A.
Linguistic knowledge and transferability of
contextual representations. In Proceedings of
Die 2019 Conference of the North American
Chapter of
the Association for Computa-
tional Linguistics: Human Language Techno-
logies, Volumen 1 (Long and Short Papers),
pages 1073–1094.

Yinhan Liu, Myle Ott, Naman Goyal, Jingfei
Von, Mandar Joshi, Danqi Chen, Omer Levy,
Mike Lewis, Luke Zettlemoyer, and Veselin
Stoyanov. 2019B. Roberta: A robustly opti-
mized bert pretraining approach. arXiv preprint
arXiv:1907.11692.

Ryan McDonald,

Joakim Nivre, Yvonne
Quirmbach-Brundage, Yoav Goldberg, Dipanjan
Das, Kuzman Ganchev, Keith Hall, Slav
Petrov, Hao Zhang, Oscar T¨ackstr¨om, et al.
2013. Universal dependency annotation for
Die
multilingual parsing. In Proceedings of
51st Annual Meeting of the Association for
Computerlinguistik (Volumen 2: Short
Papers), pages 92–97.

Judea Pearl and Dana Mackenzie. 2018. The book
of why: the new science of cause and effect.
Basic Books.

Fabian Pedregosa, Ga¨el Varoquaux, Alexandre
Gramfort, Vincent Michel, Bertrand Thirion,
Peter
Olivier Grisel, Mathieu Blondel,
Prettenhofer, Ron Weiss, and Vincent Dubourg.
2011. Scikit-learn: Machine learning in python.

172

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Journal of Machine Learning Research, 12(Oct):
2825–2830.

Matthew E. Peters, Mark Neumann, Mohit Iyyer,
Matt Gardner, Christopher Clark, Kenton Lee,
and Luke Zettlemoyer. 2018. Deep contex-
In Proceed-
tualized word representations.
ings of NAACL-HLT, pages 2227–2237. DOI:
https://doi.org/10.18653/v1/N18
-1202

Fabio Petroni, Tim Rockt¨aschel, Sebastian Riedel,
Patrick Lewis, Anton Bakhtin, Yuxiang Wu,
and Alexander Miller. 2019. Language models
as knowledge bases? In Proceedings of
Die 2019 Conference on Empirical Methods
in Natural Language Processing and the 9th
International Joint Conference on Natural
Language
(EMNLP-IJCNLP),
pages 2463–2473. DOI: https://doi.org
/10.18653/v1/D19-1250

Processing

Tiago Pimentel, Josef Valvoda, Rowan Hall
Maudslay, Ran Zmigrod, Adina Williams, Und
Ryan Cotterell. 2020.
Information-theoretic
probing for linguistic structure. DOI: https://
doi.org/10.18653/v1/2020.acl-main
.420

Shauli Ravfogel, Yanai Elazar, Hila Gonen,
Michael Twiton, and Yoav Goldberg. 2020.
Null it out: Guarding protected attributes by
iterative nullspace projection. In Proceedings of
the 58th Annual Meeting of the Association for
Computerlinguistik, pages 7237–7256,
Online. Association for Computational Linguis-
Tics. DOI: https://doi.org/10.18653
/v1/2020.acl-main.647

Abhilasha Ravichander, Yonatan Belinkov, Und
Eduard Hovy. 2020. Probing the probing
paradigm: Does probing accuracy entail task
relevance? arXiv preprint arXiv:2005.00719.

Emily Reif, Ann Yuan, Martin Wattenberg,
Fernanda B. Viegas, Andy Coenen, Adam
Pearce, and Been Kim. 2019. Visualizing and
measuring the geometry of bert. In Advances
in Neural Information Processing Systems,
pages 8592–8600.

Jorma Rissanen. 1978. Modeling by shortest data
description. Automatica, 14(5):465–471. DOI:
https://doi.org/10.1016/0005-1098
(78)90005-5

Anna Rogers, Olga Kovaleva,

and Anna
Rumshisky. 2020. A primer in bertology: Was
we know about how Bert works. arXiv preprint
arXiv:2002.12327. DOI: https://doi.org
/10.1162/tacl a 00349

Alon Talmor, Yanai Elazar, Yoav Goldberg, Und
Jonathan Berant. 2019. olmpics – on what
language model pre-training captures. DOI:
https://doi.org/10.1162/tacl a
00342

Alex Tamkin, Trisha Singh, Davide Giovanardi,
Investigating
and Noah Goodman. 2020.
transferability in pretrained language models.
arXiv
arXiv:2004.14975. DOI:
https://doi.org/10.18653/v1/2020
.findings-emnlp.125

preprint

Ian Tenney, Dipanjan Das, and Ellie Pavlick.
2019A. BERT rediscovers the classical NLP
pipeline. In Proceedings of the Conference of
the Association for Computational Linguistics,
ACL, pages 4593–4601. DOI: https://doi
.org/10.18653/v1/P19-1452

Ian Tenney, Patrick Xia, Berlin Chen, Alex Wang,
Adam Poliak, R. Thomas McCoy, Najoung
Kim, Benjamin Van Durme, Sam Bowman,
Dipanjan Das, and Ellie Pavlick. 2019B.
What do you learn from context? probing
for sentence structure in contextualized word
Darstellungen. In International Conference on
Learning Representations.

Ashish Vaswani, Noam Shazeer, Niki Parmar,
Jakob Uszkoreit, Llion Jones, Aidan N.
Gomez, Łukasz Kaiser, and Illia Polosukhin.
2017. Attention is all you need. In Advances
in neural
information processing systems,
pages 5998–6008.

Jesse Vig, Sebastian Gehrmann, Yonatan
Belinkov, Sharon Qian, Daniel Nevo, Yaron
Singer, and Stuart Shieber. 2020. Causal
mediation analysis for interpreting neural NLP:
The case of gender bias. arXiv preprint
arXiv:2004.12265.

Elena Voita and Ivan Titov. 2020. Information-
theoretic probing with minimum description
Länge. arXiv preprint arXiv:2003.12298. DOI:
https://doi.org/10.18653/v1/2020
.emnlp-main.14

173

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Alex Warstadt, Yu Cao, Ioana Grosu, Wei Peng,
Hagen Blix, Yining Nie, Anna Alsop, Shikha
Bordia, Haokun Liu, Alicia Parrish, Sheng-Fu
Wang, Jason Phang, Anhad Mohananey, Phu
Mon Htut, Paloma Jeretiˇc, and Samuel R.
Bowman. 2019. Investigating BERTs knowl-
edge of language: Five analysis methods with
NPIs. In Proceedings of the 2019 Conference
on Empirical Methods in Natural Language
Processing and the 9th International Joint
Conference on Natural Language Processing
(EMNLP-IJCNLP), pages 2870–2880. DOI:
https://doi.org/10.18653/v1/D19
-1286

Nathaniel Weir, Adam Poliak, and Benjamin
Van Durme. 2020. Probing neural language
models for human tacit assumptions. In 42nd
Annual Virtual Meeting of theCognitive Science
Society (CogSci).

Ralph Weischedel, Martha Palmer, Mitchell
Marcus, Eduard Hovy, Sameer Pradhan, Lance
Ramshaw, Nianwen Xue, Ann Taylor, Jeff
Kaufman, Michelle Franchini, and Mohammed
El-Bachouti, Robert Belvin, and Ann Houston.
release 5.0 ldc2013t19.
2013. Ontonotes
Linguistic Data Consortium, Philadelphia,
PA, 23.

Thomas Wolf, Lysandre Debut, Victor Sanh,
Julien Chaumond, Clement Delangue, Anthony
Moi, Pierric Cistac, Tim Rault, R’emi Louf,
Morgan Funtowicz, and Jamie Brew. 2019.
Huggingface’s transformers: State-of-the-art
natural language processing. ArXiv, abs/1910
.03771.

Xuhui Zhou, Yue Zhang, Leyang Cui, and Dandan
Huang. 2019. Evaluating commonsense in
pre-trained language models. arXiv preprint
arXiv:1911.11931.

Appendix A

We provide additional experiments that depict
the performance of the main task (z.B., POS)
performance during the INLP iterations
In
Figur 5.

Figur 5: LM accuracy over INLP predictions, für die
masked tokens version. We present both the Vanilla
word-prediction score (straight, blue line), sowie
Amnesic Probing (orange, small circles), and the main
task performance (Rot, large circles). For reference we
also provide the vanilla probing performance of each
Aufgabe (Grün, cross marks). Note that the number of
removed dimensions per iteration differs, basierend auf
number of classes of that property.

Appendix B Hewitt and Liang’s
Control Task

Control Task (Hewitt and Liang, 2019) has been
suggested as a way to attribute the performance
of the probe to extraction of encoded information,
as opposed to lexical memorization. Our goal in
this work, Jedoch, is not to extract information
from the representation (as is done in conventional
probing) but to measure a behavioural outcome.
Since the control
task is solved by lexical
memorization, applying INLP on control task’s
classifiers erases lexical information (d.h., erases
the ability to distinguish between arbitrary words),
which is at the core of the LM objective and
which is highly correlated with many of the other
linguistic properties, such as POS. We argue that
even if we do see a significant drop in performance
with the control
this says little on the
validity of the results of removal of the linguistic
Eigentum (z.B., POS). Jedoch, for completeness,
we provide the results in Figure 6. As can be seen
from this figure, this control’s slope is smaller than

Aufgabe,

174

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

the one of the amnesic probing, suggesting that
those directions have less behavioral influence.
Jedoch, the slopes are steeper than the ‘Rand’
Experiment. This is due to the identity removal
of groups of words, due to the label shuffle, als
suggested in their setup. This is the reason we
believe this test is not adequate in our case, Und
why we provide other tests to control for our
method: Rand and Selectivity (§2.3).

Figur 6: LM accuracy over INLP predictions, für die
masked tokens version. We present both the Vanilla
word-prediction score (straight, blue line), sowie
as Amnesic Probing (orange, small circles), und das
control performance (orange, large circles). We also
provide the Control results for selectivity test, proposed
by Hewitt and Liang (2019) (Rot, crosses). Beachten Sie, dass
the number of removed dimensions per iteration differs,
based on the number of classes of that property.

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

e
D
u

/
T

A
C
l
/

l

A
R
T
ich
C
e

P
D

F
/

D
Ö

ich
/

.

1
0
1
1
6
2

/
T

l

A
C
_
A
_
0
0
3
5
9
1
9
2
4
1
8
9

/

/
T

l

A
C
_
A
_
0
0
3
5
9
P
D

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

175Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals image
Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals image
Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals image
Amnesic Probing: Behavioral Explanation with Amnesic Counterfactuals image

PDF Herunterladen