Multimodal Pretraining Unmasked: A Meta-Analysis and a Unified
Framework of Vision-and-Language BERTs
Emanuele Bugliarello Ryan Cotterell
Naoaki Okazaki Desmond Elliott
University of Copenhagen
University of Cambridge
ETH Z¨urich
Tokyo Institute of Technology
emanuele@di.ku.dk, rcotterell@inf.ethz.ch,
okazaki@c.titech.ac.jp, de@di.ku.dk
Abstrakt
Large-scale pretraining and task-specific fine-
tuning is now the standard methodology for
many tasks in computer vision and natural
language processing. Kürzlich, a multitude of
methods have been proposed for pretraining vi-
sion and language BERTs to tackle challenges
at the intersection of these two key areas of AI.
These models can be categorized into either
single-stream or dual-stream encoders. Wir
study the differences between these two cate-
gories, and show how they can be unified under
a single theoretical framework. We then con-
duct controlled experiments to discern the em-
pirical differences between five vision and
language BERTs. Our experiments show that
training data and hyperparameters are respon-
sible for most of the differences between the
reported results, but they also reveal that the
embedding layer plays a crucial role in these
massive models.
1
Einführung
Learning generic multimodal representations from
images paired with sentences is a fundamental step
towards a single interface for vision and language
(V&L) tasks. In pursuit of this goal, many pre-
trained V&L models have been proposed in the last
Jahr, inspired by the success of pretraining in both
computer vision (Sharif Razavian et al., 2014) Und
natural language processing (Devlin et al., 2019).
All of these V&L models extend BERT (Devlin
et al., 2019) to learn representations grounded
in both modalities. They can either be classified
als (ich) single-stream, where images and text are
jointly processed by a single encoder (z.B., Zhou
et al., 2020), oder (ii) dual-stream, where the in-
puts are encoded separately before being jointly
modelled (z.B., Tan and Bansal, 2019).
The differences in downstream performance
between single- and dual-stream models are cur-
978
rently unclear, with some papers claiming the su-
periority of one family over the other (Lu et al.,
2019; Chen et al., 2020), while others arguing that
it is hard to draw any conclusion (Qi et al., 2020).
The first goal of this paper is to understand
the mathematical differences between single- Und
dual-stream models. Our analysis leads to a unified
framework in which currently proposed architec-
tures, both single- and dual-stream, are particular
instances. We then implement several of the pro-
posed encoders within this framework to empir-
ically measure their differences in a controlled
Umfeld. We believe this comparative analy-
sis is crucial to better understand and guide future
research of massive models in this vibrant area of
AI, ensuring progress is not blurred by confounds.
Tatsächlich, there are many differences in the pro-
tocols used to train V&L BERTs. In order to
better understand these models, we conduct a se-
ries of controlled studies to investigate whether
differences in downstream performance is ex-
plained by: (ich) the amount of pretraining data
and the pretraining objectives (z.B., Figur 2); (ii)
the hyperparameters used to control the learning
Verfahren; (iii) the variance caused by random ini-
tialization when pretraining (z.B., Figur 1); (iv)
the variance due to fine-tuning multiple times on a
downstream task; (v) being single- or dual-stream
architectures; oder (vi) the choice of the embedding
layer.
Zusammenfassend, our contributions in this paper are:
• We introduce a unified mathematical frame-
work in which currently proposed V&L
BERTs are only a subset of the possibilities.
• We release code for VOLTA (Visiolinguistic
Transformer architectures),1 a PyTorch im-
plementation of this framework in order to
speed up research in multimodal pretraining.
1https://github.com/e-bug/volta.
Transactions of the Association for Computational Linguistics, Bd. 9, S. 978–994, 2021. https://doi.org/10.1162/tacl a 00408
Action Editor: Jacob Eisenstein. Submission batch: 12/20; Revision batch: 4/2021; Published 9/2021.
C(cid:2) 2021 Verein für Computerlinguistik. Distributed under a CC-BY 4.0 Lizenz.
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
2.1 Input Embeddings
Language Input All V&L BERTs adopt the ap-
proach of BERT: The input sequence is first to-
kenized into sub-word units (Wu et al., 2016;
Sennrich et al., 2016) and two special tokens
[CLS] Und [SEP] are added to generate the text se-
quence {[CLS], w1, . . . , wT , [SEP]}. The embed-
ding of each token is then given by the sum of
three learnable vectors, corresponding to its form,
position in the sequence, and segment (Devlin
et al., 2019). Zusätzlich, VL-BERT (Su et al.,
2020) also adds the visual feature of the entire
image to each token.
Vision Input Typically, visual inputs are also
very similar across all V&L BERTs. For a given
Bild, a pretrained object detector is used to ex-
tract regions of interest, representing salient image
Regionen. For each region, in addition to its feature
vector, the object detector also returns the spatial
location of its bounding box, which most V&L
BERTs encode in different ways, analogously to
the word position in the language modality. Während
most approaches present very similar ways to em-
bed spatial locations, VL-BERT relies on a more
complex geometry embedding and they are, In-
stead, missing in VISUALBERT (Li et al., 2019).
Some models also include a special feature [IMG]
that denotes the representation of the entire image
(z.B., a mean-pooled visual feature with a spatial
encoding corresponding to the full image). Fi-
schließlich, PIXEL-BERT (Huang et al., 2020) nicht
rely on an object detector but directly extracts a
set of visual embeddings from the raw image.
2.2 Encoders
Single-stream Encoders The majority of V&L
BERTs follow the single-stream paradigm (Su
et al., 2020; Li et al., 2019; Chen et al., 2020;
Li et al., 2020A; Zhou et al., 2020; Lin et al.,
2020; Li et al., 2020B). Hier, a standard BERT ar-
chitecture is given the concatenation of the visual
and linguistic features of an image–text pair as in-
put (Figure 3a). This design allows for an early and
unconstrained fusion of cross-modal information.
Dual-stream Encoders VILBERT (Lu et al.,
2019), LXMERT (Tan and Bansal, 2019), Und
ERNIE-VIL (Yu et al., 2021)3 are based on a
Figur 1: How does the amount of pretraining data
affect downstream performance of V&L BERTs? Wir
find that these models perform more similarly when
trained in the same conditions. This plot shows the
results from the papers (♦), and when each model is
pretrained 10 times on the Conceptual Captions dataset
and fine-tuned once on the NLVR2 verification task
(◦). The area of a marker is proportional to the amount
of pretraining data. The result from the VISUALBERT
paper is highlighted in a dashed box.
• We conduct a series of controlled studies2
finding that several models perform similarly
when trained under the same conditions.
• While we find that single- and dual-stream
families perform equally well, Leistung
can differ significantly between two models
and the embedding layer plays a key role.
• However, these V&L BERTs are sensitive
to weight initialization and state-of-the-art
claims should not be made from single runs.
2 Vision-and-Language BERTs
Given a sequence of tokens {w1, . . . , wT } und ein
set of visual features {v1, . . . , vK}, a shared goal
of V&L BERT models is to produce cross-modal
representations that are useful for downstream
tasks grounded in both modalities.
In diesem Abschnitt, we first review how these models
embed their inputs to the feature space. Nächste, Wir
discuss the main differences in the encoders and,
finally, highlight a variety of confounds that might
affect the performance achieved by these models.
2 https://github.com/e-bug/mpre-unmasked.
3ERNIE-VIL uses the dual-stream VILBERT encoder.
979
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
dual-stream paradigm. Hier, the visual and lin-
guistic features are first processed by two in-
dependent stacks of Transformer layers.4 The
resulting representations are then fed into cross-
modal Transformer layers where intra-modal in
teractions are alternated with inter-modal inter-
Aktionen (see Figure 3b and c). Interessant, beide
VILBERT and LXMERT modeled inter-modal
interactions in the same way: Each stream first
computes its query, key, and value matrices, Sei-
fore passing the keys and values to the other
modality. By doing so, these models explicitly
constrain interactions between modalities at each
layer, inhibiting some of the interactions that are
possible in a single-stream encoder while increas-
ing their expressive power by separate sets of
learnable parameters.
2.3 Pretraining Objectives
V&L BERTs are pretrained by jointly optimiz-
ing multiple different self-supervised objectives
over tokens and image regions through (weighted)
scalarization: L (θ) =
o λoLo(θ). Hier, θ
denotes a model’s parameters, Lo is the o-th
objective, and λo is its corresponding weight.
Commonly adopted objectives are of three types:
Sprache, vision, and cross-modal predictions.
(cid:2)
For language prediction, BERT’s denoising
masked language modeling (MLM) objective is
typically used. MLM replaces some tokens with
A [MASK] symbol, which are then predicted by
using bidirectional text context and image regions.
The MLM objective has been extended to image
regions via masked region modeling objectives.
These typically take the form of either object clas-
sification or feature regression, with some papers
showing benefits when modeling both (z.B., Chen
et al., 2020). Some models, such as LXMERT, Sind
also optimized over objects’ attributes prediction.
Endlich, interactions between the two modalities
are explicitly enforced by means of cross-modal
objectives. The typical task here is that of image–
text matching (ITM; z.B., Chen et al., 2020), welche
extends BERT’s next sentence prediction objec-
tive to V&L inputs: Given a sequence of tokens
and a set of image regions, the model is tasked to
predict whether the tokens describe the image.
4In der Praxis, VILBERT directly feeds the image represen-
tations obtained from the object detector, while LXMERT
further processes them through LV layers.
Figur 2: Comparison of proposed V&L BERTs on
VQAv2 (most common downstream task) as a function
of their pretraining data (size and type).
2.4 Further Distinctions
So far, we have given an overview of the core
components in V&L BERTs. Jedoch, es gibt
several implementation differences between them.
Zum Beispiel, LXMERT presents two main vari-
ations to the above description of dual-stream
Modelle. Erste, in its inter-modal layer, the param-
eters of the attention sub-layer are shared between
the two streams. This results in the model learning
a single function to contextualize image and text
inputs, regardless of which modality plays the role
of query or context. Zweite, its intra-modal layer
only consists of the multi-head attention block.
Darüber hinaus, a wider range of choices can affect
the performance of these models. From the object
detector used (and whether it is also fine-tuned
during pretraining), to the number of image re-
gions and the maximum text sequence length,
to the number of layers and their hidden sizes, Zu
pooling methods and fine-tuning MLP sizes, to the
use of text-only data, to optimization hyperparam-
eters (such as the number of pretraining epochs).
Another important distinction is the size and
type of pretraining data, which can affect task per-
Form (Figur 2). The size of pretraining da-
tasets ranges from 3M–10M image–text pairs,
over a range of pretraining tasks. The literature
distinguishes between ‘‘in-domain’’ and ‘‘out-of-
domain’’ data, each of which may consist of mul-
tiple datasets. An in-domain dataset overlaps with
common downstream tasks, Zum Beispiel, verwenden
VQAv2 (Goyal et al., 2017) as both a pretraining
task and a downstream task, while out-of-domain
datasets have no expected overlap, Zum Beispiel,
Conceptual Captions (Sharma et al., 2018).
980
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
3 A Unified Framework
In diesem Abschnitt, we unify the recently proposed
single-stream and dual-stream architectures under
the same mathematical framework. We start by
reviewing the Transformer layer, which forms the
core of these architectures, then we explain how
this layer has been adapted to encode multimodal
data in V&L BERTs, and introduce a gated bi-
modal Transformer layer that implements all of
the architecture variants as special cases.
3.1 Transformer Layers
Transformer-based architectures consist of a stack
of Transformer layers (Vaswani et al., 2017),
each typically having a multi-head attention block
(MAB) and a feed-forward block (FFB).
Multi-head Attention Block Given Nq query
vectors, each of dimension dq, Q ∈ RNq×dq , Und
Nv key–value pairs K ∈ RNv×dq , V ∈ RNv×dv ,
an attention function Att(Q, K, V) maps queries
to output vectors with a scaled dot-product:
MAB(X, Y) = LN(X + MHA(X, Y, Y)),
(3)
where LN is layer normalization (Ba et al., 2016).
Feed-forward Block For an input matrix M ∈
RN ×d, the feed-forward block is given by:
FFB(M) = LN(M + ReLU(MW1)W2),
(4)
where W1, W(cid:5)
2
∈ Rd×df f are learnable matrices.
Standard Transformer Layer Let X ∈ RN ×d
be an embedded input sequence, a standard Trans-
former layer performing self-attention is a param-
eterized function fθ : RN ×d → RN ×d such that:
fθ(X) = FFB(MAB(X, X)).
(5)
A stack of L Transformer layers that encodes an
input X, such as BERT, is then seen as a sequence
of L Transformer layers, each parametrized by θl:
Encoder(X) = fθL
◦ · · · ◦ fθ1(X).
(6)
Att(Q, K, V) = ω(QK(cid:5))V
(1)
3.2 Single-stream Multimodal Transformers
(cid:3)
where ω denotes a row-wise, scaled softmax:
dq). Hier, S = QK(cid:5) ∈
ωi(·) = softmax(·/
RNq×Nv is a score matrix that measures the simi-
larity between each pair of query and key vectors.
The output of Eq. (1) is a weighted sum of V,
in which a value gets higher weight if its corre-
sponding key has a larger dot product with the
query.
Multi-head attention (MHA) extends this func-
tion by first projecting Q, K, V into H different
matrices and computing the attention of each pro-
jection (Eq. (1)). These H different output vectors
are concatenated together ([(cid:6)]) and the concate-
nation is projected with a linear transformation
WO:
MHA(Q, K, V) = [O1 (cid:6) . . . (cid:6) OH ]WO,
(cid:5)
QWQ
where Oh = Att
H , QWV
H
H , KWK
(cid:4)
. (2)
H , WV
H
H , WK
Hier, {WQ
}H
h=1 and WO are learned
Parameter. Usually, dq = dv = d, WO ∈ Rd×d,
∈ Rd×da where da = d/H.
and WQ
Endlich, given inputs X, Y ∈ RN ×d, a multi-
H , WK
H , WV
H
Single-stream V&L BERTs extend BERT by con-
inputs XV ∈
catenating the embedded visual
RNV ×d and the embedded textual inputs XL ∈
RNL×d as a single input, hence the name ‘‘single-
stream’’ (Figure 3a). Speziell, X = [XL (cid:6)
XV ] ∈ RN ×d, where N = NL + NV , und das
attention is over both modalities (Figure 4a).
Somit, all single-stream models are of the type
defined in the previous section: Encoder(X). Der
various approaches only differ in the initial V&L
embeddings, the pretraining tasks, and the train-
ing data.
3.3 Dual-Stream Multimodal Transformers
Both VILBERT and LXMERT concurrently intro-
duced inter-modal and intra-modal layers.
Inter-modal Transformer Layer The inter-
modal layer explicitly models cross-modal in-
teraction via a cross-modal attention module.
Speziell, let M ∈ {L, V} denote either the lin-
guistic (L) or the visual (V) modality, and \M its
complementary one. The inter-modal multi-head
attention for modality M is given by (Figure 3c):
head attention block is defined as:
MM\M = MAB(XM, XM).
(7)
981
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Figur 3: Visualization of the (A) single-stream, (B) dual-stream intra-modal, Und (C) dual-stream inter-modal
Transformer layers. (D) shows our gated bimodal layer. The inter-modal layer attends across modalities, während die
intra-model layer attends within each modality. Ours can attend to either or both.
Figur 4: Visualization of the score matrix for (A) single-stream, (B) text–text, (C) vision–vision, (D) text–vision,
Und (e) vision–text interactions. Shades of green denote the text modality, while purple ones denote the vision
modality. Dual-stream scores are sub-matrices of the single-stream scores matrix.
Note that the second input to the multi-head at-
tention block (Eq. (3)) is taken from the comple-
mentary modality, which means the keys K and
values V in scaled dot-product attention (Eq. (1))
operate across modalities (see Figure 4d and
e). The remainder of this layer follows as from
Eq. (4).
Intra-modal Transformer Layer The intra-
modal layer, andererseits, is a Transformer
layer computing the attention of each modality in-
dependently (see Figure 3b). For a modality M:
MMM = MAB(XM, XM).
(8)
The rest of the layer follows as in Eq. (4)
for VILBERT, while there is no FFB block in
LXMERT.
3.4 Dual-stream Attentions as Restricted
Single-stream Attention
Das
Recall
in single-stream models the input
to a Transformer layer is the concatenation of
both modalities, X = [XL (cid:6) XV ]. daher,
in each single-stream attention head, the query
representation is given by:
(cid:7)
(cid:6)
·L
·V
Wo
are the language and visual sub-
matrices of the input and the resulting output. A
similar expression also holds for the keys K and
values V. We note that the score matrix S can be
defined in terms of four sub-matrices (Figure 4a):
S = QK(cid:5) =
=
=
(cid:6)
(cid:10)
(cid:7)
(cid:8)
QL
QV
QLK(cid:5)
QV K(cid:5)
(cid:6)
(cid:9)
K(cid:5)
L K(cid:5)
V
L QLK(cid:5)
V
L QV K(cid:5)
V
(cid:7)
SLL SLV
SVL SVV
(cid:11)
(10)
Recall from Eq. (1) that the attention matrix is a
normalised score matrix S, so each single-stream
layer computes both intra-modal (diagonal of
S) and inter-modal attention (anti-diagonal of
S). Mit anderen Worten, the dual-stream inter-modal
and intra-modal attention functions act as re-
stricted versions of the attention function in any
single-stream layer (siehe Abbildung 4).5 Ebenso wie-
Ergebnis, by interleaving inter- and intra-modal layers,
dual-stream models introduce an inductive bias
towards which interactions the model enforces in
each layer.
Q = XWQ =
(cid:7)
(cid:6)
XL
XV
WQ =
(cid:7)
(cid:6)
QL
QV
(9)
5Note that for this to be exact, the learnable parameters of
the MHA function need to be shared between modalities (als
done, Zum Beispiel, by LXMERT in its inter-modal blocks).
982
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
3.5 Gated Bimodal Transformer Layers
In the previous section, we showed that single-
stream attention blocks capture both the inter-
modal and intra-modal
interactions, separately
modeled by dual-stream architectures. We now in-
troduce a general gated bimodal Transformer layer
(Figure 3d), in which both single- and dual-stream
layers are special cases. By doing so, we can de-
fine existing V&L BERTs within a single architec-
tur, which allows us to implement and evaluate
several of these models in a controlled environ-
ment (see next sections). In addition to textual
XL and visual embeddings XV , this layer takes
a set of fixed binary variables {γ, τ } as part
its input: γ = {γLV , γVL, γLL, γVV }, Und
von
τ = {τM HA, τLN 1, τF F , τLN 2}. The γ values act
as gates that regulate the cross-modal interactions
within a layer, while the τ values control whether
the parameters are tied between modalities.
The main difference in our gated layer is in its
attention functions, originally defined in Eq. (1)
and Eq. (2). Hier, we extend them to bimodal in-
puts with controllable multimodal interactions as:
MHA(XL, XV ) = [O1 (cid:6) . . . (cid:6) OH ]
(cid:11)
(cid:10)
WO
L
WO
V
(11)
L and WO
V
where WO
are the language and
vision output matrices. The attention output
Att(Q, K, V), with a set of gating values γ is:
it is unaltered when γ is set to 0. By having a
sub-matrix that tends to −∞, we can effectively
compute the row-wise softmax (d.h., the attention)
over the other sub-matrix, hence recovering the
inter- and intra-modal attentions.6 This is similar
to the input masking applied in autoregressive
Transformer decoders (Vaswani et al., 2017).
V = WQ in each attention head).
This formulation allows us to control the de-
gree of inter- and intra-modal attention within a
layer, allowing us to define existing architectures
within a unified mathematical framework. We can
recover an inter-modal block (Eq. (7)) by setting
γLV = γVL = 0 and γLL = γVV = 1. Ähnlich,
the single-stream block (Eq. (3)) can be recovered
by setting γ = 0 and tying the learnable pa-
rameters (τ = 1) between the two streams (z.B.,
WQ
L = WQ
Außerdem, the gated bimodal Transformer
layer allows us to model a superset of the few
combinations considered thus far for cross-modal
fusion by multimodal transformer encoders. Eins
may explore asymmetric streams in which the two
modalities interact differently with the bimodal in-
puts, or explore different ways of interleaving con-
ventional single- and dual-stream blocks, or even
different levels of parameter sharing. For exam-
Bitte, asymmetric vision-and-language layers might
be beneficial for navigation (z.B., Hill et al.,
2021) or language-conditioned image generation
(z.B., Cho et al., 2020). An exploration of these
possibilities is left for future work.
(cid:11)
(cid:11)
; γ
4 Experimental Setup
(cid:10)(cid:10)
(cid:11)
O = Att
(cid:6)(cid:6)
= Att
XLWQ
L
XVWQ
V
(cid:6)
(cid:7)
QL
QV
(cid:6)
,
(cid:7)
(cid:11)
(cid:10)
,
(cid:7)
XLWK
L
XVWK
V
(cid:7)
(cid:6)
(cid:10)
(cid:7)
,
XLWV
L
XVWV
V
KL
KV
,
VL
VV
; γ
= ω (Sγ)
VL
VV
(12)
Recall from Eq. (10) that the score matrix
Sγ can be defined in terms of intra-modal and
inter-modal submatrices. Hier, the gating values
γ = {γLL, γLV , γVL, γVV } define the permit-
ted intra-modal and inter-modal interactions. Let
ε → −∞, Sγ is given by:
(cid:6)
Sγ =
εγLLSLL εγLV SLV
εγVLSVL εγVV SVV
(cid:7)
(13)
Das ist, when an attention gate γ is set to 1,
the corresponding sub-matrix tends to −∞, while
983
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
In diesem Abschnitt, we present the experimental setup
for our controlled studies on V&L encoders.
VOLTA In order to facilitate research and devel-
opment of V&L pretraining, we release VOLTA
(Visiolinguistic Transformer architectures), ein
implementation of our unified framework in Py-
Torch (Paszke et al., 2019). Our code is built on
top of the VILBERT-MT repository,7 based on
PyTorch-Transformers, due to its support to a wide
range of V&L tasks. We stress that it is important,
for this study, to have a unified implementation
that allows us to remove possible confounds due
6In der Praxis, our implementation is efficient and does not
evaluate sub-matrices whose corresponding gate is set to 1.
7https://github.com/facebookresearch
/vilbert-multi-task/.
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
to implementation details and effectively measure
differences given by the proposed architectures.
Implementation Details V&L BERTs typically
extract
image features using a Faster R-CNN
(Ren et al., 2015) trained on the Visual Genome
dataset (VG; Krishna et al. 2017), either with a
ResNet-101 (He et al., 2016) or a ResNeXT-152
backbone (Xie et al., 2017). The number of fea-
tures varies from 10 Zu 100. Our models are
trained with 36 regions of interest extracted by
a Faster R-CNN with a ResNet-101 backbone
(Anderson et al., 2018). Each model is initial-
ized with the parameters of BERT, following the
approaches described in the original papers.8 Ran-
domly initialized weights are initialized following
the standard approach in PyTorch-Transformers
(on which these models built on): Fully-connected
and embedding layers are initialized from a normal
distribution with mean 0.0 and standard deviation
0.02, bias vectors are initially set to 0.0, Und
the Layer Normalization weight vector to 1.0. Wir
train all models on 4 NVIDIA P100 GPUs and rely
on gradient accumulation to obtain larger batches
when needed. The parameter sets giving the best
validation performance based on the pretraining
objective are used for downstream tasks.
Pretraining As discussed in §2.4, V&L BERTs
have been pretrained on datasets of varying size
and type.9 In this paper, we pretrain all of our
models on the Conceptual Captions dataset (CC;
Sharma et al. 2018), which consists of 3.3M
images with weakly associated captions automat-
ically collected from billions of Web pages. Das
stands in contrast to other datasets, Zum Beispiel,
COCO (Lin et al., 2014) or VQA (Antol et al.,
2015), where the images are strongly associated
with crowdsourced captions or question–answer
pairs. The CC dataset is a good candidate for learn-
ing generic multimodal representations because of
its size, that it was scraped from the Web, und das
it has a broad coverage of subject matter.10 Note
that due to broken links, and a subsequent pruning
Phase, where images also found in the test sets of
Dataset
VQAv2
GQA
Image Source Train Test
Metric
COCO
COCO+Flickr
655K 448K VQA-score
1.1M 12.6K Accuracy
RefCOCO+ COCO
RefCOCOg COCO
120K 10.6K Accuracy
80K 9.6K Accuracy
NLVR2
SNLI-VE
Web Crawled
Flickr
86K
7K Accuracy
529K 17.9K Accuracy
COCO
Flirckr30k
COCO
Flickr
567K
145K
1K Recall@1
1K Recall@1
Tisch 1: Statistics of the downstream V&L tasks.
common V&L tasks11 are removed, we pretrain all
our models on 2.77M image–caption pairs from
Conceptual Captions.
Downstream Evaluation Tasks We consider
the most common tasks used to evaluate V&L
BERTs, spanning four groups: vocab-based VQA
(Goyal et al., 2017; Hudson and Manning, 2019),
image–text retrieval (Lin et al., 2014; Plummer
et al., 2015), referring expression (Kazemzadeh
et al., 2014; Mao et al., 2016), and multimodal
verification (Suhr et al., 2019; Xie et al., 2019).
See Table 1 for details.12 For each model, Die
parameter set giving the best performance in the
validation set was used for test.
5 Ergebnisse
We perform carefully controlled experiments to
investigate the possible reasons for the reported
difference in performance between V&L BERTs.
5.1 Unified Data and Reimplementation
We start by examining the performance of V&L
BERTs pretrained on the same 2.7M CC dataset.
Recall from Figure 2 that V&L BERTs have been
pretrained on different combinations of datasets,
which may explain most of the claimed differ-
ences in downstream task performance. Hier, Wir
evaluate three models with official released code:
VILBERT,13 LXMERT, and VL-BERT.
8Only Tan and Bansal (2019) reported slightly better
performance when pretraining from scratch but they relied on
large corpora of in-domain, human-annotated data.
9VL-BERT also adds text-only data to avoid overfitting
on short and simple sentences typical of V&L datasets.
10We also expect this type of dataset will be easier to
collect for low-resource languages in the future.
11The datasets listed in Table 1, Visual 7W (Zhu et al.,
2016), RefCOCO (Kazemzadeh et al., 2014), GuessWhat
(de Vries et al., 2017), and VCR (Zellers et al., 2019).
12Following previous work, accuracy in referring ex-
pression is evaluated on the region proposals of Yu et al.
(2018).
13VILBERT was trained as described in Lu et al. (2020).
984
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
els (◦) closely follows the official implementa-
tions in these downstream tasks, confirming the
correctness of our framework. Es gibt, Wie-
immer, some larger differences for some of the
tasks: In VQAv2, we now see that VILBERT
performs slightly worse than the other models
(contrarily to what we obtained with the official
Code), and in GQA, LXMERT closes the gap with
VILBERT. VILBERT’s performance on NLVR2
and COCO image retrieval increases by 2–3 points
in the VOLTA framework. As VOLTA is based on the
VILBERT code base, these differences might be
due to weight initialization, an hypothesis that we
test in later sections.
With this first study, we have seen that the per-
formance of these V&L BERTs is similar when
they are trained on the same data. Darüber hinaus, Wir
demonstrated the correctness of our implementa-
tions in VOLTA, in which these models are built
following the unified framework introduced in §3.
Trotzdem, there are still many possible con-
founds in the training procedures adopted by these
models that might interfere with a fair compari-
son of these architectures. In the next section, Wir
control these variables to unmask the true gains
introduced by a number of multimodal encoders.
5.2 Controlled Setup
We define a fixed set of hyperparameters to eval-
uate VILBERT, LXMERT, VL-BERT, VISUAL-
BERT, and UNITER on four downstream tasks:
VQAv2, RefCOCO+, NLVR2, and Flickr30K.
• Inputs: Each model used a different maxi-
mum number of tokens and LXMERT did not
have an overall [IMG] feature. We fix the
same maximum number of tokens and add
Die [IMG] feature to each architecture.
• Encoders: We noticed that VILBERT used
higher dimensional representations for the
visual stream. We fix the same dimension as
in the linguistic stream for a comparison that
is fairer comparison against LXMERT, Und
more intuitive with the single-stream models.
• Pooling: While VL-BERT is the only archi-
tecture that does not have a pooling layer,
other V&L BERTs use it for the image–text
matching objective. We fix the models to use
use multiplicative pooling (Lu et al., 2019)
for all the models in order to separately learn
Figur 5: Unified data and reimplementation results.
Performance of selected V&L BERTs on multiple tasks
from the original papers (♦), and when pretrained on
2.7M Conceptual Captions with their official code ((cid:3))
or in VOLTA (◦).
Same Data, Similar Performance Figure 5
shows the results of controlling the pretraining
data and pretraining tasks. The results from the
papers are reported (♦), alongside our training of
these models using the official code ((cid:3)). There is
a drop in performance for the models we trained
on the VQAv2, NLVR2, and image retrieval tasks,
compared to the performance reported in the pa-
pers. This is not surprising given that the (cid:3) Modelle
were pretrained on less data than the papers. Im Par-
besonders, given that VILBERT was also pretrained
on CC but with more image–text pairs, our results
corroborate previous studies showing diminishing
returns with pretraining data size (z.B., Lu et al.,
2019; Li et al., 2020A). Jedoch, the claimed
performance gaps between these models narrows
when they are pretrained on the same data. Für in-
Haltung, according to the literature, LXMERT was
clearly the best model in VQA tasks, welches ist
likely due to its use of large, in-domain data and a
VQA pretraining objective.14
VOLTA Implementation We also implemented
these models in VOLTA and trained them using their
official procedures and hyperparameters. Figur 5
shows that the performance of each of these mod-
14Surprisingly, for VQAv2, each of these models used
different proportions of the validation set during training.
In our experiments, stattdessen, we use the official training set,
which explains why the largest drops in performance are seen
Hier.
985
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Modell
VQAv2 RefCOCO+ NLVR2
testd
test-P
test-dev
Flickr30k
test IR test TR
ViLBERTBASE
LXMERT
VL-BERTBASE
VisualBERT
UNITERBASE
68.7
67.1
68.3
68.2
68.8
71.4
68.8
71.1
69.7
71.9
72.4
69.1
72.6
71.3
72.9
59.8
50.4
57.9
61.1
60.9
76.7
62.5
68.5
75.5
74.2
Tisch 2: Results with our controlled setup. Jede
model is pretrained using the VOLTA framework
with the same fixed hyperparameters on the 2.7M
CC dataset, and fine-tuned on downstream tasks.
sentence-level and image-level representa-
tions and also model their interactions.
• Pretraining Objectives: Each model uses
a different set of pretraining objectives. Wir
fix them to three: MLM, masked object clas-
sification with KL-divergence,15 and ITM.
• Fine-tuning: We fine-tune each model using
the same protocols and sizes for the MLPs.
• Hyperparameters: While VILBERT and
VL-BERT were originally pretrained for 10
Epochen, LXMERT was pretrained for 20. Wir
fix the number of pretraining epochs to 10,
and set other hyperparameters (z.B., learning
rate or its warm-up proportion) to a set of
values to randomness in initialization from
the original papers that led to smooth training
of all the models, with training curves that
closely followed the ones obtained with the
original hyperparameters.16
Results Table 2 shows the results of our con-
trolled study. Erste, we note that the performance
of VILBERT and VL-BERT is similar compared
to training with their original hyperparameters.
Tatsächlich, VQAv2 performance improves for VIL-
BERT, showing that dual-stream models do not
require different sizes in the two streams. VL-
BERT also performs similarly to its official setup,
showing that the additional ITM pretraining ob-
jective in our controlled setup does not hurt down-
stream task performance (contrarily to the results
reported in their paper). Das tun wir, Jedoch, note that
LXMERT performs worse on NLVR2 and VQAv2
15Chen et al. (2020) showed that this object classification
objective is the single best one for masked regions prediction.
16Configuration files of this setup are part of our repository.
in our controlled setup than with its original hyper-
Parameter, suggesting that LXMERT may require
more pretraining steps to converge. Gesamt, Die
results show that most of the examined models per-
form similarly in our controlled setup, verglichen
to the official setups.
5.3 Fine-tuning Variance
We now turn our attention to the effect of fine-
tuning variance on task performance. It has been
observed that the fine-tuning of BERT is sensitive
to randomness in initialization and data ordering
(Dodge et al., 2020). Hier, we investigate the sen-
sitivity of the five models used in the controlled
Studie. We fine-tune each model 10 times on the
RefCOCO+ and NLVR2 tasks by varying the seed.
This changes training data order and the weight
initialization of the classification layer. Figur 7
shows violin plots of the distribution of results,
in which the dots represent the experimental ob-
servations. We also report an average standard
deviation of 0.3 points for these models across
both tasks. Jedoch, the minimum and the maxi-
mum scores of a given model often differ by 1 oder
more points, showing how a single fine-tuning run
of these models can lead to incorrect conclusions.
5.4 Pretraining Variance
In the previous section, we found substantial vari-
ance in the performance of V&L BERTs across
10 fine-tuning runs. We now investigate if the
pretraining phase is similarly affected by different
runs. Hier, each model in our controlled setup
is pretrained 10 times and fine-tuned once on
four tasks: VQAv2, RefCOCO+, NLVR2, Und
Flickr30K image–text retrieval. By varying the
seed, we modify training data order as well as all
the layers that are not initialised from BERT (z.B.,
the visual embeddings, the masked object classi-
fication head and the ITM head in single-stream
Modelle). Figur 6 shows violin plots for each task.
We start by noting that our first pretraining run
(Tisch 2) of LXMERT was the worst one (its text
retrieval recall on Flickr30K is 10 points lower
than its mean). We also confirm that LXMERT
has slower convergence rate, with its task perfor-
mance after 10 epochs showing the largest vari-
ance among the V&L BERTs we tested. Auf der
andererseits, we find that some of these architectures
are less prone to variance caused by pretraining
986
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
T
l
A
C
_
A
_
0
0
4
0
8
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Figur 6: Pretraining variance of V&L BERTs. Each model is pretrained 10 times and fine-tuned once.
Figur 7: Fine-tuning variance of V&L BERTs on
RefCOCO+ and NLVR2. Each model is pretrained
once and fine-tuned 10 times on each task.
Figur 8: Variance of V&L BERTs on the Constrastive
Set of NLVR2, when each model is pretrained 10
times and fine-tuned once (A), or pretrained once and
fine-tuned 10 mal (B).
seed, such as VILBERT for VQA and retrieval
tasks, and UNITER for referring expression. Nev-
ertheless, the performance of all of these models
can vary by more than 1 point in several tasks
solely due to random initialization.
5.5 Evaluating Local Decision Boundaries
Previous work has shown that state-of-the-art sys-
tems can exploit systematic gaps in the data to
learn simple decision rules that let them achieve
high performance on test data (Gururangan et al.,
2018; Geva et al., 2019; Ribeiro et al., 2019).
In an effort to more accurately estimate model
Leistung, Gardner et al. (2020) proposed con-
trast sets: datasets in which existing test instances
have small but label-changing modifications in
order to characterize the correct decision bound-
ary near them. Figur 8 shows the performance
of our analyzed models on the NLVR2 contrast
set. Similar to Gardner et al. (2020), we see that
LXMERT loses around 15 points when evaluated
on perturbed samples. Außerdem, models that
performed much better on the standard test set now
achieve comparable performance to LXMERT,
showing that they exploited systematic gaps. Das
Ist, all of these V&L BERTs would perform sim-
ilarly when evaluated on out-of-distribution data.
5.6 Single- or Dual-stream Architectures
One of the key design choices that distinguishes
V&L BERTs is the number of ‘‘streams’’ used by
the encoder to process visual and linguistic inputs.
Lu et al. (2019) showed how their single-stream
baseline performed worse than their dual-stream
VILBERT architecture, while Chen et al. (2020)
claimed single-stream UNITER outperformed VIL-
BERT. Our controlled study across several tasks
and different pretraining initializations allows us
to provide an answer grounded with statistical
tests. To do so, we split the models in dual- Und
single-stream architectures17 and run a one-way
ANOVA (Tisch 3). After Bonferroni correction,
we only find statistical difference at p < 0.005
(Benjamin et al., 2018) between these two groups
for the Flickr30K text retrieval task.
17We only consider VILBERT for dual-stream encoders
due to LXMERT’s sub-optimal performance in our setup.
987
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Dataset
Single/Dual Stream V&L BERTs
F-test
p-value F-test
11.40
VQAv2
0.10
RefCOCO+
8.28
NLVR2
9.64
Flickr30k IR
Flickr30k TR 31.14
p-value
12.75 8.0e-06∗
1.7e-03
7.6e-01 111.61 2.7e-18∗
13.41 5.0e-06∗
6.5e-03
13.27 5.0e-06∗
3.6e-03
2.0e-06∗ 29.74 7.5e-10∗
Table 3: ANOVA between single- and dual-stream
architectures (left) and between all the tested V&L
BERTs (right). ∗ denotes significant results at
p < 0.005 after Bonferroni correction.
On the other hand, running the same test among
the various V&L BERTs, without grouping them
as single- or dual-stream architectures, returns sta-
tistical significance in each task (Table 3). This
table tells us that the null hypothesis, the mod-
els have the same average performance, does not
hold. However, it does not allow us to discern
where statistical differences lie. To do so, we con-
duct a post-hoc exact test at significance level
p < 0.005. Figure 9 shows the corresponding
pairwise p-values and highlights significant dif-
ferences between any two models after Bonferroni
correction. For instance, VILBERT is significantly
different compared to all other models in text
retrieval on Flickr30k, while UNITER is signifi-
cantly different on RefCOCO+.
5.7 The Importance of the Embeddings
Finally, our controlled setup leads us to an in-
teresting finding: The embedding layer (§2.1)
plays a crucial role in the final performance of
V&L BERTs. In fact, the only difference among
VL-BERT, VISUALBERT, and UNITER in our
setup is their embedding layer. Figure 6 and
Figure 7 show that this can have a drastic im-
pact on the downstream performance, although
the literature has given little attention to this de-
tail. For instance, Chen et al. (2020) claim that
the main contribution of UNITER is the set of
pretraining tasks, while our results, wherein all
the models are trained on the same pretraining
tasks, highlight that their embedding layer is an
important confound on final performance. Inter-
estingly, VISUALBERT is the only model that does
not encode the locations of regions of interest in
to considerably
its embeddings. This leads it
lower performance on RefCOCO+, showing that
this information is extremely useful for this task.
988
Given this result, we conduct one additional
experiment to see whether the embedding layer
biased our conclusion for dual- and single-stream
performance. To test this, we swap the embed-
ding layers of VILBERT (best dual-stream) and
UNITER (overall better single-stream) with each
other, which we pretrain and fine-tune once
(Figure 10). Similar to our previous results, em-
beddings are especially important for the tasks of
referring expression and retrieval. However, no
single embedding layer performs better, corrob-
orating that dual- and single-stream architectures
perform on par and showing that different em-
bedding strategies are necessary to maximise
performance in these two families of V&L BERTs.
5.8 Limitations
All the experiments in this paper are limited to
models that use a specific type of pretrained and
frozen visual encoder. While most V&L BERTs
follow this paradigm, some studies find beneficial
to jointly learn the visual encoder with language
(Su et al., 2020; Huang et al., 2020; Radford et al.,
2021; Kim et al., 2021). In addition, we only con-
sider base architecture variants (initialized with
BERTBASE) and pretrained on CC. Studying the
effects of visual encoders, pretraining data and
larger models is left as future work.
Although we expect longer pretraining would be
beneficial for every model, in our controlled setup,
we pretrain each model for 10 epochs to reduce
resource consumption. Here, we also constrain our
hyperparameter search over a small grid of values
that have been used in the literature. Finally, we
leave a thorough, controlled study of the various
pretraining objectives to future work.
6 Reproducibility and the Environment
From the perspective of reproducible research,
there are several advantages to using the VOLTA
framework for V&L encoders. First, VOLTA re-
duces confounds due to differences in implemen-
tations, while also enabling fair comparisons with
related work. Second, visual and textual data only
need to be preprocessed once instead of creating
model-specific formats for every V&L BERT.
From a financial perspective, the costs involved
in pretraining hampers contributions from many
academic institutions and deters the evaluation
of multiple trained models, which we showed
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Figure 9: Exact test between any two V&L BERTs. Each box shows the p-value for the corresponding pair of
models. Green boxes denote statistical significance at 0.005 after Bonferroni correction. Boxes are dark green if
the model in the y-axis outperforms the one in the x-axis, and vice versa for light green.
posed V&L BERTs can be specified as special
cases. We conducted a series of controlled studies
within this framework to better understand the
differences between several models. We found
that the performance of the considered models
varies significantly due to random initialization,
in both pretraining and fine-tuning. We also found
that these models achieve similar performance
when trained with the same hyperparameters and
data. Notably, some models outperform others
but we found that (a) single- and dual-stream
model families are on par, and (b) embedding
layers play a crucial role towards a model’s final
performance.
Our fast-paced field rewards the contribution of
new methods and state-of-the-art results (Rogers
and Augenstein, 2020), which often contrasts
with controlled comparisons and training multi-
ple models for variance estimation. In this paper,
we showed that several methods for vision-and-
language representation learning do not signif-
icantly differ when compared in a controlled
setting. This finding echoes similar studies of
variants of LSTMs (Greff et al., 2017) and Trans-
formers (Narang et al., 2021) that are not signifi-
cantly better than the original models. Looking to
the future, we recommend that new V&L BERTs
are pretrained on similar datasets, and that re-
searchers report fine-tuning variance, in addition
to their best performing model. We hope that our
findings will encourage more controlled evalua-
tions of newly proposed architectures for vision-
and-language and beyond.
Acknowledgments
Figure 10: Results of swapping VILBERT and UNITER
embeddings ((cid:8)) compared to their performance when
pretrained 10 times (box plots).
to be extremely important for V&L BERTs. We
estimate that pretraining a single model 10× in our
controlled setup for 4 downstream tasks requires a
4-GPU machine on AWS for two months, at a cost
of ∼$6,000, corresponding to 200 GPU-compute
days. Fortunately, we had access to an internal
server, but our experiments still required 1,500
GPU days for training and evaluation. While we
were able to reduce the financial costs, there are
severe environmental and carbon footprint costs
in V&L pretraining (Strubell et al., 2019).18
We hope that VOLTA will serve as a basis for re-
search in V&L pretraining, enabling easy and fair
comparisons across architectures, and ensuring
that progress is not obfuscated by confounds.
7 Conclusion
We introduced and implemented a unified math-
ematical framework, under which recently pro-
18We distribute many of our pretrained V&L BERTs in
VOLTA to amortise the environmental costs.
We are grateful to the action editor Jacob
Eisenstein and the anonymous reviewers at TACL
for
their constructive comments and discus-
sions. This project has received funding from the
989
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
2020
research
European Union’s Horizon
the Marie
and innovation programme under
Skłodowska-Curie grant agreement no. 801199
and by ‘‘Research and Development of Deep
Learning Technology for Advanced Multilingual
the Commissioned Re-
Speech Translation,’’
search of National Institute of Information and
Communications Technology (NICT), Japan.
References
Peter Anderson, Xiaodong He, Chris Buehler,
Damien Teney, Mark Johnson, Stephen Gould,
and Lei Zhang. 2018. Bottom-up and top-down
attention for image captioning and visual ques-
tion answering. In Proceedings of the IEEE/
CVF Conference on Computer Vision and Pat-
tern Recognition (CVPR), pages 6077–6086.
https://doi.org/10.1109/CVPR
.2018.00636
Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu,
Margaret Mitchell, Dhruv Batra, C. Lawrence
Zitnick, and Devi Parikh. 2015. VQA: Vi-
sual Question Answering. In Proceedings of
the IEEE/CVF International Conference on
Computer Vision (ICCV), pages 2425–2433.
https://doi.org/10.1109/ICCV
.2015.279
Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E.
Hinton. 2016. Layer normalization. arXiv pre-
print arXiv:1607.06450.
Brian A. Nosek,
Daniel J. Benjamin, James O. Berger, Magnus
E.-J.
Johannesson,
Wagenmakers, Richard Berk, Kenneth A.
Bollen, Bj¨orn Brembs, Lawrence Brown, Colin
Camerer, David Cesarini, Christopher D.
Chambers, Merlise Clyde, Thomas D. Cook,
Paul De Boeck, Zoltan Dienes, Anna Dreber,
Kenny Easwaran, Charles Efferson, Ernst Fehr,
Fiona Fidler, Andy P. Field, Malcolm Forster,
I. George, Richard Gonzalez,
Edward
Steven Goodman, Edwin Green, Donald P.
Green, Anthony G. Greenwald, Jarrod D.
Hadfield, Larry V. Hedges, Leonhard Held,
Teck Hua Ho, Herbert Hoijtink, Daniel J.
Hruschka, Kosuke
Imai, Guido Imbens,
John P. A. Ioannidis, Minjeong Jeon, James
Holland Jones, Michael Kirchler, David
Laibson, John List, Roderick Little, Arthur
Lupia, Edouard Machery, Scott E. Maxwell,
Savalei,
Michael McCarthy, Don A. Moore, Stephen L.
Morgan, Marcus Munaf´o, Shinichi Nakagawa,
Brendan Nyhan, Timothy H. Parker, Luis
Pericchi, Marco Perugini, Jeff Rouder, Judith
Rousseau, Victoria
Felix D.
Sch¨onbrodt, Thomas Sellke, Betsy Sinclair,
Dustin Tingley, Trisha Van Zandt, Simine
Vazire, Duncan J. Watts, Christopher Winship,
Robert L. Wolpert, Yu Xie, Cristobal Young,
Jonathan Zinman, and Valen E. Johnson. 2018.
Redefine statistical significance. Nature Human
Behaviour, 2(1):6–10. https://doi.org
/10.1038/s41562-017-0189-z, Pubmed:
30980045
Yen-Chun Chen, Linjie Li, Licheng Yu, Ahmed El
Kholy, Faisal Ahmed, Zhe Gan, Yu Cheng, and
Jingjing Liu. 2020. Uniter: Universal image-
text representation learning. In European Con-
ference on Computer Vision, pages 104–120.
https://doi.org/10.1007
Springer.
/978-3-030-58577-8_7
Jiasen Lu, Dustin Schwenk,
Jaemin Cho,
Hannaneh
Aniruddha
and
Hajishirzi,
Kembhavi. 2020. X-LXMERT: Paint, Caption
and Answer Questions with Multi-Modal
Transformers. In Proceedings of the 2020 Con-
ference on Empirical Methods in Natural Lan-
guage Processing (EMNLP), pages 8785–8805,
Online. Association for Computational Linguis-
tics. https://doi.org/10.18653/v1
/2020.emnlp-main.707
Harm de Vries, Florian Strub, Sarath Chandar,
Olivier Pietquin, Hugo Larochelle, and Aaron
Courville. 2017. GuessWhat?! Visual object
discovery through multi-modal dialogue. In
Proceedings of the IEEE Conference on Com-
puter Vision and Pattern Recognition (CVPR),
pages 4466–4475.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and
Kristina Toutanova. 2019. BERT: Pre-training
of deep bidirectional transformers for language
understanding. In Proceedings of
the 2019
Conference of the North American Chapter
of the Association for Computational Linguis-
tics: Human Language Technologies, Volume 1
(Long and Short Papers), pages 4171–4186,
Minneapolis, Minnesota. Association for Com-
putational Linguistics.
990
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Jesse Dodge, Gabriel Ilharco, Roy Schwartz, Ali
Farhadi, Hannaneh Hajishirzi, and Noah Smith.
2020. Fine-tuning pretrained language models:
Weight initializations, data orders, and early
stopping. arXiv preprint arXiv:2002.06305.
Matt Gardner, Yoav Artzi, Victoria Basmov,
Jonathan Berant, Ben Bogin, Sihao Chen,
Pradeep Dasigi, Dheeru Dua, Yanai Elazar,
Ananth Gottumukkala, Nitish Gupta, Hannaneh
Hajishirzi, Gabriel Ilharco, Daniel Khashabi,
Kevin Lin, Jiangming Liu, Nelson F. Liu,
Phoebe Mulcaire, Qiang Ning, Sameer Singh,
Noah A. Smith, Sanjay Subramanian, Reut
Tsarfaty, Eric Wallace, Ally Zhang, and Ben
Zhou. 2020. Evaluating models’ local decision
boundaries via contrast sets. In Findings of
the Association for Computational Linguistics:
EMNLP 2020, pages 1307–1323, Online.
Association for Computational Linguistics.
https://doi.org/10.18653/v1/2020
.findings-emnlp.117
Mor Geva, Yoav Goldberg, and Jonathan Berant.
2019. Are we modeling the task or the annota-
tor? An investigation of annotator bias in natural
language understanding datasets. In Proceed-
the 2019 Conference on Empirical
ings of
Methods in Natural Language Processing and
the 9th International Joint Conference on Nat-
ural Language Processing (EMNLP-IJCNLP),
pages 1161–1166, Hong Kong, China. Associa-
tion for Computational Linguistics. https://
doi.org/10.18653/v1/D19-1107
Yash Goyal, Tejas Khot, Douglas Summers-Stay,
Dhruv Batra, and Devi Parikh. 2017. Making
the V in VQA matter: Elevating the role of
image understanding in Visual Question Ans-
the IEEE/CVF
wering.
Conference on Computer Vision and Pattern
Recognition
6325–6334.
https://doi.org/10.1109/CVPR.2017
.670
In Proceedings of
(CVPR),
pages
Klaus Greff, Rupesh K. Srivastava, Jan Koutn´ık,
Bas R. Steunebrink, and J¨urgen Schmidhuber.
2017. LSTM: A search space odyssey. IEEE
Transactions on Neural Networks and Learning
Systems, 28(10):2222–2232. https://doi
.org/10.1109/TNNLS.2016.2582924,
Pubmed: 27411231
991
Suchin Gururangan, Swabha Swayamdipta, Omer
Levy, Roy Schwartz, Samuel Bowman, and
Noah A. Smith. 2018. Annotation artifacts in
natural language inference data. In Proceedings
of the 2018 Conference of the North American
Chapter of the Association for Computational
Linguistics: Human Language Technologies,
Volume 2 (Short Papers), pages 107–112,
New Orleans, Louisiana. Association for Com-
putational Linguistics. https://doi.org
/10.18653/v1/N18-2017
Kaiming He, Xiangyu Zhang, Shaoqing Ren,
and Jian Sun. 2016. Deep residual learning
for image recognition. In Proceedings of the
IEEE/CVF Conference on Computer Vision and
Pattern Recognition (CVPR), pages 770–778.
Felix Hill, Olivier Tieleman, Tamara von Glehn,
Nathaniel Wong, Hamza Merzic, and Stephen
Clark. 2021. Grounded language learning fast
In International Conference on
and slow.
Learning Representations.
Zhicheng Huang, Zhaoyang Zeng, Bei Liu,
Dongmei Fu, and Jianlong Fu. 2020. Pixel-bert:
Aligning image pixels with text by deep
preprint
multi-modal
arXiv:2004.00849.
transformers.
arXiv
Drew A. Hudson and Christopher D. Manning.
2019. GQA: A new dataset for real-world
visual reasoning and compositional question
answering. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pat-
tern Recognition (CVPR), pages 6700–6709.
https://doi.org/10.1109/CVPR.2019
.00686
Sahar Kazemzadeh, Vicente Ordonez, Mark
Matten, and Tamara Berg. 2014. ReferItGame:
Referring to objects in photographs of natural
scenes. In Proceedings of the 2014 Confer-
ence on Empirical Methods in Natural Lan-
guage Processing (EMNLP), pages 787–798,
Doha, Qatar. Association for Computational
Linguistics. https://doi.org/10.3115
/v1/D14-1086
Wonjae Kim, Bokyung Son, and Ildoo Kim. 2021.
VILT: Vision-and-language transformer with-
out convolution or region supervision. arXiv
preprint arXiv:2102.03334.
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin
Johnson, Kenji Hata, Joshua Kravitz, Stephanie
Chen, Yannis Kalantidis, Li-Jia Li, David A.
Shamma, Michael S. Bernstein and Li Fei-Fe.
2017. Visual genome: Connecting language
and vision using crowdsourced dense image
annotations. International Journal of Com-
puter Vision, 123(1):32–73. https://doi
.org/10.1007/s11263-016-0981-7
Gen Li, Nan Duan, Yuejian Fang, Ming Gong, and
Daxin Jiang. 2020a. Unicoder-VL: A univer-
sal encoder for vision and language by cross-
modal pre-training. Proceedings of the AAAI
Conference on Artificial Intelligence, 34(07):
11336–11344. https://doi.org/10.1609
/aaai.v34i07.6795
Liunian Harold Li, Mark Yatskar, Da Yin, Cho-Jui
Hsieh, and Kai-Wei Chang. 2019. VisualBERT:
A simple and performant baseline for vision and
language. arXiv preprint arXiv:1908.03557.
Xiujun Li, Xi Yin, Chunyuan Li, Pengchuan
Zhang, Xiaowei Hu, Lei Zhang, Lijuan Wang,
Houdong Hu, Li Dong, Furu Wei, et al. 2020b.
Oscar: Object-semantics aligned pre-training
for vision-language tasks. In European Con-
ference on Computer Vision, pages 121–137.
Springer. https://doi.org/10.1007/978
-3-030-58577-8 8
Jingren Zhou,
Junyang Lin, An Yang, Yichang Zhang, Jie
Liu,
and Hongxia Yang.
2020. Interbert: Vision-and-language interac-
tion for multi-modal pretraining. arXiv preprint
arXiv:2003.13198.
Tsung-Yi Lin, Michael Maire, Serge Belongie,
James Hays, Pietro Perona, Deva Ramanan,
Piotr Doll´ar, and C. Lawrence Zitnick. 2014.
Microsoft COCO: Common objects in con-
text. In European Conference on Computer
Vision, pages 740–755, Cham. Springer.
https://doi.org/10.1007/978-3-319
-10602-1 48
Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan
Lee. 2019. VilBERT: Pretraining task-agnostic
visiolinguistic representations for vision-and-
language tasks. In Advances in Neural Informa-
tion Processing Systems, pages 13–23. Curran
Associates, Inc.
Jiasen Lu, Vedanuj Goswami, Marcus Rohrbach,
Devi Parikh, and Stefan Lee. 2020. 12-in-1:
Multi-task vision and language representation
learning. In Proceedings of
the IEEE/CVF
Conference on Computer Vision and Pattern
Recognition (CVPR), pages 10434–10443.
J. Mao, J. Huang, A. Toshev, O. Camburu,
A. Yuille, and K. Murphy. 2016. Generation
and comprehension of unambiguous object de-
scriptions. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern
Recognition (CVPR), pages 11–20. https://
doi.org/10.1109/CVPR.2016.9
Sharan Narang, Hyung Won Chung, Yi Tay,
William Fedus, Thibault Fevry, Michael
Matena, Karishma Malkan, Noah Fiedel, Noam
Shazeer, Zhenzhong Lan, Yanqi Zhou, Wei
Li, Nan Ding, Jake Marcus, Adam Roberts,
and Colin Raffel. 2021. Do transformer mod-
ifications transfer across implementations and
applications? arXiv preprint arXiv:2102.11972.
Adam Paszke, Sam Gross, Francisco Massa,
Adam Lerer,
James Bradbury, Gregory
Chanan, Trevor Killeen, Zeming Lin, Natalia
Gimelshein, Luca Antiga, Alban Desmaison,
Andreas Kopf, Edward Yang, Zachary DeVito,
Martin Raison, Alykhan Tejani, Sasank
Chilamkurthy, Benoit Steiner, Lu Fang, Junjie
Bai, and Soumith Chintala. 2019. PyTorch: An
imperative style, high-performance deep learn-
ing library. In H. Wallach, H. Larochelle, A.
Beygelzimer, F. d’Alch´e-Buc, E. Fox, and R.
Garnett, editors, Advances in Neural Infor-
mation Processing Systems, pages 8024–8035,
Curran Associates, Inc.
Bryan A. Plummer, Liwei Wang, Chris M.
Cervantes, Juan C. Caicedo, Julia Hockenmaier,
and Svetlana Lazebnik. 2015. Flickr30k en-
tities: Collecting region-to-phrase correspon-
dences for richer image-to-sentence models.
In Proceedings of
the IEEE/CVF Interna-
tional Conference on Computer Vision (ICCV),
pages 2641–2649, USA. IEEE Computer Soci-
ety. https://doi.org/10.1109/ICCV
.2015.303
Di Qi, Lin Su, Jia Song, Edward Cui, Taroon
Bharti, and Arun Sacheti. 2020. ImageBERT:
Cross-modal pre-training with large-scale
992
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
weak-supervised image-text data. arXiv pre-
print arXiv:2001.07966.
Alec Radford, Jong Wook Kim, Chris Hallacy,
Aditya Ramesh, Gabriel Goh, Sandhini
Agarwal, Girish Sastry, Amanda Askell,
Pamela Mishkin, Jack Clark, Gretchen Krueger,
and Ilya Sutskever. 2021. Learning transferable
visual models from natural language supervi-
sion. arXiv preprint arXiv:2103.00020.
Shaoqing Ren, Kaiming He, Ross Girshick,
and Jian Sun. 2015. Faster R-CNN: Towards
real-time object detection with region proposal
networks. In C. Cortes, N. D. Lawrence, D. D.
Lee, M. Sugiyama, and R. Garnett, editors,
Advances in Neural Information Processing
Systems, pages 91–99. Curran Associates, Inc.
Marco Tulio Ribeiro, Carlos Guestrin, and Sameer
Singh. 2019. Are red roses red? Evaluating
consistency of question-answering models. In
Proceedings of the 57th Annual Meeting of
the Association for Computational Linguistics,
pages 6174–6184, Florence, Italy. Association
for Computational Linguistics. https://
doi.org/10.18653/v1/P19-1621
Anna Rogers and Isabelle Augenstein. 2020. What
can we do to improve peer review in NLP? In
Findings of the Association for Computational
Linguistics: EMNLP 2020, pages 1256–1262,
Online. Association for Computational Linguis-
tics. https://doi.org/10.18653/v1
/2020.findings-emnlp.112
Rico Sennrich, Barry Haddow, and Alexandra
Birch. 2016. Neural machine translation of rare
words with subword units. In Proceedings of
the 54th Annual Meeting of the Association for
Computational Linguistics (Volume 1: Long
Papers), pages 1715–1725, Berlin, Germany.
Association for Computational Linguistics.
https://doi.org/10.18653/v1/P16
-1162
Ali Sharif Razavian, Hossein Azizpour, Josephine
Sullivan, and Stefan Carlsson. 2014. CNN
features off-the-shelf: An astounding base-
line for recognition. In Proceedings of
the
IEEE/CVF Conference on Computer Vision
and Pattern Recognition (CVPR) Workshops,
512–519. https://doi.org/10
pages
.1109/CVPRW.2014.131
Piyush Sharma, Nan Ding, Sebastian Goodman,
and Radu Soricut. 2018. Conceptual captions:
A cleaned, hypernymed, image alt-text dataset
for automatic image captioning. In Proceedings
of the 56th Annual Meeting of the Associa-
tion for Computational Linguistics (Volume 1:
Long Papers), pages 2556–2565, Melbourne,
Australia. Association
for Computational
Linguistics.
Emma Strubell, Ananya Ganesh, and Andrew
McCallum. 2019. Energy and policy conside-
rations for deep learning in NLP. In Pro-
the 57th Annual Meeting of
ceedings of
the Association for Computational Linguistics,
pages 3645–3650, Florence, Italy. Associa-
tion for Computational Linguistics. https://
doi.org/10.18653/v1/P18-1238
Weijie Su, Xizhou Zhu, Yue Cao, Bin Li, Lewei
Lu, Furu Wei, and Jifeng Dai. 2020. Vl-BERT:
Pre-training of generic visual-linguistic rep-
resentations. In International Conference on
Learning Representations.
Alane Suhr, Stephanie Zhou, Ally Zhang, Iris
Zhang, Huajun Bai, and Yoav Artzi. 2019. A
corpus for reasoning about natural language
grounded in photographs. In Proceedings of
the 57th Annual Meeting of the Association for
Computational Linguistics, pages 6418–6428,
Florence, Italy. Association for Computational
Linguistics.
Hao Tan and Mohit Bansal. 2019. LXMERT:
Learning cross-modality encoder representa-
tions from transformers. In Proceedings of
the 2019 Conference on Empirical Meth-
ods in Natural Language Processing and the
9th International Joint Conference on Natu-
ral Language Processing (EMNLP-IJCNLP),
pages 5100–5111, Hong Kong, China. Associa-
tion for Computational Linguistics. https://
doi.org/10.18653/v1/D19-1514
Ashish Vaswani, Noam Shazeer, Niki Parmar,
Jakob Uszkoreit, Llion Jones, Aidan N. Gomez,
Łukasz Kaiser, and Illia Polosukhin. 2017.
Attention is all you need. I. Guyon, U. V.
Luxburg, S. Bengio, H. Wallach, R. Fergus,
S. Vishwanathan, and R. Garnett, editors,
Advances in Neural Information Processing
Systems, pages 5998–6008. Curran Associates,
Inc.
993
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Yonghui Wu, Mike Schuster, Zhifeng Chen,
Quoc V. Le, Mohammad Norouzi, Wolfgang
Macherey, Maxim Krikun, Yuan Cao, Qin Gao,
Klaus Macherey, Jeff Klingner, Apurva Shah,
Melvin Johnson, Xiaobing Liu, Łukasz Kaiser,
Stephan Gouws, Yoshikiyo Kato, Taku Kudo,
Hideto Kazawa, Keith Stevens, George Kurian,
Nishant Patil, Wei Wang, Cliff Young,
Jason Smith, Jason Riesa, Alex Rudnick, Oriol
Vinyals, Greg Corrado, Macduff Hughes, and
Jeffrey Dean. 2016. Google’s neural machine
translation system: Bridging the gap between
human and machine translation. arXiv preprint
arXiv:1609.08144.
Ning Xie, Farley Lai, Derek Doran, and Asim
Kadav. 2019. Visual entailment: A novel task
for fine-grained image understanding. arXiv
preprint arXiv:1901.06706.
Saining Xie, Ross Girshick, Piotr Doll´ar, Zhuowen
Tu, and Kaiming He. 2017. Aggregated resid-
ual transformations for deep neural networks.
In Proceedings of the IEEE/CVF Conference
on Computer Vision and Pattern Recognition
(CVPR), pages 5987–5995.
Fei Yu, Jiji Tang, Weichong Yin, Yu Sun,
Hao Tian, Hua Wu, and Haifeng Wang.
2021. Ernie-vil: Knowledge enhanced vision-
language representations through scene graph.
Proceedings of
Artificial Intelligence.
the AAAI Conference on
Licheng Yu, Zhe Lin, Xiaohui Shen, Jimei
Yang, Xin Lu, Mohit Bansal, and Tamara L.
Berg. 2018. Mattnet: Modular attention network
for
In
referring expression comprehension.
Proceedings of the IEEE Conference on Com-
puter Vision and Pattern Recognition (CVPR),
pages 1307–1315.
Rowan Zellers, Yonatan Bisk, Ali Farhadi, and
Yejin Choi. 2019. From recognition to cog-
In
nition: Visual commonsense reasoning.
Proceedings of
the IEEE/CVF Conference
on Computer Vision and Pattern Recognition
(CVPR), pages 6713–6724. https://doi
.org/10.1109/CVPR.2019.00688
Luowei Zhou, Hamid Palangi, Lei Zhang,
Houdong Hu, Jason Corso, and Jianfeng Gao.
2020. Unified vision-language pre-training for
image captioning and vqa. Proceedings of
the AAAI Conference on Artificial
Intelli-
gence, 34(07):13041–13049. https://doi
.org/10.1609/aaai.v34i07.7005
Y. Zhu, O. Groth, M. Bernstein, and L. Fei-Fei.
2016. Visual7w: Grounded question answering
in images. In 2016 IEEE Conference on Com-
puter Vision and Pattern Recognition (CVPR),
pages 4995–5004. https://doi.org/10
.1109/CVPR.2016.540
l
D
o
w
n
o
a
d
e
d
f
r
o
m
h
t
t
p
:
/
/
d
i
r
e
c
t
.
m
i
t
.
e
d
u
/
t
a
c
l
/
l
a
r
t
i
c
e
-
p
d
f
/
d
o
i
/
.
1
0
1
1
6
2
/
t
l
a
c
_
a
_
0
0
4
0
8
1
9
6
3
7
3
4
/
/
t
l
a
c
_
a
_
0
0
4
0
8
p
d
.
f
b
y
g
u
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
994