Compositional Generalization in Multilingual
Semantic Parsing over Wikidata
Ruixiang Cui, Rahul Aralikatte, Heather Lent, and Daniel Hershcovich
Department of Computer Science
University of Copenhagen, Denmark
{rc, rahul, hcl, dh}@di.ku.dk
Abstrakt
Semantic parsing (SP) allows humans to lever-
age vast knowledge resources through natural
interaction. Jedoch, parsers are mostly de-
signed for and evaluated on English resources,
such as CFQ (Keysers et al., 2020), the cur-
rent standard benchmark based on English
data generated from grammar rules and ori-
ented towards Freebase, an outdated knowl-
edge base. We propose a method for creating a
multilingual, parallel dataset of question-query
pairs, grounded in Wikidata. We introduce
such a dataset, which we call Multilingual
Compositional Wikidata Questions (MCWQ),
and use it to analyze the compositional gen-
eralization of semantic parsers in Hebrew,
Kannada, Chinese, and English. While within-
language generalization is comparable across
languages, experiments on zero-shot cross-
lingual transfer demonstrate that cross-lingual
compositional generalization fails, even with
state-of-the-art pretrained multilingual encod-
ers. Außerdem, our methodology, dataset,
and results will facilitate future research on
SP in more realistic and diverse settings than
has been possible with existing resources.
1
Einführung
Semantic parsers grounded in knowledge bases
(KBs) enable knowledge base question answer-
ing (KBQA) for complex questions. Many seman-
tic parsers are grounded in KBs such as Freebase
(Bollacker et al., 2008), DBpedia (Lehmann et al.,
2015), and Wikidata (Pellissier Tanon et al.,
2016), and models can learn to answer questions
about unseen entities and properties (Herzig and
Berant, 2017; Cheng and Lapata, 2018, Shen et al.,
2019; Sas et al., 2020). An important desired abil-
ity is compositional generalization—the ability
to generalize to unseen combinations of known
components (Oren et al., 2020; Kim and Linzen,
2020).
937
One of the most widely used datasets for mea-
suring compositional generalization in KBQA is
CFQ (Compositional Freebase Questions; Keyser
et al., 2020), which was generated using grammar
rules, and is based on Freebase, an outdated and
unmaintained English-only KB. While the need to
expand language technology to many languages
is widely acknowledged (Joshi et al., 2020), Die
lack of a benchmark for compositional generaliza-
tion in multilingual semantic parsing (SP) hinders
KBQA in languages other than English. Weiter-
mehr, progress in both SP and KB necessitates
that benchmarks can be reused and adapted for
future methods.
Wikidata is a multilingual KB, with entity and
property labels in a multitude of languages. Es
has grown continuously over the years and is an
important complement to Wikipedia. Much effort
has been made to migrate Freebase data to Wiki-
Daten (Pellissier Tanon et al., 2016; Diefenbach
et al., 2017; Hogan et al., 2021), but only in En-
glish. Investigating compositional generalization
in cross-lingual SP requires a multilingual data-
set, a gap we address in this work.
We leverage Wikidata and CFQ to create
Multilingual Compositional Wikidata Questions
(MCWQ), a new multilingual dataset of compo-
sitional questions grounded in Wikidata (see Fig-
ure 1 for an example). Beyond the original English,
an Indo-European language using the Latin script,
we create parallel datasets of questions in He-
brew, Kannada, and Chinese, which use different
scripts and belong to different language families:
Afroasiatic, Dravidian, and Sino-Tibetan, bzw-
aktiv. Our dataset includes questions in the four
languages and their associated SPARQL queries.
Our contributions are:
• a method to automatically migrate a KBQA
dataset to another KB and extend it to diverse
languages and domains,
Transactions of the Association for Computational Linguistics, Bd. 10, S. 937–955, 2022. https://doi.org/10.1162/tacl a 00499
Action Editor: Emily Pitler. Submission batch: 4/2022; Revision batch: 6/2022; Published 9/2022.
C(cid:2) 2022 Verein für Computerlinguistik. Distributed under a CC-BY 4.0 Lizenz.
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Parsers trained on CFQ transform these ques-
tions into SPARQL queries, which can subse-
quently be executed against Freebase to answer
the original questions (in diesem Fall, ‘‘Yes’’).
the test set
CFQ uses the Distribution-Based Composi-
tionality Assessment (DBCA) method to generate
multiple train-test splits with maximally divergent
examples in terms of compounds, while main-
taining a low divergence in terms of primitive
Elemente (atoms). In these maximum compound
divergence (MCD) splits,
is con-
strained to examples containing novel compounds,
das ist, new ways of composing the atoms seen
during training. For measuring compositional gen-
eralizations, named entities in the questions are
anonymized so that models cannot simply learn the
relationship between entities and properties. CFQ
contains 239,357 English question-answer pairs,
which encompass 49,320 question patterns and
34,921 SPARQL query patterns. Tisch 1 zeigt an
selected fields of an example in CFQ. In their ex-
periments, Keysers et al. (2020) trained seman-
tic parsers using several architectures on various
train-test splits. They demonstrated strong nega-
tive correlation between models’ accuracy (cor-
rectness of the full generated SPARQL query) Und
compound divergence across a variety of system
architectures—all models generalized poorly in
the high-divergence settings, highlighting the need
to improve compositional generalization in SP.
By the time CFQ was released, Freebase
had already been shut down. On that account, Zu
our knowledge,
there is no existing SP data-
set targeting compositional generalization that is
grounded in a currently usable KB, which con-
tains up-to-date information. We therefore mi-
grate the dataset to such a KB, nämlich, Wikidata,
in §3.
Darüber hinaus, only a few studies have evaluated
semantic parsers’ performance in a multilingual
setting, due to the scarcity of multilingual KBQA
datasets (Perevalov et al., 2022B). No compara-
ble benchmark exists for languages other than
English, and it is therefore not clear whether
results are generalizable to other languages. Com-
positional generalization in typologically distant
languages may pose completely different chal-
Längen, as these languages may have different
ways to compose meaning (Evans and Levinson,
2009). We create such a multilingual dataset in
§4, leveraging the multilinguality of Wikidata.
Figur 1: An example from the MCWQ dataset. Der
question in every language corresponds to the same
Wikidata SPARQL query, welche, upon execution,
returns the answer (which is positive in this case).
• a benchmark for measuring compositional
generalization in SP for KBQA over Wiki-
data in four typologically diverse languages,
• monolingual experiments with different SP
architectures in each of the four languages,
demonstrating similar within-language gen-
eralization, Und
• zero-shot cross-lingual experiments using
pretrained multilingual encoders,
show-
ing that compositional generalization from
English to the other languages fails.
Our code for generating the dataset and for
the experiments, as well as the dataset itself and
trained models, are publicly available on https://
github.com/coastalcph/seq2sparql.
2 Limitations of CFQ
CFQ (Keysers et al., 2020) is a dataset for mea-
suring compositional generalization in SP. It tar-
gets the task of parsing questions in English into
SPARQL queries executable on the Freebase KB
(Bollacker et al., 2008). CFQ contains questions
as in Table 1, as well as the following English
question (with entities surrounded by brackets):
‘‘Was [United Artists]
founded by
[Herr. Fix-it]’s star, founded by [D. W.
Griffith], founded by [Mary Pickford],
and founded by [The Star Boarder]'S
star?’’
938
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
CFQ field
questionWithBrackets
questionPatternModEntities
questionWithMids
sparql
sparqlPatternModEntities
Content
Did [‘Murder’ Legendre]’s male actor marry [Lillian Lugosi]
Did M0 ’s male actor marry M2
Did m.0h4y854 ’s male actor marry m.0hpnx3b
SELECT count(*) WHERE { ?x0 ns:film.actor.film/ns:film.performance
.character
ns:M.0h4y854 . ?x0 ns:people.person.gender ns:M.05zppz . ?x0
ns:people.person.spouse s/ns:fictional universe.marriage of
fictional characters.spouses ns:M.0hpnx3b . FILTER ( ?x0 !=
ns:M.0hpnx3b )}
SELECT count(*) WHERE { ?x0 ns:film.actor.film/ns:film.performance
.character
M0 . ?x0 ns:people.person.gender ns:M.05zppz . ?x0
ns:people.person.spouse s
/ns:fictional universe.marriage of fictional characters.spouses M2
. FILTER
( ?x0 != M2 )}
Tisch 1: Selected fields in a CFQ entry. questionWithBrackets is the full English question
with entities surrounded by brackets. questionPatternModEntities is the question with entites
replaced by placeholders. In questionWithMids, the entity codes (Freebase machine IDs; MIDs)
are given instead of their labels. sparql is the fully executable SPARQL query for the question, Und
in sparqlPatternModEntities the entity codes are replaced by placeholders.
3 Migration to Wikidata
Wikidata is widely accepted as the replacement
for Freebase. It is actively maintained and repre-
sents knowledge in a multitude of languages and
domains, and also supports SPARQL. Migrating
Freebase queries to Wikidata, Jedoch, is not triv-
ial, as there is no established full mapping between
the KBs’ properties and entities. An obvious alter-
native to migration would be a replication of the
original CFQ generation process but with Wiki-
data as the KB. Before delving into the details of
the migration process, let us motivate the deci-
sion not to pursue that option: The grammar used
to generate CFQ was not made available to oth-
ers by Keysers et al. (2020) and is prohibitively
too complex to reverse-engineer. Our migration
Verfahren, andererseits, is general and can sim-
ilarly be applied for migrating other datasets from
Freebase to Wikidata. Endlich, many competitive
models with specialized architecture have been
developed for CFQ (Guo et al., 2020; Herzig
et al., 2021; Gai et al., 2021. Our migrated data-
set is formally similar and facilitates their eval-
uation and the development of new methods.
3.1 Property Mapping
Verwendet 51 unique properties in its SPARQL queries,
mostly belonging to the cinematography domain.
These Freebase properties cannot be applied di-
rectly to Wikidata, which uses different property
codes known as P-codes (z.B., P21). We there-
fore need to map the Freebase properties into
Wikidata properties.
As a first step in the migration process, Wir
check which Freebase properties used in CFQ
have corresponding Wikidata properties. Using a
publicly available repository providing a partial
mapping between the KBs,1 we identify that 22
out of the 51 Freebase properties in CFQ can
be directly mapped to Wikidata properties.2 The
andere 29 require further processing:
Fourteen properties are the reverse of other
properties, which do not have Wikidata counter-
Teile. Zum Beispiel, ns:film.director.film
is the reverse of ns:film.film.directed by,
and only the latter has Wikidata mapping, P57.
We resolve the problem by swapping the entities
around the property.
The other 15 properties deal with judging
whether an entity has a certain quality. In CFQ,
?x1 a ns:film.director asks whether ?x1
is a director. Wikidata does not contain such
unary properties. daher, we need to treat these
Wie in der Tabelle zu sehen ist 1, the WHERE clause in a
SPARQL query consists of a list of triples, Wo
the second element in each triple is the property
(z.B., ns:people.person.gender). CFQ
1https://www.wikidata.org/wiki/Wikidata
:WikiProject Freebase/Mapping.
2While some Freebase properties have multiple cor-
responding Wikidata properties, we consider a property
mappable as long as it has at least one mapping.
939
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
CFQ properties as entities in Wikidata. Zum Beispiel-
reichlich, director is wd:Q2526255, so we para-
phrase the query as ?x1 wdt:P106 wd:
Q2526255, asking whether ?x1’s occupation
(P106) is director. Zusätzlich, we substitute the
art director property from CFQ with the com-
poser property because the former has no equiv-
alent in Wikidata. Endlich, we filter out queries
with reverse marks over properties, Zum Beispiel,
?x0 ˆns:people.person.gender M0, fällig
to incompatibility with the question generation
Verfahren (§3.2).
After filtering, we remain with 236,304 entries
with only fully-mappable properties—98.7% of all
entries in CFQ. We additionally make necessary
SPARQL syntax modification for Wikidata.3
3.2 Entity Substitution
A large number of entities in Freebase are absent
in Wikidata. Zum Beispiel, neither of the entities
in Table 1 exist in Wikidata. Außerdem, nicht wie
the case of properties, to our knowledge, Dort
is no comprehensive or even partial mapping of
Freebase entity IDs (d.h., Freebase machine IDs,
MIDs, such as s:M.05zppz) to Wikidata en-
tity IDs (d.h., Q-codes, such as wd:Q6581097).
We replicate the grounding process carried out
by Keysers et al. (2020), substituting entity place-
holders with compatible entities codes by execut-
ing the queries against Wikidata:
1. Replacing entity placeholders with SPARQL
Variablen (z.B., ?v0), we obtain queries that
return sets of compatible candidate entity
assignments instead of simply an answer for
a given assignment of entities.
2. We add constraints for the entities to be
distinct, to avoid nonsensical redundancies
(z.B., due to conjunction of identical clauses).
3. Special entities, representing nationalities
and genders, are regarded as part of the
question patterns in CFQ (and are not re-
placed with placeholders). Before running
the queries, we thus replace all such entities
with corresponding Wikidata Q-codes (In-
stead of variables).
4. We execute the queries against the Wikidata
query service4 to get the satisfying assign-
ments of entity combinations, mit welchem
we replace the placeholders in sparql-
PatternModEntities fields.
5. Endlich, we insert
the Q-codes into the
English questions in the questionWith-
Mids field and the corresponding entity la-
bels into the questionWithBrackets to
obtain the English questions for our dataset.
Along this process, 52.5% of the queries have
at least one satisfying assignment. The resulting
question-query pairs constitute our English data-
set. They maintain the SPARQL patterns in CFQ,
but the queries are all executable on Wikidata.
We obtain 124,187 question-query pairs, von
welche 67,523 are yes/no questions and 56,664
are wh- Fragen. The expected responses of
yes/no questions in this set are all ‘‘yes’’ due to
our entity assignment process. To make MCWQ
comparable to CFQ, which has both positive and
negative answers, we sample alternative queries
by replacing entities with ones from other queries
whose preceding predicates are the same. Unser
negtive sampling results in 30,418 questions with
‘‘no’’ answers.
3.3 Migration Example
Consider the SPARQL pattern from Table 1:
SELECT count(*) WHERE { ?x0 ns:film.actor.
film/ns:film.performance.character M0 .
?x0 ns:people.person.gender ns:M.05zppz .
?x0 ns:people.person.spouse s/ns:
fictional universe.
marriage of fictional characters.spouses
M2 . FILTER ( ?x0 != M2 )}
We replace the properties and special enti-
Krawatten (here the gender male: ns:M.05zppz →
wd:Q6581097):
SELECT count(*) WHERE {?x0 wdt:P453 M0 .?x0
wdt:P21 wd:Q6581097 .
FILTER ( ?x0 != M2 )}
?x0 wdt:P26 M2 .
Then we replace placeholders (z.B., M0) with vari-
ables and add constraints for getting only one
Abtretung (which is enough for our purposes)
with distinct entities. The resulting query is:
3CFQ uses SELECT count(*) WHERE to query
yes/no questions, but this syntax is not supported by Wiki-
Daten. We replace it with ASK WHERE, intended for Boolean
Abfragen.
SELECT ?v0 ?v1 WHERE ?x0 wdt:P453 ?v0. ?x0
wdt:P21 wd:Q6581097.
?x0 wdt:P26
?v1.
4https://query.wikidata.org/.
940
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
FILTER ( ?x0 != ?v1 ). FILTER ( ?v0 !=
?v1 ) LIMIT 1
We execute the query and get wd:Q50807639
(Lohengrin) and wd:Q1560129 (Margarete
Joswig) as satisfying answers for v0 and v1,
jeweils. Note that these are different from the
entities in the original question (‘Murder’ Legendre
and Lillian Lugosi)—in general, there is no guar-
antee that the same entities from CFQ will be
preserved in our dataset. Then we put back
these answers into the query, and make neces-
sary SPARQL syntax modification for Wikidata.
The final query for this entry is:
ASK WHERE {?x0 wdt:P453 wd:Q50807639. ?x0 wdt:
P21 wd:Q6581097
Q1560129 . FILTER ( ?x0 != wd:Q1560129 )}
?x0 wdt:P26 wd:
.
As for the English question, we map the Free-
base entities in the questionWithMids field
with the labels of the obtained Wikidata entities.
daher, the English question resulting from this
process is:
Did [Lohengrin] ’s male actor marry
[Margarete Joswig]?
Figur 2: Complexity distribution of the MCD1 split
of CFQ (über) and MCWQ (below).
3.4 Dataset Statistics
We compare the statistics of MCWQ with CFQ
in Table 3. MCWQ has 29,312 unique question
patterns (mod entities, verbs, etcs), das ist, 23.6%
of questions cover all question patterns, com-
pared to 20.6% in CFQ. Außerdem, MCWQ
hat 86,353 unique query patterns (mod enti-
Krawatten), ergebend 69.5% of instances covering
all SPARQL patterns, 18% higher than CFQ. Unser
dataset thus poses a greater challenge for com-
positional SP, and exhibits less redundancy in
terms of duplicate query patterns. It is worth not-
ing that less unique query percentage in MCWQ
than CFQ results from the loss during swapping
the entities in §3.1.
To be compositionally challenging, Keyser
et al. (2020) generated the MCD splits to have
high compound divergence while maintaining
low atom divergence. As atoms in MCWQ are
mapped from CFQ while leaving the composi-
tional structure intact, we derive train-test splits
of our dataset by inducing the train-test splits from
CFQ on the corresponding subset of instances in
our dataset.
The complexity of questions in CFQ is mea-
sured by recursion depth and reflects the number
of rule applications used to generate a ques-
tion, which encompasses grammar, Wissen,
inference, and resolution rules. While each ques-
tion’s complexity in MCWQ is the same as the
corresponding CFQ question’s, some cannot be
migrated (see §3.1 and §3.2). To verify the com-
pound divergence is not affected, we compare
the question complexity distribution of the two
datasets in one of the three compositional splits
(MCD1) in Abbildung 2. The training, Entwicklung,
and test sets of the split in CFQ and MCWQ
follow a similar trend in general. The fluctuation
in the complexity of questions in the MCWQ
splits reflects the dataset’s full distribution—see
Figur 3.
Stemming from its entities and properties, CFQ
questions are limited to the domain of movies. Der
entities in MCWQ, Jedoch, can in principle come
from any domain, owing to our flexible entity
replacing method. Though MCWQ’s properties
are still a subset of those used in CFQ, they are
primarily in the movies domain. We also observe a
few questions from literature, Politik, and history
in MCWQ.
941
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Tisch 2: The MCWQ example from Figure 1. The English question is generated from the CFQ
entry in Table 1 by the migration process described in §3.3, and the questions in the other languages
are automatically translated (§4.1). The questionWithBrackets, questionPatternMod-
Entities, sparql, and sparqlPatternModEntities fields are analogous to the CFQ ones.
recursionDepth (which quantifies the question complexity) and expectedResponse (welche
is the answer returned upon execution of the query) are copied from the CFQ entry.
CFQ
MCWQ
Unique questions
Questions patterns
Unique queries
Query patterns
239,357
124,187
49,320 (20.6%)
228,149 (95.3%)
123,262 (51.5%)
29,312 (23.6%)
101,856 (82%)
86,353 (69.5%)
Yes/no questions
Wh- Fragen
130,571 (54.6%)
108,786 (45.5%)
67,523 (54.4%)
56,664 (45.6%)
Tisch 3: Dataset statistics comparison for MCWQ
and CFQ. Percentages are relative to all unique
Fragen. Questions patterns refer to mod enti-
Krawatten, verbs, usw. while query patterns refer to mod
entities only.
4.1 Generating Translations
Both question patterns and bracketed questions
are translated separately with Google Cloud Trans-
lation5 from English.6 SPARQL queries remain
unchanged, as both property and entity IDs are
language-independent in Wikidata, which con-
tains labels in different languages for each. Tisch 2
shows an example for a question in our dataset
(which is generated from the same question as
5https://cloud.google.com/translate.
6We attempted to translate bracketed questions and sub-
sequently replace the bracketed entities with placeholders as
question patterns. In preliminary experiments, we found that
separate translation of question patterns is of higher trans-
lation quality. daher, we choose to translate question
patterns and bracketed questions individually.
Figur 3: Complexity distribution of MCWQ, mea-
sured by recursion depth, compared to CFQ.
4 Generating Multilingual Questions
To create a typologically diverse dataset, start-
ing from our English dataset (an Indo-European
language using the Latin script), we use machine
translation to three other languages from differ-
ent families (Afroasiatic, Dravidian, and Sino-
Tibetan), which use different scripts: Hebrew,
Kannada, and Chinese (§4.1). For a comparison
to machine translation and a more realistic eval-
uation with regard to compositional SP, we man-
ually translate a subset of the test sets of the
three MCD splits (§4.2) and evaluate the machine
translation quality (§4.3).
942
the CFQ instance from Table 1), as well as the
resulting translations.
As an additional technical necessity, we add
a question mark to the end of each question be-
fore translation (as the original dataset does not
include question marks) and remove trailing ques-
tion marks from the translated question before
including it in our dataset. We find this step to be
essential for translation quality.
4.2 Gold Test Set
CFQ and other datasets for evaluating compo-
sitional generalization (Lake and Baroni, 2018;
Kim and Linzen, 2020) are generated from gram-
mars. Jedoch, It has not been investigated how
well models trained on them generalize to hu-
man questions. As a step towards that goal,
we evaluate whether models trained with auto-
matically generated and translated questions can
generalize to high-quality human-translated ques-
tionen. For that purpose, we obtain the intersection
of the test sets of the MCD splits (1,860 In-
versucht), and sample two translated questions with
yes/no questions and two with wh- questions for
each complexity level (if available). This sample,
termed test-intersection-MT, hat 155 entries in
total. The authors (one native speaker for each
Sprache) manually translate the English ques-
tions into Hebrew, Kannada, and Chinese. Wir
term the resulting dataset test-intersection-gold.
4.3 Translation Quality
test-intersection-MT against
We compute the BLEU (Papineni et al., 2002)
test-
scores of
intersection-gold using SacreBLEU (Post, 2018),
ergebend 87.4, 76.6, Und 82.8 for Hebrew,
Kannada, and Chinese, jeweils. This indicates
high quality of the machine translation outputs.
Zusätzlich, one author for each language
manually assesses translation quality for one sam-
pled question from each complexity level from
the full dataset (40 in Summe). We rate the trans-
lations on a scale of 1–5 for fluency and for
meaning preservation, mit 1 being poor, Und 5 Sei-
ing optimal. Despite occasional translation issues,
mostly attributed to lexical choice or morphologi-
cal agreement, we confirm that the translations are
of high quality. Across languages, über 80% von
examples score 3 or higher in fluency and meaning
preservation. The average meaning preservation
scores for Hebrew, Kannada, and Chinese are 4.4,
3.9, Und 4.0, jeweils. For fluency, they are
3.6, 3.9, Und 4.4, jeweils.
As a control, one of the authors (a native
English speaker) evaluated English fluency for
the same sample of 40 Fragen. Nur 62% von
patterns were rated 3 or above. While all English
questions are grammatical, many suffer from poor
fluency, tracing back to their automatic genera-
tion using rules. Some translations are rated higher
in terms of fluency, mainly due to annotator le-
niency (focusing on disfluencies that might result
from translation) and paraphrasing of unnatural
constructions by the MT system (especially for
lower complexities).
5 Experimente
While specialized architectures have achieved
state-of-the-art results on CFQ (Guo et al., 2020,
2021; Gai et al., 2021), these approaches are
English- or Freebase-specific. We therefore exper-
iment with sequence-to-sequence (seq2seq) mod-
els, among which T5 (Raffel et al., 2020) has been
shown to perform best on CFQ (Herzig et al.,
2021). We evaluate these models for each lan-
guage separately (§5.1), and subsequently evaluate
their cross-lingual compositional generalization
(§5.2).
5.1 Monolingual Experiments
We evaluate six models’ monolingual parsing
performance on the three MCD splits and a random
split of MCWQ. As done by Keysers et al. (2020),
entities are masked during training, except those
that are part of the question patterns (genders and
nationalities).
We experiment with two seq2seq architectures
on MCWQ for each language, with the same hy-
perparameters tuned by Keysers et al. (2020) An
the CFQ random split: LSTM (Hochreiter and
Schmidhuber, 1997) with attention mechanism
(Bahdanau et al., 2015) and Evolved Trans-
ehemalig (So et al., 2019), both implemented using
Tensor2Tensor (Vaswani et al., 2018). Separate
models are trained and evaluated per language,
with randomly initialized (not pretrained) In-
coders. We train a model for each of the three
MCD splits plus a random split for each language.
We also experiment with pretrained language
Modelle (PLMs), to assess whether multilingual
PLMs, mBERT (Devlin et al., 2019) and mT5
943
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
MCD1
MCD2
MCD3
MCDmean
Random
Exact Match (%)
En
Er
Kn
Zh
En
Er
Kn
Zh
En
Er
Kn
Zh
En
Er
Kn
Zh
En
Er
Kn
Zh
LSTM+Attention
E. Transformer
mBERT
T5-base+RIR
mT5-small+RIR
mT5-base+RIR
38.2
53.3
49.5
57.4
77.6
55.5
29.3
35
38.7
–
57.8
59.5
27.1
30.7
34.4
–
55
49.1
26.1
31
35.6
–
52.8
30.2
6.3
16.5
13.4
14.6
13
27.7
5.6
8.7
11.4
–
12.6
16.6
9.9
11.9
12.3
–
8.2
16.6
7.5
10.2
15.1
–
21.1
23
13.6
18.2
17
12.3
24.3
18.2
11.5
13
18
–
17.5
23.4
15.7
18.1
18.1
–
31.4
30.5
15.1
15.5
19.4
–
34.9
35.6
19.4
29.3
26.6
28.1
38.3
33.8
15.5
18.9
22.7
–
29.3
33.2
17.6
20.2
21.6
–
31.5
32.1
16.2
18.9
23.4
–
36.3
29.6
96.6
99
98.7
98.5
98.6
99.1
80.8
90.4
91
–
90
90.6
88.7
93.7
95.1
–
93.8
94.2
86.8
92.2
93.3
–
91.8
92.2
Tisch 4: Monolingual evaluation: Exact match accuracies on MCWQ. MCDmean is the mean accu-
racy of all three MCD splits. Random represents a random split of MCWQ. This is an upper bound
on the performance shown only for comparison. As SPARQL BLEU scores are highly correlated with
accuracies in this experiment, we only show the latter here.
(Xue et al., 2020), are as effective for monolin-
gual compositional generalization as an English-
only PLM using the Transformers library (Wolf
et al., 2020).
For mBERT, we fine-tune a multi cased
L-12 H-768 A-12 encoder and a randomly
initialized decoder of the same architecture. Wir
train for 100 epochs with patience of 25, batch
size of 128, and learning rate of 5 × 10−5 with a
linear decay.
For T5, we fine-tune T5-base on MCWQ
English, and mT5-small and mT5-base on
each language separately. We use the default hy-
perparameter settings except trying two learning
Tarife, 5e−4 and 3e−5 (see results below). SPARQL
queries are pre-processed using reversible inter-
mediate representations (RIR), previously shown
(Herzig et al., 2021) to facilitate compositional
generalization for T5. We fine-tune all models for
50K steps.
We use six Titan RTX GPUs for training, mit
batch size of 36 for T5-base, 24 for mT5-
small, Und 12 for mT5-base. We use two
random seeds for T5-base. It takes 384 hours
to finish a round of mT5-small experiments,
120 hours for T5-base, Und 592 hours for
mT5-base.
In addition to exact-match accuracy, we report
the BLEU scores of the predictions computed
with SacreBLEU, as a large portion of the gen-
erated queries is partially (but not fully) correct.
Results The results are shown in Table 4. Während
models generalize almost perfectly in the random
split for all four languages, the MCD splits are
much harder, with the highest mean accuracies
von 38.3%, 33.2%, 32.1%, Und 36.3% for English,
Hebrew, Kannada, and Chinese, jeweils. Für
comparison, on CFQ, T5-base+RIR has an ac-
curacy of 60.8% on MCDmean (Herzig et al.,
2021). One reason for this decrease in performance
is the smaller training data: The MCWQ dataset
hat 52.5% the size of CFQ. Außerdem, MCWQ
has less redundancy than CFQ in terms of dupli-
cate questions and SPARQL patterns, Rendern
models’ potential strategy of simply memorizing
patterns less effective.
Contrary to expectation, mT5-base does not
outperform mT5-small. During training, Wir
found mT5-base reached minimum loss early
(after 1k steps). By changing the learning rate from
the default 3e−5 to 5e−4, we seem to have over-
come the local minimum. Training mT5-small
with learning rate 5e−4 also renders better per-
Form. Außerdem, the batch size we use for
mT5-base may not be optimal, but we could
not experiment with larger batch sizes due to
resource limitations.
Comparing the performance across languages,
mT5-base performs best on Hebrew and Kan-
nada on average, while mT5-small has the best
performance on English and Chinese. Due to re-
source limitations, we were not able to look deeper
into the effect of hyperparameters or evaluate
larger models. Jedoch, our experiments show
that while multilingual compositional generaliza-
tion is challenging for seq2seq semantic parsers,
within-language generalization is comparable be-
tween languages. dennoch, English is always
the easiest (at least marginally). A potential cause
is that most semantic query languages were ini-
tially designed to represent and retrieve data stored
in English databases, and thus have a bias towards
English. Folglich, SPARQL syntax is closer
to English than Hebrew, Kannada, and Chinese.
While translation errors might have an effect as
well, we have seen in §4.3 that translation quality
is high.
To investigate further, we plot the complexity
distribution of true predictions (exactly matching
944
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
MCDmean
Random
SPARQL BLEU
En
Er
Kn
Zh
En
Er
Kn
Zh
mT5-small+RIR
mT5-base+RIR
87.5
86.4
53.8
46.4
53.2
46
59
52.7
99.9
99.9
60.4
63.2
59.9
63.5
63.8
70.6
Exact Match (%)
mT5-small+RIR
mT5-base+RIR
38.3
33.8
0.2
0.4
0.3
0.7
0.2
1.5
98.6
99.1
0.5
1.1
0.4
0.9
1.1
7.2
Tisch 5: Mean BLEU scores and exact match ac-
curacies on the three MCD splits and on a random
split in zero-shot cross-lingual transfer experi-
ments on MCWQ. The gray text represents the
models’ monolingual performance on English,
given for reference (the exact match accuracies
are copied from Table 4). The black text indicates
the zero-shot cross-lingual transfer performances
on Hebrew, Kannada, and Chinese of a model
trained on English. While the scores for individ-
ual MCD splits are omitted for brevity, in all three
MCD splits, the accuracies are below 1% (except
on MCD2 Chinese, Sein 4%).
et al., 2020; Sherborne and Lapata, 2022). Sei-
cause translating datasets and training KBQA
systems is expensive, it is beneficial to leverage
multilingual PLMs, fine-tuned on English data,
for generating SPARQL queries over Wikidata
given natural language questions in different lan-
guages. While compositional generalization is
difficult even in a monolingual setting, it is inter-
esting to investigate whether multilingual PLMs
can transfer in cross-lingual SP over Wikidata.
Simple seq2seq T5/mT5 models perform reason-
ably well (> 30% accuracy) on monolingual SP
on some splits (see §5.1). We investigate whether
the learned multilingual representations of such
models enable compositional generalization even
without target language training. We use mT5-
small+RIR and mT5-base+RIR, the best two
models trained and evaluated on English from
previous experiments,
to predict on the other
languages.
Results The results are shown in Table 5. Beide
BLEU and exact match accuracy of the predicted
SPARQL queries drop drastically when the model
is evaluated on Hebrew, Kannada, and Chinese.
mT5-small+RIR achieves 38.3% accuracy on
MCDmean English, but less than 0.3% in zero-shot
parsing on three non-English languages.
Even putting aside compositionality evaluation,
as seen in the random split, the exact match ac-
curacy in the zero-shot cross-lingual setting is
Figur 4: Two mT5 models’ number of correct pre-
the three MCD splits in
dictions summing over
monolingual experiments, plotted by complexity level.
Each line represents a language. While mT5-small
generalizes better overall, mT5-base is better in
lower complexities (which require less compositional
generalization).
the gold SPARQL) per language by the two best
systems in Figure 4. We witness a near-linear
performance decay from complexity level 19. Wir
find that mT5-base is better than mT5-small
on lower complexity despite the latter’s supe-
rior overall performance. Interessant, übersetzt
questions seem to make the parsers generalize
better at higher complexity, as shown in the
figure. For mT5-small, the three non-English
models successfully parse more questions within
the complexity range 46–50 than English, für
mT5-base 44–50. As is discussed in §4.3,
machine-translated questions tend to have higher
fluency than English questions; we conjecture
that such a smoothing method helps the parser
to understand and learn from higher complexity
Fragen.
5.2 Zero-shot Cross-lingual Parsing
Zero-shot cross-lingual SP has witnessed new
advances with the development of PLMs (Shao
945
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
still low. The relatively high BLEU scores can
be attributed to the small overall vocabulary
used in SPARQL queries. Interessant, while
mT5-base+RIR on MCDmean English does not
outperform mT5-small+RIR,
it yields better
performance in the zero-shot setting. For He-
brew, Kannada, and Chinese, the accuracies are
0.2%, 0.4%, Und 1.3% higher, jeweils. Für
mT5-base, Chinese is slightly easier than Kan-
nada and Hebrew to parse in the zero-shot setting,
outperforming 1.1% Und 0.8%.
To conclude, zero-shot cross-lingual transfer
from English to Hebrew, Kannada, and Chinese
fails to generate valid queries in MCWQ. A po-
tential cause for such unsuccessful transfer is that
all four languages in MCWQ belong to different
language families and have low linguistic sim-
ilarities. It remains to be investigated whether
such cross-lingual transfer will be more effective
on related languages, such as from English to
Deutsch (Lin et al., 2019).
6 Analyse
6.1 Evaluation with Gold Translation
Most existing compositional generalization data-
sets focus on SP (Lake and Baroni, 2018; Kim
and Linzen, 2020; Keysers et al., 2020). Diese
datasets are composed either with artificial lan-
guage or in English using grammar rules. Mit
test-intersection-gold proposed in §4.2, wir drin-
vestigate whether models can generalize from
a synthetic automatically translated dataset to a
manually translated dataset.
We use the monolingual models trained on
three MCD splits to parse test-intersection-gold.
In Table 6, we present the mean BLEU scores and
exact match accuracy of the predicted SPARQL
Abfragen. There is no substantial difference be-
tween the performances on the two intersection
sets, except for Kannada, which has a 4% accu-
racy drop on average. These results testify that
MCWQ has sufficiently high translation qual-
ity and that models trained with such synthetic
data can be used to generalize to high-quality
manually-translated questions.
6.2 Categorizing Errors
In an empirical analysis, we categorize typical
prediction errors on test-intersection-gold and
test-intersection-MT into six types: missing prop-
erty, extra property, wrong property (bei dem die
SPARQL BLEU
En
Er
Kn
Zh
En
Er
Kn
Zh
test-intersection-MT
test-intersection-gold
mT5-small+RIR
mT5-base+RIR
86.1
85.5
82.5
83.7
78.9
81.8
85.1
83.2
Exact Match (%)
mT5-small+RIR
mT5-base+RIR
45.6
40.4
35.7
41.9
32.7
40.2
38.5
38.7
–
–
–
–
81.8
83.8
77.7
80.9
86
83.8
35.9
41.1
28.2
34
39.8
38.9
Tisch 6: Mean BLEU scores and accuracies of
monolingual models (§5.1) on test-intersection-
MT and test-intersection-gold. The numbers are
averaged over the accuracies of the predictions
from the monolingual models trained on three
MCD splits. Gesamt, there is no substantial dif-
ference between the performances on the two
intersection sets, demonstrating the reliability of
evaluating on machine translated data in this case.
Figur 5: Number of errors per category in differ-
ent SPARQL predictions on test-intersection-MT and
test-intersection-gold, averaged across monolingual
mT5-small+RIR models trained on the three MCD
splits. The total number of items in each test set is 155.
two property sets have the same numbers of prop-
erties, but the elements do not match), missing
entity, extra entity and wrong entity (wieder, same
number of entities but different entity sets). Wir
plot the mean number of errors per category, als
well as the number of predictions with multiple
Fehler, in Abbildung 5 for monolingual mT5-small
Modelle. Gesamt, model predictions tend to have
more missing properties and entities than extra
ones. Different languages, Jedoch, vary in er-
ror types. Zum Beispiel, on Hebrew, models make
more missing property/entity errors than other
languages; but on Kannada they make more ex-
tra property/entity errors than the others. About
70 out of the 155 examples contain multiple
errors for all languages, with Kannada having
slightly more.
946
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Figur 7: Example of an error reflecting incorrect
predicate-argument structure. wdt:P57 is director
and wdt:P58 is screenwriter. Incorrect triples are
shown in red and missed triples in blue.
placeholders. In the example shown in Figure 7,
we see that the model generates M1 wdt:P57
M2 instead of M0 wdt:P57 M2, which in-
dicates incorrect predicate-argument structure
Deutung.
7 Related Work
Compositional Generalization Compositional
generalization has witnessed great developments
in recent years. SCAN (Lake and Baroni, 2018),
a synthetic dataset consisting of natural language
and command pairs, is an early dataset designed
to systematically evaluate neural networks’ gen-
eralization ability. CFQ and COGS are two more
realistic benchmarks following SCAN. Es gibt
various approaches developed to enhance com-
positional generalization, Zum Beispiel, by using
hierarchical poset decoding (Guo et al., 2020),
combining relevant queries (Das et al., 2021)
using span representation (Herzig and Berant,
2021), and graph encoding (Gai et al., 2021).
In addition to pure language, the evaluation of
compositional generalization has been expanded
to image captioning and situated language under-
Stehen (Nikolaus et al., 2019; Ruis et al., 2020).
Multilingual and cross-lingual compositional gen-
eralization is an important and challenging field
to which our paper aims to bring researchers’
attention.
reading
to machine
Knowledge Base Question Answering Com-
paring
comprehension
(Rajpurkar et al., 2016; Joshi et al., 2017; Shao
et al., 2018; Dua et al., 2019; d’Hoffschmidt
et al., 2020), KBQA is less diverse in terms of
datasets. Datasets such as WebQuestions (Berant
et al., 2013), SimpleQuestions (Bordes et al., 2015),
ComplexWebQuestions (Talmor and Berant, 2018),
FreebaseQA (Jiang et al., 2019), GrailQA (Gu
et al., 2021), CFQ and *CFQ (Tsarkov et al., 2021)
were proposed on Freebase, a now-discontinued
KB. SimpleQuestions2Wikidata (Diefenbach et al.,
Figur 6: Number of errors per category in differ-
ent zero-shot cross-lingual SPARQL predictions on
test-intersection-MT, averaged across mT5-small+
RIR models trained on the three MCD splits in En-
glish. Zusätzlich, mean error counts on the English
set are given for comparison. The total number of
items in each test set is 155.
Comparing errors on test-intersection-gold and
test-intersection-MT, we find missing properties
are more common in gold for all languages. Für
Hebrew and Kannada, extra properties and enti-
ties are also more common in gold. Jedoch, für
Chinese, these and missing entities are less com-
mon in gold compared to MT.
In Abbildung 6 we plot the error statistics for zero-
transfer using mT5-small
shot cross-lingual
Modelle. We can see that there are drastically more
error occurrences. For both missing and extra
property/entity,
the numbers are about double
those from monolingual experiments. Die Nummer
of wrong property/entity errors remain similar,
due to the difficulty of even predicting a set of
the correct size in this setting. For all three target
languages, nearly all predictions contain multi-
ple errors. The statistics indicate the variety and
pervasiveness of errors.
6.3 Other Observations
We also find that, comparatively, parsers perform
well on short questions on all four languages.
This is expected as the compositionality of these
questions is inherently low. On languages other
than English, the models perform well when the
translations are faithful. On occasions when they
are less faithful or fluent but still generate cor-
rect queries, we hypothesize that translation acts
as data regularizers, especially at higher com-
plexities, as demonstrated in Figure 4.
Among wrong entity errors, the most common
cause across languages is the shuffling of entity
947
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
2017) and ComplexSequentialQuestions (Saha
et al., 2018) are based on Wikidata, Aber, like most
Andere, they are monolingual English datasets.
Related to our work is RuBQ (Korablinov and
Braslavski, 2020; Rybin et al., 2021), an English-
Russian dataset for KBQA over Wikidata. Während
the dataset is bilingual, it uses crowdsourced ques-
tions and is not designed for compositionality
Analyse. Kürzlich, Thorne et al. (2021) proposed
WIKINLDB, a Wikidata-based English KBQA
dataset, focusing on scalability rather than compo-
sitionality. Other related datasets include QALM
(Kaffee et al., 2019), a dataset for multilingual
question answering over a set of different popular
knowledge graphs, intended to help determine the
multilinguality of those knowledge graphs. Simi-
larly, QALD-9 (Ngomo, 2018) and QALD-9-plus
(Perevalov et al., 2022A) support the develop-
ment of multilingual question answering systems,
tied to DBpedia and Wikidata, jeweils. Der
goal of both datasets is to expand QA systems to
more languages rather than improving composi-
tionality. KQA Pro (Cao et al., 2022), a concurrent
work to us, is an English KBQA dataset over
Wikidata with a focus on compositional reasoning.
Wikidata has been leveraged across many NLP
tasks such as coreference resolution (Aralikatte
et al., 2019), frame-semantic parsing (Sas et al.,
2020), entity linking (Kannan Ravi et al., 2021),
and named entity recognition (Nie et al., 2021).
As for KBQA, the full potential of Wikidata is
yet to be explored.
Multilingual
and Cross-lingual Modeling
Benchmarks such as XGLUE (Liang et al., 2020)
and XTREME (Hu et al., 2020) focus on multi-
lingual classification and generation tasks. Cross-
lingual learning has been studied across multiple
fields, such as sentiment analysis (Abdalla and
Hirst, 2017), document classification (Dong and
de Melo, 2019), POS tagging (Kim et al., 2017),
and syntactic parsing (Rasooli and Collins, 2017).
In den vergangenen Jahren, multilingual PLMs have been a
primary tool for extending NLP applications to
low-resource languages, as these models amelio-
rate the need to train individual models for each
Sprache, for which less data may be available.
Several studies have attempted to explore the
limitations of such models in terms of practical
usability for low-resource languages (Wu and
Dredze, 2020), and also the underlying elements
that make cross-lingual transfer learning viable
(Dufter and Sch¨utze, 2020). Beyond these PLMs,
other works focus on improving cross-lingual
learning by making particular changes to the
encoder-decoder architecture, such as adding
adapters to attune to specific information (Artetxe
et al., 2020B; Pfeiffer et al., 2020).
For cross-lingual SP, Sherborne and Lapata
(2022) explored zero-shot SP by aligning latent
Darstellungen. Zero-shot cross-lingual SP has
also been studied in dialogue modeling (Nicosia
et al., 2021). Yang et al. (2021) present augmenta-
tion methods for Discourse Representation Theory
(Liu et al., 2021B). Oepen et al. (2020) explore
cross-framework and cross-lingual SP for mean-
ing representations. To the best of our knowledge,
our work is the first on studying cross-lingual
transfer learning in KBQA.
8 Limitations
MCWQ is based on CFQ, a rule-base generated
dataset, and hence it has the inherited unnatural-
ness in question-query pairs of high complexity.
Zweitens, we use machine translation to make
MCWQ multilingual. Although this is the domi-
nant approach for generating multilingual datasets
(Ruder et al., 2021) and we have provided ev-
idences that MCWQ has reasonable translation
accuracy and fluency with human evaluation and
comparative experiments in §4.3 and §5.1, ma-
chine translation would nevertheless create sub-
standard translation artifacts (Artetxe et al.,
2020A). One alternative is to write rules for tem-
plate translation. The amount of work can possibly
be reduced by refering to a recent work (Goodwin
et al., 2021) in which English rules are provided
for syntactic dependency parsing on CFQ’s ques-
tion fields.
Außerdem, the assumption that an English KB
is a ‘‘canonical’’ conceptualization is unjustified,
as speakers of other languages may know and care
about other entities and relationships (Liu et al.,
2021A; Hershcovich et al., 2022A). daher,
future work must create multilingual SP datasets
by sourcing questions from native speakers rather
than translating them.
9 Abschluss
The field of KBQA has been saturated with
work on English, due to both the inherent chal-
lenges of translating datasets and the reliance on
English-only DBs. In this work, we presented a
948
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
MCWQ mT5-base+RIR
Verweise
Information
Einheit
1. Model publicly available?
2. Time to train final model
3. Time for all experiments
4. Energy consumption
5. Location for computations
6. Energy mix at location
7. CO2eq for final model
8. CO2eq for all experiments
Ja
592 hours
1315 hours
2209.2 kWh
Denmark
191 gCO2eq/ kWh
189.96 kg
421.96 kg
Tisch 7: Climate performance model card for
mT5-base+RIR fine-tuned on all splits and
languages.
method for migrating the existing CFQ dataset
to Wikidata and created a challenging multilin-
gual dataset, MCWQ,
targeting compositional
generalization in multilingual and cross-lingual
SP. In our experiments, we observe that pre-
trained multilingual language models struggle to
transfer and generalize compositionally across
languages. Our dataset will facilitate building
robust multilingual semantic parsers by serving
as a benchmark for evaluation of cross-lingual
compositional generalization.
10 Environmental Impact
Following the climate-aware practice proposed
by Hershcovich et al. (2022B), we present a cli-
mate performance model card in Table 7. ‘‘Time
to train final model’’ is the sum over splits and
languages for mT5-base+RIR, while ‘‘Time for
all experiments’’ also includes the experiments
with the English-only T5-base+RIR across all
splits. Although the work does not have direct pos-
itive environmental impact, better understanding
of compositional generalization, resulting from
unsere Arbeit, will facilitate more efficient modeling
and therefore reduce emissions in the long term.
Danksagungen
The authors thank Anders Søgaard and Miryam
de Lhoneux for their comments and suggestions,
as well as the TACL editors and several rounds of
reviewers for their constructive evaluation. Das
project has received funding from the European
Union’s Horizon 2020 research and innovation
programme under the Marie Skłodowska-Curie
grant agreement No. 801199 (Heather Lent).
Mohamed Abdalla and Graeme Hirst. 2017.
Cross-lingual sentiment analysis without (good)
Übersetzung. In Proceedings of the Eighth In-
ternational Joint Conference on Natural Lan-
guage Processing (Volumen 1: Long Papers),
pages 506–515, Taipeh, Taiwan. Asian Federa-
tion of Natural Language Processing.
Rahul Aralikatte, Heather Lent, Ana Valeria
Gonzalez, Daniel Herschcovich, Chen Qiu,
Anders Sandholm, Michael Ringaard, Und
Anders Søgaard. 2019. Rewarding coreference
resolvers for being consistent with world knowl-
edge. In Proceedings of the 2019 Conference
on Empirical Methods in Natural Language
Processing and the 9th International Joint
Conference on Natural Language Processing
(EMNLP-IJCNLP), pages 1229–1235, HongKong,
China. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/D19-1118
Mikel Artetxe, Gorka Labaka,
and Eneko
Agirre. 2020A. Translation artifacts in cross-
transfer learning. In Proceedings of
lingual
Die 2020 Conference on Empirical Methods
in Natural Language Processing (EMNLP),
pages 7674–7684, Online. Association for Com-
putational Linguistics. https://doi.org/10
.18653/v1/2020.emnlp-main.618
Mikel Artetxe, Sebastian Ruder, and Dani
Yogatama. 2020B. On the cross-lingual trans-
ferability of monolingual representations. In
Proceedings of the 58th Annual Meeting of
the Association for Computational Linguistics,
pages 4623–4637, Online. Association for Com-
putational Linguistics. https://doi.org
/10.18653/v1/2020.acl-main.421
Dzmitry Bahdanau, Kyung Hyun Cho, and Yoshua
Bengio. 2015. Neural machine translation by
jointly learning to align and translate. In 3rd
International Conference on Learning Repre-
Sendungen, ICLR 2015.
Jonathan Berant, Andrew Chou, Roy Frostig, Und
Percy Liang. 2013. Semantic parsing on Free-
base from question-answer pairs. In Proceed-
ings of
Die 2013 Conference on Empirical
Methods in Natural Language Processing,
pages 1533–1544, Seattle, Washington, USA.
Verein für Computerlinguistik.
949
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Kurt Bollacker, Colin Evans, Praveen Paritosh,
Tim Sturge, and Jamie Taylor. 2008. Freebase:
A collaboratively created graph database for
structuring human knowledge. In Proceedings
of the 2008 ACM SIGMOD International Con-
ference on Management of Data, SIGMOD ’08,
pages 1247–1250, New York, New York, USA. Asso-
ciation for Computing Machinery. https://
doi.org/10.1145/1376616.1376746
Antoine Bordes, Nicolas Usunier, Sumit Chopra,
and Jason Weston. 2015. Large-scale simple
question answering with memory networks.
arXiv preprint arXiv:1506.02075.
Shulin Cao, Jiaxin Shi, Liangming Pan, Lunyiu
Nie, Yutong Xiang, Lei Hou, Juanzi Li, Bin He,
and Hanwang Zhang. 2022. KQA pro: A dataset
with explicit compositional programs for com-
plex question answering over knowledge base.
In Proceedings of the 60th Annual Meeting of
the Association for Computational Linguistics
(Volumen 1: Long Papers), pages 6101–6119,
Dublin, Ireland. Association for Computational
Linguistik. https://doi.org/10.18653
/v1/2022.acl-long.422
Jianpeng Cheng and Mirella Lapata. 2018.
Weakly-supervised neural semantic parsing
with a generative ranker. In Proceedings of
the 22nd Conference on Computational Natural
Language Learning, pages 356–367, Brussels,
Belgien. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/K18-1035
Rajarshi Das, Manzil Zaheer, Dung Thai, Ameya
Godbole, Ethan Perez, Jay Yoon Lee, Lizhen
Bräunen, Lazaros Polymenakos,
and Andrew
McCallum. 2021. Case-based reasoning for nat-
ural language queries over knowledge bases.
In Proceedings of
Die 2021 Conference on
Empirical Methods in Natural Language Pro-
Abschließen, pages 9594–9611, Online and Punta
Cana, Dominican Republic. Association for
Computerlinguistik.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, Und
Kristina Toutanova. 2019. BERT: Pre-training
of deep bidirectional transformers for language
Verständnis. In Proceedings of
Die 2019
Conference of the North American Chapter
of the Association for Computational Linguis-
Tics: Human Language Technologies, Volumen 1
(Long and Short Papers), pages 4171–4186,
Minneapolis, Minnesota. Association for Com-
putational Linguistics.
Martin d’Hoffschmidt, Wacim Belblidia, Quentin
Heinrich, Tom Brendl´e, and Maxime Vidal.
2020. FQuAD: French question answering
dataset. In Findings of
the Association for
Computerlinguistik: EMNLP 2020,
pages 1193–1208, Online. Association for Com-
putational Linguistics. https://doi.org/10
.18653/v1/2020.findings-emnlp.107
Dennis Diefenbach, Thomas Pellissier Tanon,
K. Singh, and P. Maret. 2017. Question an-
swering benchmarks for Wikidata. In Interna-
tional Semantic Web Conference.
Xin Dong and Gerard de Melo. 2019. A robust
self-learning framework for cross-lingual text
classification. In Proceedings of the 2019 Con-
ference on Empirical Methods in Natural Lan-
guage Processing and the 9th International
Joint Conference on Natural Language Pro-
Abschließen (EMNLP-IJCNLP), pages 6306–6310,
Hongkong, China. Association for Computa-
tional Linguistics. https://doi.org/10
.18653/v1/D19-1658
Dheeru Dua, Yizhong Wang, Pradeep Dasigi,
Gabriel Stanovsky, Sameer Singh, and Matt
Gardner. 2019. DROP: A reading comprehen-
sion benchmark requiring discrete reasoning
over paragraphs. In Proceedings of the 2019
Conference of the North American Chapter
of the Association for Computational Linguis-
Tics: Human Language Technologies, Volumen 1
(Long and Short Papers), pages 2368–2378,
Minneapolis, Minnesota. Association for Com-
putational Linguistics.
Philipp Dufter and Hinrich Sch¨utze. 2020. Iden-
tifying elements essential for BERT’s multilin-
guality. In Proceedings of the 2020 Conference
on Empirical Methods in Natural Language
Processing (EMNLP), pages 4423–4437, Online.
Verein für Computerlinguistik.
https://doi.org/10.18653/v1/2020
.emnlp-main.358
Nicholas Evans and Stephen C. Levinson.
2009. The myth of language universals: Lan-
guage diversity and its importance for cogni-
tive science. Behavioral and Brain Sciences,
32(5):429–448. https://doi.org/10.1017
/S0140525X0999094X
950
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Yu Gai, Paras
Jain, Wendi Zhang,
Joseph
Gonzalez, Dawn Song, and Ion Stoica. 2021.
Grounded graph decoding improves composi-
tional generalization in question answering. In
Findings of the Association for Computational
Linguistik: EMNLP 2021, pages 1829–1838,
Punta Cana, Dominican Republic. Association
für Computerlinguistik. https://doi
.org/10.18653/v1/2021.findings
-emnlp.157
Emily Goodwin, Siva Reddy, Timothy J.
O’Donnell, and Dzmitry Bahdanau. 2021. Com-
positional generalization in dependency parsing.
Yu Gu, Sue Kase, Michelle Vanni, Brian Sadler,
Percy Liang, Xifeng Yan, and Yu Su. 2021.
Beyond iid: Three levels of generalization
for question answering on knowledge bases.
In Proceedings of the Web Conference 2021,
pages 3477–3488. https://doi.org/10
.1145/3442381.3449992
Yinuo Guo, Zeqi Lin, Jian-Guang Lou, Und
Dongmei Zhang. 2020. Hierarchical poset
decoding for compositional generalization in
Sprache. Advances in Neural
Information
Processing Systems, 33:6913–6924.
Yinuo Guo, Hualei Zhu, Zeqi Lin, Bei Chen,
Jian-Guang Lou, and Dongmei Zhang. 2021.
Revisiting iterative back-translation from the
perspective of compositional generalization. In
AAAI’21.
Daniel Hershcovich, Stella Frank, Heidekraut
Lent, Miryam de Lhoneux, Mostafa Abdou,
Stephanie Brandl, Emanuele Bugliarello, Laura
Cabello Piqueras, Ilias Chalkidis, Ruixiang
Cui, Constanza Fierro, Katerina Margatina,
Phillip Rust, and Anders Søgaard. 2022A. Chal-
lenges and strategies in cross-cultural NLP. In
Proceedings of the 60th Annual Meeting of
the Association for Computational Linguistics
(Volumen 1: Long Papers), pages 6997–7013,
Dublin, Ireland. Association for Computational
Linguistik. https://doi.org/10.18653
/v1/2022.acl-long.482
Daniel Hershcovich, Nicolas Webersinke,
Mathias Kraus, Julia Anna Bingler, and Markus
Leippold. 2022B. Towards climate awareness
in NLP research. arXiv preprint arXiv:2205.
05071.
Jonathan Herzig and Jonathan Berant. 2017. Neu-
ral semantic parsing over multiple knowledge-
951
the 55th Annual
bases. In Proceedings of
the Association for Computa-
Meeting of
tional Linguistics (Volumen 2: Short Papers),
pages 623–628, Vancouver, Kanada. Associa-
tion for Computational Linguistics. https://
doi.org/10.18653/v1/P17-2098
Jonathan Herzig and Jonathan Berant. 2021.
Span-based semantic parsing for compositional
generalization. In Proceedings of the 59th An-
nual Meeting of the Association for Computa-
tional Linguistics and the 11th International
Joint Conference on Natural Language Process-
ing (Volumen 1: Long Papers), pages 908–921,
Online. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/2021.acl-long.74
Jonathan Herzig, Peter Shaw, Ming-Wei Chang,
Kelvin Guu, Panupong Pasupat, and Yuan
Zhang. 2021. Unlocking compositional gen-
eralization in pre-trained models using inter-
mediate representations. arXiv preprint arXiv:
2104.07478.
Sepp Hochreiter and J¨urgen Schmidhuber. 1997.
Long short-term memory. Neural computation,
9(8):1735–1780. https://doi.org/10.1162
/neco.1997.9.8.1735
Aidan Hogan, Eva Blomqvist, Michael Cochez,
Claudia d’Amato, Gerard de Melo, Claudio
Gutierrez, Jos´e Emilio Labra Gayo, S. Kirrane,
Sebastian Neumaier, Axel Polleres, R. Navigli,
Axel-Cyrille Ngonga Ngomo, Sabbir M. Rashid,
Anisa Rula, Lukas Schmelzeisen, Juan Sequeda,
Steffen Staab, and Antoine Zimmermann. 2021.
Die
Knowledge graphs. Communications of
ACM, 64:96–104. https://doi.org/10
.1145/3418294
Junjie Hu, Sebastian Ruder, Aditya Siddhant,
Graham Neubig, Orhan Firat, and Melvin
Johnson. 2020. XTREME: A massively
multilingual multi-task benchmark for eval-
uating cross-lingual generalisation. In Inter-
national Conference on Machine Learning,
pages 4411–4421. PMLR.
Kelvin Jiang, Dekun Wu, and Hui
Jiang.
2019. FreebaseQA: A new factoid QA data
set matching trivia-style question-answer pairs
with Freebase. In Proceedings of
Die 2019
Conference of the North American Chapter
of the Association for Computational Linguis-
Tics: Human Language Technologies, Volumen 1
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
(Long and Short Papers), pages 318–323,
Minneapolis, Minnesota. Association for Com-
putational Linguistics.
Mandar Joshi, Eunsol Choi, Daniel Weld, Und
Luke Zettlemoyer. 2017. TriviaQA: A large
scale distantly supervised challenge dataset for
reading comprehension. In Proceedings of the
55th Annual Meeting of the Association for
Computerlinguistik (Volumen 1: Long
Papers), pages 1601–1611, Vancouver, Kanada.
Verein für Computerlinguistik.
https://doi.org/10.18653/v1/P17
-1147
Pratik Joshi, Sebastin Santy, Amar Budhiraja,
Kalika Bali, and Monojit Choudhury. 2020.
The state and fate of linguistic diversity and in-
clusion in the NLP world. In Proceedings of
the 58th Annual Meeting of the Association for
Computerlinguistik, pages 6282–6293,
Online. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/2020.acl-main.560
Lucie-Aim´ee Kaffee, Kemele M. Endris, Elena
Simperl, and Maria-Esther Vidal. 2019. Rank-
ing knowledge graphs by capturing knowledge
about languages and labels. In Proceedings of
the 10th International Conference on Knowl-
edge Capture, K-CAP 2019, Marina Del Rey,
CA, USA, November 19–21, 2019. ACM.
https://doi.org/10.1145/3360901
.3364443
Manoj Prabhakar Kannan Ravi, Kuldeep Singh,
Isaiah Onando Mulang’, Saeedeh Shekarpour,
Johannes Hoffart, and Jens Lehmann. 2021.
CHOLAN: A modular approach for neural
entity linking on Wikipedia and Wikidata.
the 16th Conference of
In Proceedings of
the European Chapter of
the Association
für Computerlinguistik: Main Volume,
pages 504–514, Online. Association for Com-
putational Linguistics. https://doi.org
/10.18653/v1/2021.eacl-main.40
Daniel Keysers, Nathanael Sch¨arli, Nathan Scales,
Hylke Buisman, Daniel Furrer, Sergii Kashubin,
Nikola Momchev, Danila Sinopalnikov, Lukasz
Stafiniak, Tibor Tihon, Dmitry Tsarkov, Xiao
Wang, Marc van Zee, and Olivier Bousquet.
2020. Measuring compositional generaliza-
tion: A comprehensive method on realistic
Daten. In International Conference on Learning
Darstellungen.
Joo-Kyung Kim, Young-Bum Kim, Ruhi Sarikaya,
and Eric Fosler-Lussier. 2017. Cross-lingual
transfer learning for POS tagging without cross-
lingual resources. In Proceedings of the 2017
Conference on Empirical Methods in Natu-
ral Language Processing, pages 2832–2838,
Copenhagen, Denmark. Association for Com-
putational Linguistics.
Najoung Kim and Tal Linzen. 2020. COGS: A
compositional generalization challenge based
on semantic interpretation. In Proceedings of
Die 2020 Conference on Empirical Methods
in Natural Language Processing (EMNLP),
pages 9087–9105, Online. Association for
Computerlinguistik.
Vladislav Korablinov and Pavel Braslavski.
2020. RuBQ: A Russian dataset for question
answering over Wikidata. In International Se-
mantic Web Conference. https://doi.org
/10.1007/978-3-030-62466-8_7
Brenden M. Lake and Marco Baroni. 2018.
Generalization without systematicity: Auf der
compositional skills of sequence-to-sequence
recurrent networks. In Proceedings of the 35th
International Conference on Machine Learn-
ing, ICML 2018, Stockholmsm¨assan, Stock-
holm, Schweden, July 10–15, 2018, Volumen 80 von
Proceedings of Machine Learning Research,
pages 2879–2888. PMLR.
Jens Lehmann, Robert Isele, Max Jakob, Anja
Jentzsch, Dimitris Kontokostas, Pablo N.
Sebastian Hellmann, Mohamed
Mendes,
Morsey, Patrick van Kleef, S¨oren Auer, Und
Christian Bizer. 2015. DBpedia – a large-
scale, multilingual knowledge base extracted
from Wikipedia. Semantic Web, 6(2):167–195.
https://doi.org/10.3233/SW-140134
Yaobo Liang, Nan Duan, Yeyun Gong, Ning Wu,
Fenfei Guo, Weizhen Qi, Ming Gong, Linjun
Shou, Daxin Jiang, Guihong Cao, Xiaodong
Fan, Ruofei Zhang, Rahul Agrawal, Edward
Cui, Sining Wei, Taroon Bharti, Ying Qiao,
Jiun-Hung Chen, Winnie Wu, Shuguang Liu,
Fan Yang, Daniel Campos, Rangan Majumder,
and Ming Zhou. 2020. XGLUE: A new bench-
for cross-lingual pre-training,
mark dataset
understanding and generation. In Proceedings
of the 2020 Conference on Empirical Methods
952
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
in Natural Language Processing (EMNLP),
pages 6008–6018, Online. Association for
Computerlinguistik. https://doi.org
/10.18653/v1/2020.emnlp-main.484
(CoNLL), pages 87–98, Hongkong, China.
Verein für Computerlinguistik.
https://doi.org/10.18653/v1/K19
-1009
Yu-Hsiang Lin, Chian-Yu Chen, Jean Lee, Zirui
Li, Yuyan Zhang, Mengzhou Xia, Shruti
Rijhwani, Junxian He, Zhisong Zhang, Xuezhe
Ma, Antonios Anastasopoulos, Patrick Littell,
and Graham Neubig. 2019. Choosing trans-
fer languages for cross-lingual
learning. In
Proceedings of the 57th Annual Meeting of
the Association for Computational Linguistics,
pages 3125–3135, Florence, Italien. Association
für Computerlinguistik.
Fangyu Liu, Emanuele Bugliarello, Edoardo
Maria Ponti, Siva Reddy, Nigel Collier, Und
Desmond Elliott. 2021A. Visually grounded
reasoning across languages and cultures. In Pro-
ceedings of the 2021 Conference on Empiri-
cal Methods in Natural Language Processing,
pages 10467–10485, Online and Punta Cana,
Dominican Republic. Association for Compu-
tational Linguistics.
Jiangming Liu, Shay B. Cohen, Mirella Lapata,
and Johan Bos. 2021B. Universal Discourse
Representation Structure Parsing. Computa-
tional Linguistics, 47(2):445–476.
Ngonga Ngomo. 2018. 9th challenge on question
answering over linked data (qald-9). Language,
7(1):58–64.
Massimo Nicosia, Zhongdi Qu, and Yasemin
Altun. 2021. Translate & Fill: Improving zero-
shot multilingual semantic parsing with syn-
thetic data. In Findings of
the Association
für Computerlinguistik: EMNLP 2021,
pages 3272–3284, Punta Cana, Dominican
Republik. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/2021.findings-emnlp.279
Binling Nie, Ruixue Ding, Pengjun Xie, Fei
Huang, Chen Qian, and Luo Si. 2021.
Knowledge-aware named entity recognition with
alleviating heterogeneity. Verfahren der
AAAI Conference on Artificial Intelligence.
Mitja Nikolaus, Mostafa Abdou, Matthew Lamm,
Rahul Aralikatte, and Desmond Elliott. 2019.
Compositional generalization in image cap-
tioning. In Proceedings of the 23rd Conference
on Computational Natural Language Learning
Stephan Oepen, Omri Abend, Lasha Abzianidze,
Johan Bos, Jan Hajic, Daniel Hershcovich, Bin
Li, Tim O’Gorman, Nianwen Xue, and Daniel
Zeman. 2020. MRP 2020: The second shared
task on cross-framework and cross-lingual
meaning representation parsing. In Proceedings
von
the CoNLL 2020 Shared Task: Cross-
Framework Meaning Representation Parsing,
pages 1–22, Online. Association for Computa-
tional Linguistics. https://doi.org/10
.18653/v1/2020.conll-shared.1
Inbar Oren, Jonathan Herzig, Nitish Gupta, Matt
Gardner, and Jonathan Berant. 2020. Improv-
ing compositional generalization in semantic
parsing. In Findings of
the Association for
Computerlinguistik: EMNLP 2020,
pages 2482–2495, Online. Association for
Computerlinguistik. https://doi.org
/10.18653/v1/2020.findings-emnlp.225
Kishore Papineni, Salim Roukos, Todd Ward,
and Wei-Jing Zhu. 2002. BLEU: A method for
automatic evaluation of machine translation. In
Proceedings of the 40th Annual Meeting of
the Association for Computational Linguistics,
pages 311–318, Philadelphia, Pennsylvania,
USA. Verein für Computerlinguistik.
https://doi.org/10.3115/1073083
.1073135
Thomas Pellissier Tanon, Denny Vrandeˇci´c,
Sebastian Schaffert, Thomas Steiner, and Lydia
Pintscher. 2016. From Freebase to Wikidata:
The great migration. In Proceedings of the
25th International Conference on World Wide
Netz, WWW ’16, pages 1419–1428. https://
doi.org/10.1145/2872427.2874809
Aleksandr
Perevalov, Dennis Diefenbach,
Ricardo Usbeck, and Andreas Both. 2022A.
QALD-9-plus: A multilingual dataset for ques-
tion answering over DBpedia and Wikidata
translated by native speakers. In 2022 IEEE
16th International Conference on Semantic
Computing (ICSC). IEEE. https://doi.org
/10.1109/ICSC52841.2022.00045
Aleksandr Perevalov, Axel-Cyrille Ngonga
Ngomo, and Andreas Both. 2022B. Enhancing
the accessibility of knowledge graph question
953
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
answering systems through multilingualization.
In 2022 IEEE 16th International Conference
on Semantic Computing (ICSC), pages 251–256.
https://doi.org/10.1109/ICSC52841
.2022.00048
Jonas Pfeiffer, Ivan Vuli´c, Iryna Gurevych, Und
Sebastian Ruder. 2020. MAD-X: An Adapter-
Based Framework for Multi-Task Cross-Lingual
Transfer. In Proceedings of the 2020 Confer-
ence on Empirical Methods in Natural Lan-
guage Processing (EMNLP), pages 7654–7673,
Online. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/2020.emnlp-main.617
Matt Post. 2018. A call for clarity in reporting
the Third
BLEU scores. In Proceedings of
Conference on Machine Translation: Forschung
Papers, pages 186–191, Brussels, Belgien.
Verein für Computerlinguistik.
https://doi.org/10.18653/v1/W18
-6319
Colin Raffel, Noam Shazeer, Adam Roberts,
Katherine Lee, Sharan Narang, Michael
Matena, Yanqi Zhou, Wei Li, und Peter J. Liu.
2020. Exploring the limits of transfer learning
with a unified text-to-text transformer. Zeitschrift
of Machine Learning Research, 21(140):1–67.
Pranav Rajpurkar,
Text.
In Proceedings of
Jian Zhang, Konstantin
Lopyrev, and Percy Liang. 2016. SQuAD:
100,000+ questions for machine comprehen-
Die 2016
Version von
Conference on Empirical Methods in Natural
Language Processing, pages 2383–2392, Austin,
Texas. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/D16-1264
Mohammad Sadegh Rasooli and Michael Collins.
2017. Cross-lingual syntactic transfer with
limited resources. Transactions of the Associa-
tion for Computational Linguistics, 5:279–293.
https://doi.org/10.1162/tacl a 00061
Sebastian Ruder, Noah Constant, Jan Botha,
Jinlan Fu,
Aditya Siddhant, Orhan Firat,
Junjie Hu, Dan Garrette,
Pengfei Liu,
Graham Neubig, and Melvin Johnson. 2021.
XTREME-R: Towards more challenging and
nuanced multilingual evaluation. In Proceed-
Die 2021 Conference on Empirical
ings of
Methods in Natural Language Processing,
pages 10215–10245, Online and Punta Cana,
954
Dominican Republic. Association for Compu-
tational Linguistics. https://doi.org/10
.18653/v1/2021.emnlp-main.802
Laura Ruis, Jacob Andreas, Marco Baroni, Diane
Bouchacourt, and Brenden M. Lake. 2020.
A benchmark for systematic generalization in
grounded language understanding. Advances
in Neural Information Processing Systems,
33:19861–19872.
Ivan Rybin, Vladislav Korablinov, Pavel Efimov,
and Pavel Braslavski. 2021. RuBQ 2.0: Ein
innovated Russian question answering data-
set. In Eighteenth Extended Semantic Web
Conference – Resources Track. https://doi
.org/10.1007/978-3-030-77385-4 32
Amrita Saha, Vardaan Pahuja, Mitesh M. Khapra,
Karthik Sankaranarayanan, and A. P. S. Chandar.
2018. Complex sequential question answer-
ing: Towards learning to converse over linked
question answer pairs with a knowledge graph.
In AAAI. https://doi.org/10.1609/aaai
.v32i1.11332
Cezar Sas, Meriem Beloucif, and Anders Søgaard.
2020. WikiBank: Using Wikidata to improve
multilingual frame-semantic parsing. In Pro-
ceedings of
the 12th Language Resources
and Evaluation Conference, pages 4183–4189,
Marseille, Frankreich. European Language Re-
sources Association.
Bo Shao, Yeyun Gong, Weizhen Qi, Nan Duan,
and Xiaola Lin. 2020. Multi-level alignment
pretraining for multi-lingual semantic pars-
ing. In Proceedings of the 28th International
Conference on Computational Linguistics,
pages 3246–3256, Barcelona, Spanien (Online).
International Committee on Computational Lin-
guistics. https://doi.org/10.18653/v1
/2020.coling-main.289
Chih Chieh Shao, Trois Liu, Yuting Lai, Yiying
Tseng, and Sam Tsai. 2018. DRCD: A Chinese
machine reading comprehension dataset.
Tao Shen, Xiubo Geng, Tao Qin, Daya Guo, Duyu
Tang, Nan Duan, Guodong Long, and Daxin
Jiang. 2019. Multi-task learning for conversa-
tional question answering over a large-scale
knowledge base. In Proceedings of the 2019
Conference on Empirical Methods in Natural
Language Processing and the 9th International
Joint Conference on Natural Language Pro-
Abschließen (EMNLP-IJCNLP), pages 2442–2451,
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3
Ashish Vaswani, Samy Bengio, Eugene Brevdo,
Francois Chollet, Aidan N. Gomez, Stephan
Gouws, Llion Jones, Łukasz Kaiser, Nal
Kalchbrenner, Niki Parmar, Ryan Sepassi,
Noam Shazeer, and Jakob Uszkoreit. 2018.
Tensor2tensor for neural machine translation.
CoRR, abs/1803.07416.
Funtowicz,
Joe Davison,
Thomas Wolf, Lysandre Debut, Victor Sanh,
Julien Chaumond, Clement Delangue, Anthony
Moi, Pierric Cistac, Tim Rault, Remi Louf,
Morgan
Sam
Shleifer, Patrick von Platen, Clara Ma, Yacine
Jernite, Julien Plu, Canwen Xu, Teven Le
Scao, Sylvain Gugger, Mariama Drame,
Quentin Lhoest, and Alexander Rush. 2020.
Transformers: State-of-the-art natural language
Die 2020
In Proceedings of
Verarbeitung.
Conference on Empirical Methods in Natural
Language Processing: Systemdemonstrationen,
pages 38–45, Online. Association for Com-
putational Linguistics. https://doi.org
/10.18653/v1/2020.emnlp-demos.6
Shijie Wu and Mark Dredze. 2020. Are all lan-
guages created equal in multilingual BERT? In
Proceedings of the 5th Workshop on Representa-
tion Learning for NLP, pages 120–130, Online.
Verein für Computerlinguistik.
Linting Xue, Noah Constant, Adam Roberts,
Mihir Kale, Rami Al-Rfou, Aditya Siddhant,
Aditya Barua, and Colin Raffel. 2020. mT5:
A massively multilingual pre-trained text-to-
text transformer. arXiv preprint arXiv:2010.
11934.
Jingfeng Yang, Federico Fancellu, Bonnie
Webber, and Diyi Yang. 2021. Frustrat-
ingly simple but surprisingly strong: Using
language-independent features for zero-shot
cross-lingual semantic parsing. In Proceed-
ings of
Die 2021 Conference on Empirical
Methods in Natural Language Processing,
pages 5848–5856, Online and Punta Cana,
Dominican Republic. Association for Computa-
tional Linguistics. https://doi.org/10
.18653/v1/2021.emnlp-main.472
Hongkong, China. Association for Computa-
tional Linguistics. https://doi.org/10
.18653/v1/D19-1248
Tom Sherborne and Mirella Lapata. 2022.
Zero-shot cross-lingual semantic parsing. In
Proceedings of the 60th Annual Meeting of
the Association for Computational Linguistics
(Volumen 1: Long Papers), pages 4134–4153,
Dublin, Ireland. Association for Computational
Linguistik. https://doi.org/10.18653
/v1/2022.acl-long.285
David So, Quoc Le, and Chen Liang. 2019. Der
evolved transformer. In Proceedings of the 36th
International Conference on Machine Learn-
ing, Volumen 97 of Proceedings of Machine
Learning Research, pages 5877–5886. PMLR.
Alon Talmor and Jonathan Berant. 2018. Der
web as a knowledge-base for answering com-
plex questions. In Proceedings of the 2018
Conference of the North American Chapter
of the Association for Computational Linguis-
Tics: Human Language Technologies, Volumen 1
(Long Papers), pages 641–651, New Orleans,
Louisiana. Association for Computational Lin-
guistics. https://doi.org/10.18653/v1
/N18-1059
James Thorne, Majid Yazdani, Marzieh Saeidi,
Fabrizio Silvestri, Sebastian Riedel, and Alon
Halevy. 2021. Database reasoning over text.
In Proceedings of the 59th Annual Meeting of
the Association for Computational Linguistics
and the 11th International Joint Conference
on Natural Language Processing (Volumen 1:
Long Papers), pages 3091–3104, Online.
Verein für Computerlinguistik.
https://doi.org/10.18653/v1/2021
.acl-long.241
Dmitry Tsarkov, Tibor Tihon, Nathan Scales,
Nikola Momchev, Danila Sinopalnikov, Und
Nathanael Sch¨arli. 2021. *-CFQ: Analyzing the
scalability of machine learning on a composi-
tional task. Proceedings of the AAAI Conference
on Artificial Intelligence.
955
l
D
Ö
w
N
Ö
A
D
e
D
F
R
Ö
M
H
T
T
P
:
/
/
D
ich
R
e
C
T
.
M
ich
T
.
e
D
u
/
T
A
C
l
/
l
A
R
T
ich
C
e
–
P
D
F
/
D
Ö
ich
/
.
1
0
1
1
6
2
/
T
l
A
C
_
A
_
0
0
4
9
9
2
0
4
2
5
8
8
/
/
T
l
A
C
_
A
_
0
0
4
9
9
P
D
.
F
B
j
G
u
e
S
T
T
Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3