Sequence-Level Training for
Non-Autoregressive Neural
机器翻译
Chenze Shao
Key Laboratory of Intelligent
Information Processing
Institute of Computing Technology
Chinese Academy of Sciences
shaochenze18z@ict.ac.cn
Yang Feng
Key Laboratory of Intelligent
Information Processing
Institute of Computing Technology
Chinese Academy of Sciences
fengyang@ict.ac.cn
Jinchao Zhang
Pattern Recognition Center
WeChat AI, Tencent Inc
dayerzhang@tencent.com
Fandong Meng
Pattern Recognition Center
WeChat AI, Tencent Inc
fandongmeng@tencent.com
Jie Zhou
Pattern Recognition Center
WeChat AI, Tencent Inc
withtomzhou@tencent.com
最近几年, Neural Machine Translation (NMT) has achieved notable results in various
translation tasks. 然而, the word-by-word generation manner determined by the autoregres-
sive mechanism leads to high translation latency of the NMT and restricts its low-latency ap-
并发症. Non-Autoregressive Neural Machine Translation (NAT) removes the autoregressive
提交材料已收到: 8 六月 2021; 收到修订版: 18 八月 2021; 接受出版:
26 八月 2021.
https://doi.org/10.1162/COLI 00421
© 2021 计算语言学协会
根据知识共享署名-非商业性-禁止衍生品发布 4.0 国际的
(CC BY-NC-ND 4.0) 执照
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
mechanism and achieves significant decoding speedup by generating target words independently
and simultaneously. 尽管如此, NAT still takes the word-level cross-entropy loss as the
training objective, which is not optimal because the output of NAT cannot be properly evaluated
due to the multimodality problem. 在本文中, we propose using sequence-level training
objectives to train NAT models, which evaluate the NAT outputs as a whole and correlates
well with the real translation quality. 第一的, we propose training NAT models to optimize
sequence-level evaluation metrics (例如, 蓝线) based on several novel reinforcement algorithms
customized for NAT, which outperform the conventional method by reducing the variance of
gradient estimation. 第二, we introduce a novel training objective for NAT models, which aims
to minimize the Bag-of-N-grams (BoN) difference between the model output and the reference
句子. The BoN training objective is differentiable and can be calculated efficiently without
doing any approximations. 最后, we apply a three-stage training strategy to combine these two
methods to train the NAT model. We validate our approach on four translation tasks (WMT14
En↔De, WMT16 En↔Ro), which shows that our approach largely outperforms NAT baselines
and achieves remarkable performance on all translation tasks. The source code is available at
https://github.com/ictnlp/Seq-NAT.
1. 介绍
Machine translation used to be one of the most challenging tasks in natural language
加工, but recent advances in neural machine translation make it possible to
translate with an end-to-end model architecture. NMT models are typically built on
the encoder-decoder framework. The encoder network encodes the source sentence to
distributed representations, and the decoder network reconstructs the target sentence
from these representations in an autoregressive manner. The target sentence is gener-
ated word-by-word where the previously predicted words are fed back to the decoder
as context. In the past few years, autoregressive NMT models have achieved notable
results in various translation tasks (Cho et al. 2014; 吸勺, Vinyals, and Le 2014;
Bahdanau, 给, and Bengio 2015; Wu et al. 2016; Vaswani et al. 2017). 然而, 这
word-by-word generation manner determined by the autoregressive mechanism leads
to high translation latency of the NMT and restricts its low-latency applications.
Non-Autoregressive Neural Machine Translation (NAT) (Gu et al. 2018) is proposed
to reduce the latency of NMT. By removing the autoregressive mechanism, NAT can
generate target words independently and simultaneously, thereby achieving significant
decoding speedup. 尽管如此, NAT still takes the word-level cross-entropy loss as
the training objective, which is not optimal because the output of NAT cannot be
properly evaluated. Due to the multimodality of language, the reference sentence may
have many variants that are composed of different words but have the same seman-
抽动症. For the autoregressive model, the teacher forcing algorithm (Williams and Zipser
1989) can provide it with sequential information that guides the model to generate
the reference sentence. 然而, the sequential information is not available during
the training of NAT, so NAT may generate any translation variant with the target
语义学. Once the NAT model tends to generate a variant that is not aligned verbatim
with the reference sentence, the cross-entropy loss will give it a large penalty with
no regard to the translation quality. 最后, the correlation between the cross-
entropy loss and translation quality becomes weak, which has a negative impact on the
NAT performance.
892
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
数字 1
One example of NAT that is not aligned verbatim with the reference sentence. The red line
indicates the misalignment that will receive a large penalty from the word-level cross-entropy
loss. The purple arrow indicates the possible overcorrection error.
桌子 1
A translation case on the validation set of WMT14 De-En. Source and Target are the source
sentence and reference sentence, 分别. AT and NAT are the output of the autoregressive
Transformer and non-autoregressive Transformer, 分别.
来源
Es gibt Krebsarten, die aggressiv und andere, die indolent sind.
参考
There are aggressive cancers and others that are indolen.
AT
NAT
There are cancers that are aggressive and others that are indolent.
There are cancers cancer aggressive aggressive others are indindent.
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
如图 1, though the translation “I have to get up and start working.”
has similar semantics to the reference sentence, the word-level cross-entropy loss will
give it a large penalty because it is not aligned verbatim with the reference sentence.
Under the guidance of cross-entropy loss, the translation may be further corrected to
“I have to up up start start working.”. This is preferred by the cross-entropy loss but
the translation quality will actually get worse, which is called the overcorrection error
(张等人. 2019). The essential reason for the overcorrection error is that the loss
function evaluates the generation quality of each position independently and does not
model the sequential dependency. 因此, NAT tends to focus on local correctness
while ignoring the overall translation quality, and therefore generates influent transla-
tions with many over-translation and under-translation errors. 如表所示 1,
the output of NAT is incomplete and contains repeated words like “cancer” and
“aggressive.”
在本文中, we propose using sequence-level training objectives to train NAT
型号, which evaluate the NAT outputs as a whole and correlate well with the real
translation quality. 第一的, we propose training NAT models to optimize sequence-level
evaluation metrics (例如, 蓝线 [Papineni et al. 2002], GLEU [Wu et al. 2016], and ROUGE
[林 2004]). These metrics are usually non-differentiable, and reinforcement learning
技巧 (Sutton 1984; 威廉姆斯 1992; Sutton et al. 1999) are widely applied to train
autoregressive NMT to optimize these discrete objectives (Ranzato et al. 2016; Bahdanau
等人. 2017). 然而, the training procedure is usually unstable due to the high variance
of the gradient estimation. Using the appealing characteristics of non-autoregressive
一代, we propose several novel reinforcement algorithms customized for NAT,
which outperform the conventional method by reducing the variance of gradient es-
timation. 第二, we introduce a novel training objective for NAT models, 哪个
893
计算语言学
体积 47, 数字 4
aims to minimize the Bag-of-N-grams (BoN) difference between the model output and
the reference sentence. As the word-level loss cannot properly model the sequential
依赖性, we propose to evaluate the NAT output at the n-gram level. 自从
output of NAT may not be aligned verbatim with the reference, we do not require the
strict alignment and optimize the BoN for NAT. Optimizing such an objective usually
faces the difficulty of the exponential search space, and we find that the difficulty
can be overcome through using the characteristics of non-autoregressive generation. 在
summary, the BoN training objective has many appealing properties. It is differentiable
and can be calculated efficiently without doing any approximations. Most importantly,
the BoN objective correlates well with the overall translation quality, as we demonstrate
在实验中.
The reinforcement learning method can train NAT with any sequence-level objec-
主动的, but it requires a lot of calculations on the CPU to reduce the variance of gradient
estimation. The bag-of-n-grams method can efficiently calculate the BoN objective with-
out doing any approximations, but the choice of training objectives is very limited. 这
cross-entropy loss also has strengths such as high-speed training and is suitable for
model warmup. 所以, we apply a three-stage training strategy to combine the two
sequence-level training methods and the word-level training to train the NAT model.
We validate our approach on four translation tasks (WMT14 En↔De, WMT16 En↔Ro),
which shows that our approach largely outperforms NAT baselines and achieves re-
markable performance on all translation tasks.
This article extends our conference papers on non-autoregressive translation (Shao
等人. 2019, 2020) in three major directions. 第一的, we propose several novel sequence-
level training algorithms in this article. In the context of reinforcement learning, 我们
propose the traverse-based method Traverse-Ref, which practically eliminates the vari-
ance of gradient estimation and largely outperforms the best method Reinforce-Top-k
proposed in Shao et al. (2019). We also propose to use bag-of-words as the training
objective of NAT. The bag-of-words vector can be explicitly calculated, so it supports a
variety of distance metrics such as BoW-L1, BoW-L2, and BoW-Cos as loss functions,
which enables us to analyze the performance of different distance metrics on NAT.
第二, we explore the combination of the reinforcement learning based method and
the bag-of-n-grams method and propose a three-stage training strategy to better com-
bine their advantages. 最后, we conduct experiments on a stronger baseline model
(Ghazvininejad et al. 2019) and a larger batch size setting to show the effectiveness of
our approach, and we also provide a more detailed analysis. The article is structured
as follows. We explain the vanilla non-autoregressive translation and sequence-level
training in Section 2. We introduce our sequence-level training methods in Section 3. 我们
review the related works on non-autoregressive translation and sequence-level training
in Section 4. 在部分 5, we introduce the experimental design, conduct experiments
to evaluate the performance of our methods and conduct a series of analyses to un-
derstand the underlying key components in them. 最后, we conclude in Section 6 经过
summarizing the contributions of our work.
2. Background
2.1 Autoregressive Neural Machine Translation
Deep neural networks with the autoregressive encoder-decoder framework have
achieved state-of-the-art results on machine translation, with different choices of ar-
chitectures such as Recurrent Neural Network (RNN), Convolutional Neural Network
894
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
(CNN), and Transformer. RNN-based models (Bahdanau, 给, and Bengio 2015; 给
等人. 2014) have a sequential architecture that makes them difficult to be parallelized.
CNN (Gehring et al. 2017) and self-attention (Vaswani et al. 2017) based models have
highly parallelized architectures, which solves the parallelization problem during the
训练. 然而, during the inference, the translation has to be generated word-by-
word due to the autoregressive mechanism.
Given a source sentence X = {x1, . . . , xn} and a target sentence Y = {y1, . . . , yT}, 这
autoregressive NMT models the translation probability from X to Y sequentially as:
磷(是 |X, 我) =
时间
(西德:89)
t=1
p(yt|y
many distance metrics like L2 distance and cosine distance cannot be calculated. 为了-
tunately, we find that the L1 distance between the two BoN vectors, denoted as BoN-
L1, can be simplified using the sparsity of bag-of-n-grams. As shown in Equation (19),
for NAT, its bag-of-n-grams vector BoNθ is dense. 相反, assume that the
reference sentence is ˆY, the vector BoN ˆY is very sparse where only a few entries of it
have non-zero values. Using this property, we can write BoN-L1 as follows:
Theorem 4
BoN-L1 = 2(T − n + 1 -
(西德:88)
G
min(BoNθ(G), BoN ˆY (G)))
(21)
Proof. 第一的, we show that the L1-norm of BoNY and BoNθ are both T − n + 1:
(西德:88)
G
(西德:88)
G
T−n
(西德:88)
(西德:88)
t=0
(西德:88)
G
(西德:88)
BoNY (G) =
BoNθ(G) =
1{yt+1:t+n = g} = T − n + 1
磷(是|X, 我) · BoNY (G)
是
磷(是|X, 我) ·
G
(西德:88)
是
=
(西德:88)
G
BoNY (G) = T − n + 1
(22)
905
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
Algorithm 6 BoN-L1
输入: probability distribution p(·|X, 我)), reference sentence ˆY, prediction length T, n
输出: BoN distance BoN-L1
1: construct the bag-of-n-grams BoN ˆY for the reference sentence
2: ref-n-grams = {G|BoN ˆY (G) != 0}
3: match = 0
4: for g in ref-n-grams do
5:
6:
7: end for
8: BoN-L1 = 2 · (T-n+1-match)
9: return BoN-L1
calculate BoNθ(G) according to Equation (19)
match += min(BoNθ(G), BoN ˆY (G))
以这个为基础, we can convert BoN-L1 to the following form:
BoN-L1 =
=
(西德:88)
G
(西德:88)
G
|BoNθ(G) − BoN ˆY (G)|
(BoNθ(G) + BoN ˆY (G) - 2 min(BoNθ(G), BoN ˆY (G))
(23)
= 2(T − n + 1 -
(西德:88)
G
min(BoNθ(G), BoN ˆY (G)))
(西德:3)
The minimum between BoNθ(G) and BoN ˆY (G) can be understood as the number of
matches for the n-gram g, and the L1 distance measures the number of n-grams pre-
dicted by NAT that fails to match the reference sentence. Notice that the minimum will
be nonzero only if the n-gram g appears in the reference sentence. Hence we can only
focus on n-grams in the reference, which significantly reduces the computational effort
and storage requirement. Algorithm 6 illustrates the calculation process of BoN-L1.
We normalize distances to range [0, 1] as training objectives. For BoW distances, 我们
keep BoW-Cos unchanged and divide BoW-L1 and BoW-L2 by the constant 2T:
LBoW-L1 (我) = BoW-L1
2时间
LBoW-L2 (我) = BoW-L2
2时间
For BoN, we divide BoN-L1 by the constant 2(T − n + 1):
LBoN-L1 (我) = BoN-L1
2(T − n + 1)
(24)
(25)
3.3 Training Strategy
The reinforcement learning method can train NAT with any sequence-level objective
that correlates well with the translation quality, but it requires a lot of calculations
on CPU to reduce the variance of gradient estimation. The bag-of-n-grams method
can efficiently calculate the BoN objective without doing any approximations, 但是
training objective is limited to the L1 distance. The word-level cross-entropy loss cannot
906
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
evaluate the output of NAT properly, but it also has strengths like high-speed training
and it is suitable for model warmup.
所以, we propose to use a three-stage training strategy to combine the
strengths of the two training methods and the cross-entropy loss. 第一的, 我们使用
cross-entropy loss to pretrain the NAT model, and this process takes the most training
脚步. Then we use the bag-of-n-grams objective to finetune the pretrained model for
a few training steps. 最后, we apply the reinforcement learning method to finetune
the model to optimize the sequence-level objective, where this process takes the least
training steps. There are also other training strategies like two-stage training and joint
训练, and we will show the efficiency of three-stage training in the experiment.
The loss based on reinforcement learning or bag-of-n-grams can also be used alone to
finetune the model pretrained by the cross-entropy loss. We will adopt this strategy
when analyzing these methods separately.
4. 相关工作
4.1 Non-Autoregressive Translation
Gu et al. (2018) proposed non-autoregressive translation to reduce the translation
潜伏, which generates all target tokens simultaneously. Although accelerating the
decoding process significantly, the acceleration comes at the cost of translation quality.
所以, intensive efforts have been devoted to improving the performance of NAT,
which can be roughly divided into the following categories.
Latent Variables. NAT suffers from the multimodality problem, which can be mitigated
by introducing a latent variable to directly model the nondeterminism in the translation
过程. Gu et al. (2018) proposed to use fertility scores specifying the number of
output words each input word generates to model the latent variable. Kaiser et al.
(2018) autoencoded the target sequence into a sequence of discrete latent variables and
decoded the output sequence from the latent sequence in parallel. Based on variational
inference, 马等人. (2019) proposed FlowSeq to model sequence-to-sequence generation
using generative flow, and Shu et al. (2020) introduced LaNMT with continuous latent
variables and deterministic inference. Bao et al. (2019); Ran et al. (2021) used the
position information as latent variables to explicitly model the reordering information
in the decoding procedure.
Decoding Methods. The fully non-autoregressive transformer generates all target
words in one run, which suffers from the large performance degradation. 所以,
researchers were interested in alternative decoding methods that are slightly slower
but can significantly improve the translation quality. 李, Mansimov, and Cho (2018)
proposed the iterative decoding based NAT model IRNAT to iteratively refine the
translation where the outputs of the decoder are fed back as the decoder inputs
in the next iteration. The pattern of iterative decoding was adopted by many non-
autoregressive models. Ghazvininejad et al. (2019) and Kasai et al. (2020) refine model
output iteratively by masking part of the translation and predicting the masks in
each iteration. Gu, 王, and Zhao (2019) introduced the Levenshtein Transformer to
iteratively refine the translation with insertion and deletion operations. In addition to
iterative decoding, Sun et al. (2019) proposed to incorporate the Conditional Random
Fields in the top of NAT decoder to help the NAT decoding. 王, 张, and Chen
(2018) introduced the semi-autoregressive decoding mechanism that generates a group
907
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
of words each time. Ran et al. (2020) proposed another semi-autoregressive model
named RecoverSAT, which generates a translation as a sequence of simultaneously
generated segments.
Training Objectives. As the cross-entropy loss cannot evaluate NAT outputs properly,
researchers attempt to improve the model performance by introducing better training
目标. In addition to sequence-level training (Shao et al. 2019, 2020), Wang et al.
(2019) proposed the similarity regularization and reconstruction regularization to re-
duce errors of repeated and incomplete translations. Libovick ´y and Helcl (2018); Saharia
等人. (2020) applied the Connectionist Temporal Classification loss to marginalize out
latent alignments using dynamic programming. Ghazvininejad et al. (2020) proposed
the aligned cross-entropy loss, which uses a differentiable dynamic program based on
the best monotonic alignment between target tokens and model predictions.
Other Improvements. Besides the above mentioned categories, some researchers im-
prove the NAT performance from other perspectives. Guo et al. (2019) proposed to
enhance the inputs of NAT decoder with phrase-table lookup and embedding mapping.
Akoury, Krishna, and Iyyer (2019) introduced syntactically supervised Transformers,
which first autoregressively predict a chunked parse tree and then generate all target to-
kens conditioned on it. Zhou and Keung (2020) proposed to improve NAT performance
with source-side monolingual data. Shan, 冯, and Shao (2021) proposed to model the
coverage information for NAT. 李等人. (2019); Wei et al. (2019) improve the performance
of NAT by exploring better methods to learn from autoregressive models. 周, Gu,
and Neubig (2020) investigated the knowledge distillation technique in NAT. Tu et al.
(2020) introduced the energy-based inference networks as an alternative to knowledge
distillation.
4.2 Sequence-Level Training for Autoregressive NMT
Neural machine translation models are usually trained with the word-level loss under
the teacher forcing algorithm (Williams and Zipser 1989), which forces the model to
generate the next word based on the previous ground-truth words other than the
model outputs during the training. 然而, this training method suffers from the
exposure bias problem (Ranzato et al. 2016) because the model is exposed to different
data distributions during training and inference. To alleviate the exposure bias problem,
some researchers improve the teacher forcing algorithm to professor forcing (Goyal
等人. 2016) or seer forcing (Feng et al. 2021). Scheduled sampling (Bengio et al. 2015;
Venkatraman, Hebert, and Bagnell 2015) is the direct solution for exposure bias, 哪个
attempts to alleviate the exposure bias problem through mixing ground-truth words
and previously predicted words as inputs during the training. 然而, the generated
sequence may not be aligned with the target sequence, which is inconsistent with the
word-level loss. 所以, it is a natural solution to apply sequence-level training to
eliminate the exposure bias in the autoregressive NMT.
Sequence-level training objectives are usually non-differentiable, and reinforcement
learning techniques (威廉姆斯 1992; Sutton et al. 1999) are widely applied to train au-
toregressive NMT to optimize discrete objectives. Ranzato et al. (2016) first pointed out
the exposure bias problem and proposed the MIXER algorithm to alleviate the exposure
bias, which is a combination of the word-level cross-entropy loss and the sequence-
level loss optimized by the REINFORCE algorithm. Bahdanau et al. (2017) presented an
approach to training neural networks to generate sequences using actor-critic methods
908
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
from reinforcement learning. He et al. (2016) proposed a dual learning approach to train
the forward NMT model using reward signals provided by the backward model. 吴
等人. (2016) introduced a new sequence evaluation metrics GLEU for the sequence-
level training of Google’s NMT System. Yu et al. (2017) proposed a sequence generation
framework called SeqGAN to overcome the differentiable difficulty of GAN through
reinforcement learning, which is then applied by Wu et al. (2018乙), Yang et al. (2018) 到
train NMT under the generator-discriminator framework. Wu et al. (2018A) 实施
a systematic study on the reinforcement learning based training method for NMT.
In addition to reinforcement learning based methods, there are also some ap-
proaches that can train NMT with sequence-level objectives. 沉等人. (2016) intro-
duced Minimum Risk Training for NMT to minimize the expected risk on training
数据. Norouzi et al. (2016) proposed Reward Augmented Maximum Likelihood to in-
corporate sequence-level reward into a maximum likelihood framework. Edunov
等人. (2018) surveyed a range of classical objective functions and applied them to neural
sequence to sequence models. 马等人. (2018) proposed to optimize NMT by the bag-of-
words training objective. Shao, 陈, and Feng (2018) introduced probabilistic n-gram
matching to transform the discrete sequence-level objective into the differentiable form.
As shown above, sequence-level training has attracted much attention of re-
searchers and has been deeply studied on autoregressive models. 然而, 尽管
sequence-level training is more essential on non-autoregressive models, its application
on NAT has not been well studied before.
5. 实验
5.1 Setup
数据集. We evaluate our proposed methods on four translation tasks: WMT14
English↔German (WMT14 En↔De) and WMT16 English↔Romanian (WMT16
En↔Ro). We use the standard tokenized BLEU (Papineni et al. 2002) to evaluate the
translation quality. For WMT14 En↔De, we use the WMT 2016 corpus consisting of
4.5M sentence pairs for the training. The validation set is newstest2013 and the test
set is newstest2014. We learn a joint BPE model (Sennrich, Haddow, and Birch 2016)
with 32K operations to process the data and share the vocabulary for source and
target languages. For WMT16 En↔Ro, we use the WMT 2016 corpus consisting of
610K sentence pairs for the training. We take newsdev-2016 and newstest-2016 as
development and test sets. We learn a joint BPE model with 32K operations to process
the data and share the vocabulary for source and target languages.
Knowledge Distillation. Knowledge distillation (欣顿, Vinyals, and Dean 2015;
Kim and Rush 2016) is proved to be crucial for successfully training NAT models.
For all the translation tasks, we first train an autoregressive model as the teacher and
apply sequence-level knowledge distillation to construct the distillation corpus where
the target side of the training corpus is replaced by the output of the autoregressive
Transformer model. We use the distillation corpora to train NAT models.
基线. We take the base version of Transformer (Vaswani et al. 2017) as our
autoregressive baseline as well as the teacher model. The NAT baseline takes the same
structure as the base Transformer except that we modify the attention mask of the
decoder, so it does not mask out the future tokens. We perform uniform copy from
source embeddings (Gu et al. 2018) to construct decoder inputs. We use a target length
909
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
predictor to predict the length of the target sentence, which takes the encoder hidden
states as inputs and feeds it to a softmax classifier after an affine transformation. 我们
use the golden length during the training and the predicted length during the inference.
Rescoring. For inference, we follow the common practice of noisy parallel decoding (Gu
等人. 2018), which generates a number of decoding candidates in parallel and selects
the best translation via rescoring with the autoregressive teacher. We generate multiple
translation candidates by predicting the target length T and generate translations with
lengths ranging from [T − B, 时间 + 乙], where B is the beam size. The autoregressive
teacher calculates the cross-entropy loss of the n = 2B + 1 translations and selects the
translation with the lowest loss.
Hyperparameters. In the main experiments, we set the top-k size k to 5, the sampling
times n to 10, and the N in BoN to 2. We use the ROUGE-2 score as the reward in
reinforcement learning. In the pretraining stage of the three-stage training, 号码
of training steps is 300k for WMT14 En↔De and 150k for WMT16 En↔Ro. In the second
阶段, we use the BoN objective to finetune the model for 3k steps. In the final stage, 我们
use sequence-level evaluation metrics to finetune the model for 300 脚步. The batch size
is 128k for pretraining and 512k for finetuning. For WMT14 En↔De, we use a dropout
率 0.3 during the pretraining and a dropout rate of 0.1 during the finetuneing. 为了
WMT16 En↔Ro, we use a dropout rate of 0.3 during the pretraining and finetuneing.
We also use 0.01 L2 weight decay and label smoothing with (西德:15) = 0.1 for regularization.
We follow the weight initialization schema from BERT (Devlin et al. 2019). All models
are optimized with Adam (Kingma and Ba 2014) with β = (0.9, 0.98) 和 (西德:15) = 10−8. 这
learning rate warms up to 5 · 10−4 within 10k steps, and then decays with the inverse
square-root schedule. We use 8 GeForce RTX 3090 GPUs for the training.
5.2 Main Results
We first compare the performance of our proposed methods,
including the Re-
inforcement Learning (RL) based methods (IE。, Reinforce-Base, Reinforce-Step,
Reinforce-Top-k, and Traverse-Ref) and the Bag-of-N-grams (BoN) based methods (IE。,
BoW-Cos, BoW-L2, BoW-L1, and BoN-L1 (N= 2)). We also adopt the three-stage training
strategy to combine the best performing methods of the above two categories (IE。,
Traverse-Ref and BoN-L1 (N= 2)), which is denoted as BoN+RL. 桌子 3 reports the
experiment results of our methods, from which we have the following observations.
training can effectively improve the performance of non-
1. Sequence-level
autoregressive models. All the methods listed in Table 3 can greatly improve the
translation quality of non-autoregressive models. Even the simplest method Reinforce-
Base achieves more than 3 BLEU improvements on the WMT14 data set, indicating that
sequence-level training is very suitable for non-autoregressive models.
2. The methods we propose for variance reduction are helpful to enhance the
performance of the reinforcement learning. Comparing the reinforcement learning
based methods, Reinforce-Step reduces the estimation variance by replacing the
sentence reward with step reward, which improves Reinforce-Base by about 0.5 蓝线.
Reinforce-Top-k further improves Reinforce-Step by about 0.8 BLEU by eliminating the
variance of important words. 最后, Traverse-Ref gives a method to traverse the whole
search space for Reference-Based rewards, which improves Reinforce-Top-k by about
910
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
0.4 蓝线. 总之, the methods we propose for variance reduction are helpful to
enhance the performance of reinforcement learning.
3. Among the three BoW training objectives, the L1 distance is very suitable for the
training of non-autoregressive models. Comparing the three Bag-of-Words objectives,
BoW-L1 achieves the best performance and largely outperforms the other two
目标, indicating that the L1 distance of BoW is very suitable for the training
of non-autoregressive models. Regarding the bag-of-n-grams objective, the main
limitation is that many distance metrics like L2 distance and cosine distance cannot be
calculated, and the observation on BoW can alleviate this concern to some extent.
4. Three-stage training can effectively combine reinforcement learning and bag-of-n-
克. Three-stage training achieves the best performance by combining the best meth-
ods of the two categories (IE。, Traverse-Ref and BoN-L1 (N= 2)), which improves the
NAT baseline by more than 5 BLEU scores on the WMT14 data set and more than 2
BLEU scores on the WMT16 data set. We use Seq-NAT to represent this method.
5.3 Sequence-Level Training for Iterative NAT
In the previous section, we have verified the effect of sequence-level training on the
vanilla NAT, which is non-iterative and uses a single forward pass during the decoding.
在这个部分, we conduct experiments to evaluate the effect of sequence-level training
on iterative NAT, which is an important class of NAT models. We use the Conditional
Masked Language Model (CMLM) with mask-predict decoding (Ghazvininejad et al.
2019) as our baseline model, which is a strong iterative NAT model. We apply sequence-
level training to finetune the CMLM baseline and call this method Seq-CMLM.
数字 2 shows the BLEU scores of CMLM and Seq-CMLM under a different number of
迭代.
桌子 3
The performance (test set BLEU) of our methods on all of our benchmarks. All models except
Transformer are purely non-autoregressive, using a single forward pass during the argmax
decoding. NAT-Base is the baseline NAT model. All other NAT models are finetuned from the
NAT-Base.
模型
Speedup
WMT14
WMT16
EN-DE DE-EN EN-RO RO-EN
Base
RL
BoN
Transformer
NAT-Base
Reinforce-Base
Reinforce-Step
Reinforce-Top-k
Traverse-Ref
BoW-Cos
BoW-L2
BoW-L1
BoN-L1 (N= 2)
3-Stage RL+BoN
1.0×
15.6×
15.6×
15.6×
15.6×
15.6×
15.6×
15.6×
15.6×
15.6×
15.6×
27.42
19.51
23.23
23.76
24.68
25.15
23.90
24.22
24.75
25.28
25.54
31.63
24.47
27.59
28.08
29.05
29.50
28.11
29.03
29.41
29.66
29.91
34.18
28.89
29.67
30.14
30.73
31.12
29.72
30.08
31.01
31.37
31.69
33.72
29.35
29.85
30.31
30.88
31.34
29.69
29.93
31.19
31.51
31.78
911
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
数字 2
BLEU scores of CMLM and Seq-CMLM on the test set of WMT14 En-De under a different
number of iterations. For the elegance of the figure, we removed the 1 iteration result, 这是
17.47 for CMLM and 24.90 for Seq-CMLM.
From Figure 2, we can see that Seq-CMLM consistently outperforms CMLM on all
numbers of iterations. 即使 10 迭代, Seq-CMLM can achieve an improvement
的 0.42 BLEU to CMLM, reaching a BLEU score of 27.36, showing that sequence-level
training is also very effective on Iterative NAT.
5.4 Speedup in Batch Decoding
Non-autoregressive models have high speedup in sentence by sentence translation, 但
this advantage will gradually decrease when we increase the size of decoding batch,
making the advantage of NAT in practical application questioned. We resolve this
concern by measuring the translation latency of NAT and AT models under different
sizes of decoding batches. We conduct experiments on the test set of WMT14 En→De
and show the results in Figure 3.
数字 3
The translation latency of AT, Seq-NAT, and Seq-CMLM measured on the test set of WMT14
En→De. The decoding batch of 400 sentences is not applicable to Seq-CMLM due to the memory
limit.
912
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
桌子 4
The Pearson correlation between loss functions and translation quality. n = k represents the
bag-of-k-grams training objective. CE represents the cross-entropy loss.
n = 3
0.84
Loss function
Correlation
n = 2
0.87
n = 4
0.79
CE
0.56
From Figure 3, as the size of decoding batch increases, both NAT models have
higher translation latency. 尤其, the iterative model Seq-CMLM becomes even much
slower than the autoregressive model when using large batch size. 相反, 这
one-iteration model Seq-NAT still maintains more than 5× speedup during the batch
decoding, demonstrating the efficiency of non-autoregressive generation.
5.5 Correlation with Translation Quality
在这个部分, we conduct experiments to analyze the correlation between loss func-
tions and the translation quality. We are interested in how the cross-entropy loss and
BoN objective correlate with the translation quality. We do not analyze the reinforce-
ment learning based methods because they do not calculate the loss function, 但
directly estimate the gradient of the loss. We use the GLEU score to represent the
translation quality, which is more accurate than BLEU in sentence-level evaluation (吴
等人. 2016). We conduct experiments on the validation set of WMT14 En→De, 哪个
包含 3,000 句子. 第一的, we load the NAT-Base model and calculate the loss of
every sentence in the validation set. Then we use the NAT-Base model to decode the
validation set and calculate the GLEU score of every sentence. 最后, we calculate the
Pearson correlation between the 3,000 GLEU scores and losses.
For the cross-entropy loss, we normalize it by the target sentence length. The BoN
training objective is the L1 distance normalized by 2(T − n + 1). We respectively set n to
2, 3, 和 4 to test different n-gram sizes. 桌子 4 lists the correlation results.
From Table 4, we can see that all three BoN objectives outperform the cross-entropy
loss by large margins, and the n = 2 setting achieves the highest correlation 0.87. 到
find out where the improvements come from, we analyze the effect of sentence length
in the following experiment. We evenly divide the data set into two parts according to
the source length. The first part consists of 1,500 short sentences and the second part
consists of 1,500 long sentences. We respectively measure the Pearson correlation on the
two parts and report the results in Table 5:
From Table 5, we can see that the correlation of the cross-entropy loss drops as
the sentence length increases, where the BoN objective still has a strong correlation on
long sentences. The reason is not difficult to explain. The cross-entropy loss requires the
strict alignment between the translation and reference. As the sentence length grows,
it becomes harder for NAT to align the translation with the reference, which leads to a
decrease of correlation between cross-entropy loss and translation quality. 相比之下,
the BoN objective is robust to unaligned situations, so its correlation with translation
quality stays strong when translating long sentences.
5.6 Effect of Training Strategy
在这个部分, we analyze the effect of training strategy that combines the word-level
loss and the two methods based on reinforcement learning and bag-of-n-grams. 前
913
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
桌子 5
The Pearson correlation between loss functions and translation quality on short sentences and
long sentences.
Cross-Entropy
BoN (n = 2)
全部
0.56
0.87
short
0.68
0.89
长的
0.44
0.86
桌子 6
The training time required for different methods to process 64k tokens. The time is measured on
the training set of WMT14 En-De with a single GeForce RTX 3090 GPU. CE is the cross-entropy
loss. RF is the abbreviation of Reinforce.
方法
时间
CE
1.2s
BoW BoN (n = 2)
1.5s
7.1s
RF-Base
2.2s
RF-Step
16.9s
RF-Top-k
31.3s
Traverse-Ref
63.2s
桌子 7
Validation BLEU scores and training time of different training strategies on WMT14 En-De. CE
represents the cross-entropy loss and TR represents the Traverse-Ref loss. The training time is
measured on 8 GeForce RTX 3090 GPUs.
战略
1. CE—BoN+TR
2. CE+BoN—BoN—TR
3. CE—TR—BoN
4. CE—BoN—TR
BLEU Time
65.6H
24.56
93.1H
24.82
63.3H
24.63
37.5H
24.69
discussing the training strategy, we first give the training speed of each method in
桌子 6. As we can see, Traverse-Ref is the slowest method, which is nearly 10 次
slower than BoN. 所以, when choosing a training strategy, it is necessary to avoid
a large number of calculations of Traverse-Ref.
We consider four training strategies that involve the word-level cross-entropy loss,
the Traverse-Ref loss and the BoN loss. 第一的, we consider the two-stage strategy that
uses the cross-entropy loss for pretraining and finetunes the model by the weighted
summation of the Traverse-Ref and BoN losses. The second strategy follows the joint
training strategy in Shao et al. (2020) to combine the BoN and cross-entropy loss for
pretraining, and then finetunes the model sequentially by BoN and Traverse-Ref. 这
latter two strategies adopt the three-stage strategy that uses the cross-entropy loss for
pretraining and sequentially uses Traverse-Ref and BoN for finetuning. We report the
BLEU scores of the four strategies together with the training time in Table 7.
桌子 7 shows that the second strategy achieves the best performance but suffers
from high training cost. The fourth strategy is more economical; it achieves a slightly
lower BLEU but greatly shortens the training time. Compared with the other two strate-
吉斯, it outperforms them on both BLEU and training time. 所以, we finally adopt
the fourth strategy to combine the word-level training and sequence-level training
方法.
914
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
数字 4
BLEU scores of Reinforce-Top-k with different k on the validation set of WMT14 En-De.
5.7 Effect of Hyperparameters
在这个部分, we analyze the effect of some hyperparameters in our method that will
affect the model performance, including the top-k size k and the reward function in
reinforcement learning, the n-gram size n in bag-of-n-grams training, and the batch size
for finetuning.
Top-k Size. Reinforce-Top-k is proposed to reduce the estimation variance by traversing
the top-k words, which is important in the gradient estimation. 直观地, a larger l will
make the model stronger. When k is 0, Reinforce-Top-k degenerates to Reinforce-Step.
When k equals the vocabulary size |V|, Reinforce-Top-k has the same performance with
Traverse-Ref. 然而, using such a large k will make the training very slow. 所以,
we need to find an appropriate k to balance the performance and training cost. 我们
conduct experiments on the validation set of WMT14 En-De to see the effect of top-k
size k and illustrate our results in Figure 4.
From Figure 4, we can see that the model performance steadily improves as k rises
从 0 到 5. When k rises from 5 到 10, the model performance is also slightly improved.
然而, we can barely see improvements from k = 10 to k = 20, showing that the
appropriate k is between 5 到 10. 此外, we use Traverse-Ref to show the k = |V|
结果, which achieves considerable improvements to Reinforce-Top-k.
Reward Function. The performance of reinforcement learning based methods is influ-
enced by the reward function it uses. Our methods have almost no restriction on the
reward function, where only Traverse-Ref requires the reward function to Reference-
Based. 所以, we choose three widely used Reference-Based rewards BLEU
(Papineni et al. 2002), GLEU (Wu et al. 2016), and ROUGE-2 (林 2004) as candidates.
We use the three rewards to finetune the NAT baseline and report their results in Table 8.
We also directly evaluate the three rewards by the Pearson correlation coefficient with
translation quality. We use the WMT16 DAseg De-En data set for evaluation, 哪个
consists of 560 source sentences, model translations, reference sentences, and human
scores. We obtain the rewards of the model translations and calculate the Pearson
correlation coefficient between rewards and human scores, 如表所示 8. 我们
can see that there is no significant difference in the BLEU performance of these three
915
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
桌子 8
BLEU scores on the validation set of WMT14 En-De when using BLEU, GLEU, or ROUGE-2 as
reward functions, and the Pearson correlation coefficient of these rewards.
蓝线
GLEU
ROUGE-2
Correlation
0.389
0.482
0.483
Reinforce-Base
22.58
22.56
22.39
Reinforce-Step
23.01
23.17
23.14
Traverse-Ref Average
24.12
24.16
24.25
23.24
23.30
23.26
桌子 9
Validation BLEU scores of BoN-L1 with different n on WMT14 En-De and the time required to
process 64k tokens during the training. The time is measured with a single GeForce RTX 3090
GPU.
n = 2
n
蓝线 24.37
7.1s
时间
n = 3
24.29
9.7s
n = 4
24.07
12.3s
rewards. In terms of the correlation, BLEU underperforms ROUGE-2 and GLEU by a
large margin, which is possibly due to instability of BLEU as there are usually little
matches of 3-gram or 4-gram in sentence-level evaluation. We finally use the ROUGE-2
as the reward function because of its overall performance and fast calculation in our
implementation.
N-gram Size. 桌子 3 has shown that the bag-of-n-grams (N= 2) objective outperforms
the bag-of-words objective, but the effect of different n-gram sizes n has not been
分析过的. 所以, we conduct experiments on the validation set of WMT14 En-De
to see the performance of bag-of-n-grams objectives with different choices of n, 和我们
also provide the training speed of BoN-L1 with different n. Results are listed in Table 9.
We can see that n = 2 slightly outperforms other choices of n, which is consistent with
the correlation result in Table 4. 此外, BoN-L1 with n = 2 is much faster than
other choices of n during the training, so we set n = 2 in the main experiment.
Batch Size for Finetuning. In the training of deep neural models, a larger batch size
usually leads to stronger performance, which comes with the cost of greater training
成本. In the sequence-level training scenario, since we only need to finetune the model
for a few steps, we can increase the batch size within a reasonable range, which only
slightly increases the training cost but brings considerable improvements on the model
表现. To show the effect of the batch size, we use different batch sizes during
the BoN finetuning and report the corresponding BLEU scores and total training time
表中 10. We can see that the BLEU score steadily increases as the batch size for
finetuning increases. In terms of training time, even when we use a batch size of 512k,
这是 4 times the size of the pretraining, the training time is only 1.25 times the NAT
基线.
5.8 Effect of Sentence Length
在部分 5.5, we analyze the correlation between loss functions and the translation
quality under different sentence lengths, which shows that sequence-level losses greatly
916
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
桌子 10
Validation BLEU scores on WMT14 En-De and training costs when using different batch size
during the finetuning. NoFT represents the NAT baseline without finetuning. The training time
is measured on 8 GeForce RTX 3090 GPUs.
Batch Size NoFT
19.08
蓝线
24.8H
时间
32k
23.77
25.2H
64k
23.93
25.6H
128k
24.15
26.3H
256k
24.27
27.9H
512k
24.37
31.0H
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
数字 5
Validation BLEU scores of baseline methods and Seq-NAT on WMT14 En-De under different
length buckets.
outperform the word-level loss in terms of correlation when evaluating long sentences.
在这个部分, we calculate the BLEU performance of baseline methods and our model
on different sentence lengths and see whether the better correlation contributes to better
BLEU performance. We conduct the experiment on the validation of WMT14 En→De
and divide the sentence pairs into different length buckets according to the length
of the source sentence. We use Seq-NAT to represent our best performing method,
and calculate the BLEU scores of baseline models and Seq-NAT under different length
buckets. The results are shown in Figure 5.
From Figure 5, we can see that NAT-Base and Seq-NAT have similar perfor-
mance when translating short sentences. 然而, the translation quality of NAT-Base
drops quickly as sentence length increases, where the autoregressive Transformer and
Seq-NAT have stable performance over different sentence lengths, which is in good
agreement with the correlation results. As the sentence length grows, the correlation
between the cross-entropy loss and the translation quality drops, which leads to the
weakness of NAT in translating long sentences. 相反, sequence-level losses
evaluate the translation quality of long sentences with high correlations, so Seq-NAT
has stable performance on long sentences.
917
计算语言学
体积 47, 数字 4
桌子 11
Performance comparison between our method Seq-NAT and existing methods. The speedup is
measured on the WMT14 En-De test set with batch size 1. “—” indicates that the result is not
报道. n is the number of candidates rescored by the autoregressive teacher.
模型
Autoregressive
Transformer
WMT14
WMT16
EN-DE DE-EN EN-RO RO-EN
Speedup
27.42
31.63
34.18
33.72
1.0×
Non-Autoregressive w/o Rescoring
NAT-FT (Gu et al. 2018)
LT (Kaiser et al. 2018)
CTC (Libovick ´y and Helcl 2018)
ENAT (Guo et al. 2019)
NAT-REG (Wang et al. 2019)
NAT-Hints (李等人. 2019)
Reinforce-NAT (Shao et al. 2019)
BoN-Joint+FT (Shao et al. 2020)
imitate-NAT (Wei et al. 2019)
FlowSeq (马等人. 2019)
NART-DCRF (Sun et al. 2019)
ReorderNAT (Ran et al. 2021)
PNAT (Bao et al. 2019)
FCL-NAT (Junliang et al. 2020)
AXE (Ghazvininejad et al. 2020)
EM (Sun and Yang 2020)
Imputer (Saharia et al. 2020)
Seq-NAT (ours)
Non-Autoregressive w/ Rescoring
NAT-FT (n = 10) (Gu et al. 2018)
NAT-FT (n = 100) (Gu et al. 2018)
LT (n = 10) (Kaiser et al. 2018)
ENAT (n = 9) (Guo et al. 2019)
NAT-REG (n = 9) (Wang et al. 2019)
NAT-Hints (n = 9) (李等人. 2019)
imitate-NAT (n = 7) (Wei et al. 2019)
FlowSeq (n = 15) (马等人. 2019)
NART-DCRF (n = 9) (Sun et al. 2019)
PNAT (n = 7) (Bao et al. 2019)
FCL-NAT (n = 9) (Junliang et al. 2020)
EM (n = 9) (Sun and Yang 2020)
Seq-NAT (n = 9, ours)
5.9 Performance Comparision
17.69
19.80
17.68
20.65
20.65
21.11
19.15
20.90
22.44
23.72
23.44
22.79
23.05
21.70
23.53
24.54
25.80
25.54
18.66
19.17
21.0
24.28
24.61
25.20
24.15
25.03
26.07
—
25.75
25.75
26.35
21.47
—
19.80
23.02
24.77
25.24
22.52
24.61
25.67
28.39
27.22
27.28
27.18
25.32
27.90
27.93
28.40
29.91
22.41
23.20
—
26.10
28.90
29.52
27.28
30.48
29.68
27.90
29.50
29.29
30.70
27.29
—
19.93
30.08
—
—
27.09
28.31
28.61
29.73
—
29.30
—
—
30.75
—
32.30
31.69
29.02
29.79
—
34.51
—
—
31.45
31.89
—
—
—
—
33.21
29.06
—
24.71
—
—
—
27.93
29.29
28.90
30.72
—
29.50
—
—
31.54
—
31.70
31.78
30.76
31.44
—
—
—
—
31.81
32.43
—
—
—
—
33.28
15.6×
3.8×
—
25.3×
27.6×
30.8×
10.77×
10.73×
18.6×
—
10.4×
16.1×
7.3×
28.9×
—
16.4×
—
15.6×
7.68×
2.36×
—
12.4×
15.1×
17.8×
9.70×
—
6.14×
3.7×
16.0×
9.14×
9.0×
We use Seq-NAT to represent our best performing method. 表中 11, we com-
pare the performance of Seq-NAT against the autoregressive Transformer and strong
non-iterative NAT baseline models. 桌子 11 shows that Seq-NAT outperforms most
existing NAT systems, and the performance gap between Seq-NAT and the autoregres-
sive teacher is about 2 BLEU on average. Rescoring 9 candidates further improves the
translation quality and narrows the performance gap to about 0.8 BLEU on average. 它
918
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
桌子 12
Three translation cases in the validation set of WMT14 De-En. Source and Target are,
分别, the source sentence and reference sentence. AT is the output of the autoregressive
Transformer. NAT-Base is the output of the NAT baseline. Seq-NAT is the output of our model.
来源
Es gibt Krebsarten, die aggressiv und andere, die indolent sind .
目标
There are aggressive cancers and others that are indolent .
AT
There are cancers that are aggressive and others that are indolent .
NAT-Base
There are cancers cancer aggressive aggressive others are indindent .
Seq-NAT
There are cancers that are aggressive and others that indolent .
来源
目标
AT
Wir wissen ohne den Schatten eines Zweifels, dass wir ein echtes
neues Teilchen haben, und dass es dem vom Standardmodell
vorausgesagten Higgs-Boson stark ˜Ad’hnelt .
We know without a shadow of a doubt that it is a new authentic
particle, and greatly resembles the Higgs boson predicted by the
Standard Model .
We know without the shadow of a doubt that we have a real new
particle, and that it is very similar to the Higgs Boson predicted by
the standard model .
NAT-Base
We know without without shadow shadow of doubt doubt that we
have a new particle le that it is very similar similar to HiggsgsBoson
predicted by the standard model .
Seq-NAT
We know without the shadow of a doubt that we have a real new
particle and that it is very similar to the Higgs-Boson predicted by
the standard model .
来源
目标
AT
und noch tragischer ist, dass es Oxford war – eine Universit ˜Ad’t, die
nicht nur 14 Tory-Premierminister hervorbrachte, sondern sich bis
heute hinter einem unverdienten Ruf von Gleichberechtigung und
Gedankenfreiheit versteckt .
even more tragic is that it was Oxford, which not only produced 14
Tory prime ministers, 但, to this day, hides behind an ill-deserved
reputation for equality and freedom of thought .
And more tragically, it was Oxford – a university that not only
produced 14 Tory prime ministers but hides to this day behind an
undeserved reputation of equality and freedom of thought .
NAT-Base
More more tragic tragic it was Oxford Oxford Oxford university
university university not only 14 14 14 Torprime prime ministers but
but hihihihiday behind an unved call for equality and freedom of
想法 .
Seq-NAT
More is even more tragic that it was Oxford – a university that not
only produced 14 Tory prime ministers, but continues continues far
hidden behind an unved call for equality and freedom of thought .
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
919
计算语言学
体积 47, 数字 4
is also worth noting that Seq-NAT does not affect the translation speed, which has the
same speedup 15.6× as NAT-Base. After rescoring 9 candidates, Seq-NAT still maintains
9.0× speedup.
5.10 Case Study
表中 12, we present three translation cases from the validation set of WMT14
De-En to analyze how sequence-level training improves the translation quality of NAT.
We can see from the three cases that the NAT baseline suffers from over-translation
and under-translation errors especially when translating long sentences. The output of
NAT-Base contains many repeated translations like “aggressive,” “shadow,” and “14.”
此外, the translation is incomplete since much information is missing. 和我们一样
mentioned before, this is due to the limitation of the word-level cross-entropy loss we
使用, which evaluates the generation quality of each position independently and does
not model the target-side sequential dependency, making NAT only focus on local
correctness and ignore the overall translation quality.
When we look at the translation results of Seq-NAT, we can see that the errors of
over-translation and under-translation are significantly reduced. Although there are still
a few repeated translations when translating long sentences, the translation results are
basically accurate and comparable to the autoregressive Transformer. 与相比
the NAT baseline, Seq-NAT focuses more on the overall accuracy after the sequence-
level training, which greatly improves the translation quality.
6. 结论
Non-autoregressive translation achieves significant decoding speedup through gener-
ating target words independently and simultaneously. 然而, the word-level cross-
entropy loss cannot evaluate the output of NAT properly. 因此, NAT has a
relatively low translation quality and tends to generate translations with over-
translation and under-translation errors. 在本文中, we propose to train NAT with
sequence-level training objectives. 第一的, we propose to train NAT to optimize the
sequence-level evaluation metric based on novel reinforcement algorithms customized
for NAT. Then we introduce a novel bag-of-n-grams objective for NAT, which is differ-
entiable and can be calculated efficiently. 最后, we use a three-stage training strat-
egy to combine the strengths of the two training methods and the word-level loss.
Experimental results show that our method achieves remarkable performance on all
translation tasks.
参考
Akoury, Nader, Kalpesh Krishna, and Mohit
伊耶尔. 2019. Syntactically supervised
transformers for faster neural machine
翻译. In Proceedings of the 57th
Annual Meeting of the Association for
计算语言学, pages 1269–1281,
Florence. https://doi.org/10.18653
/v1/P19-1122
Bahdanau, Dzmitry, Philemon Brakel, Kelvin
徐, Anirudh Goyal, Ryan Lowe, Joelle
Pineau, Aaron C. 考维尔, and Yoshua
本吉奥. 2017. An actor-critic algorithm for
sequence prediction. In 5th International
Conference on Learning Representations, ICLR
2017, Toulon, 法国, 四月 24-26, 2017,
Conference Track Proceedings.
Bahdanau, Dzmitry, Kyunghyun Cho, 和
Yoshua Bengio. 2015. Neural machine
translation by jointly learning to align and
translate. In 3rd International Conference on
Learning Representations, ICLR 2015,
圣地亚哥, CA, 美国, 可能 7-9, 2015,
Conference Track Proceedings.
Bao, 于, Hao Zhou, Jiangtao Feng, Mingxuan
王, Shujian Huang, Jiajun Chen, and LI
920
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
Lei. 2019. Non-autoregressive transformer
by position learning. arXiv 预印本
arXiv:1911.10677.
本吉奥, Samy, Oriol Vinyals, Navdeep Jaitly,
and Noam Shazeer. 2015. Scheduled
sampling for sequence prediction with
recurrent neural networks. In Proceedings
of the 28th International Conference on
Neural Information Processing Systems –
体积 1, NIPS’15, pages 1171–1179,
剑桥, 嘛.
体积 70, ICML’17, pages 1243–1252,
JMLR.org.
Ghazvininejad, Marjan, Vladimir Karpukhin,
Luke Zettlemoyer, and Omer Levy.
2020. Aligned cross entropy for
non-autoregressive machine translation. 在
Proceedings of the 37th International
Conference on Machine Learning, ICML 2020,
13-18 七月 2020, Virtual Event, 体积 119,
of Proceedings of Machine Learning Research,
pages 3515–3523.
给, Kyunghyun, Bart van Merri¨enboer,
Ghazvininejad, Marjan, Omer Levy, Yinhan
Caglar Gulcehre, Dzmitry Bahdanau, Fethi
Bougares, Holger Schwenk, and Yoshua
本吉奥. 2014. Learning phrase
representations using RNN
encoder–decoder for statistical machine
翻译. 在诉讼程序中 2014
Conference on Empirical Methods in Natural
语言处理 (EMNLP),
pages 1724–1734, Doha. https://土井
.org/10.3115/v1/D14-1179
刘, and Luke Zettlemoyer. 2019.
Mask-predict: Parallel decoding of
conditional masked language models. 在
诉讼程序 2019 会议
自然语言的经验方法
Processing and the 9th International Joint
Conference on Natural Language Processing
(EMNLP-IJCNLP), pages 6112–6121.
https://doi.org/10.18653/v1/D19
-1633
Devlin, 雅各布, Ming-Wei Chang, Kenton Lee,
Goyal, Anirudh, Alex Lamb, Ying Zhang,
and Kristina Toutanova. 2019. BERT:
Pre-training of deep bidirectional
transformers for language understanding.
在诉讼程序中 2019 Conference of the
North American Chapter of the Association
for Computational Linguistics: 人类
语言技术, 体积 1 (Long and
Short Papers), pages 4171–4186,
明尼阿波利斯, 明尼苏达州.
Edunov, Sergey, Myle Ott, Michael Auli,
David Grangier, and Marc’Aurelio
Ranzato. 2018. Classical structured
prediction losses for sequence to sequence
学习. 在诉讼程序中 2018
Conference of the North American Chapter of
the Association for Computational Linguistics:
人类语言技术, 体积 1
(Long Papers), pages 355–364,
New Orleans, 这. https://doi.org/10
.18653/v1/N18-1033
冯, 哪个, Shuhao Gu, Dengji Guo,
Zhengxin Yang, and Chenze Shao. 2021.
Guiding teacher forcing with seer forcing
for neural machine translation. 在
Proceedings of the 59th Annual Meeting of the
Association for Computational Linguistics and
the 11th International Joint Conference on
自然语言处理 (体积 1: 长的
文件), pages 2862–2872, 在线的.
https://doi.org/10.18653/v1/2021
.acl-long.223
Gehring, Jonas, Michael Auli, 大卫
Grangier, Denis Yarats, and Yann N.
Dauphin. 2017. Convolutional sequence to
sequence learning. In Proceedings of the 34th
International Conference on Machine Learning –
Saizheng Zhang, Aaron C. 考维尔, 和
Yoshua Bengio. 2016. Professor forcing: A
new algorithm for training recurrent
网络. In Advances in Neural Information
Processing Systems 29: Annual Conference on
Neural Information Processing Systems 2016,
十二月 5-10, 2016, 巴塞罗那, 西班牙,
pages 4601–4609.
Gu, Jiatao, James Bradbury, Caiming Xiong,
Victor O. K. 李, and Richard Socher. 2018.
Non-autoregressive neural machine
翻译. In 6th International Conference
on Learning Representations, ICLR 2018,
Vancouver, BC, 加拿大, 四月 30 – 可能 3,
2018, Conference Track Proceedings.
Gu, Jiatao, Changhan Wang, and Junbo
赵. 2019. Levenshtein transformer. 在
神经信息处理的进展
系统 32: Annual Conference on Neural
Information Processing Systems 2019,
神经信息处理系统 2019, 十二月 8-14, 2019,
Vancouver, BC, 加拿大, pages 11179–11189.
Guo, Junliang, Xu Tan, Di He, Tao Qin,
Linli Xu, and Tie-Yan Liu. 2019.
Non-autoregressive neural machine
translation with enhanced decoder input.
Proceedings of the AAAI Conference on
人工智能, 33(01):3723–3730.
https://doi.org/10.1609/aaai.v33i01
.33013723
他, 从, Yingce Xia, Tao Qin, Liwei Wang,
Nenghai Yu, Tieyan Liu, and Wei-Ying Ma.
2016. Dual learning for machine
翻译. In Advances in Neural
Information Processing Systems 29,
pages 820–828.
921
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
欣顿, 杰弗里, Oriol Vinyals, and Jeff
院长. 2015. Distilling the knowledge in a
neural network. arXiv 预印本
arXiv:1503.02531.
Joachims, Thorsten. 1998. Text categorization
with support vector machines: 学习
with many relevant features. In Proceedings
of the 10th European Conference on Machine
学习, ECML’98, pages 137–142.
https://doi.org/10.1007/BFb0026683
Junliang, Guo, Xu Tan, Linli Xu, Tao Qin,
Enhong Chen, and Tie-Yan Liu. 2020.
Fine-tuning by curriculum learning for
non-autoregressive neural machine
翻译. In Proceedings of the AAAI
Conference on Artificial Intelligence,
34:7839–7846. https://doi.org/10
.1609/aaai.v34i05.6289
Kaiser, Lukasz, Samy Bengio, Aurko Roy,
Ashish Vaswani, Niki Parmar, Jakob
Uszkoreit, and Noam Shazeer. 2018. Fast
decoding in sequence models using
discrete latent variables. 在诉讼程序中
the 35th International Conference on Machine
学习, 体积 80 of Proceedings of
Machine Learning Research,
pages 2390–2399, 斯德哥尔摩.
Kasai, Jungo, James Cross, Marjan
Ghazvininejad, and Jiatao Gu. 2020.
Non-autoregressive machine translation
with disentangled context transformer.
In Proceedings of the 37th International
Conference on Machine Learning, ICML 2020,
13-18 七月 2020, Virtual Event, 体积 119
of Proceedings of Machine Learning Research,
pages 5144–5155.
Kim, Yoon and Alexander M. 匆忙. 2016.
Sequence-level knowledge distillation.
在诉讼程序中 2016 会议
自然语言的经验方法
加工, pages 1317–1327, 奥斯汀, TX.
https://doi.org/10.18653/v1/D16-1139
Kingma, Diederik P. and Jimmy Ba. 2014.
亚当: A method for stochastic
优化. arXiv 预印本
arXiv:1412.6980.
李, Jason, Elman Mansimov, 和
Kyunghyun Cho. 2018. Deterministic
non-autoregressive neural sequence
modeling by iterative refinement. 在
诉讼程序 2018 会议
自然语言的经验方法
加工, pages 1173–1182, 布鲁塞尔.
https://doi.org/10.18653/v1/D18-1149
李, Bofang, Zhe Zhao, Tao Liu, Puwei Wang,
and Xiaoyong Du. 2016. Weighted neural
bag-of-n-grams model: New baselines for
text classification. 在诉讼程序中
科林 2016, the 26th International
922
Conference on Computational Linguistics:
技术论文, pages 1591–1600, 大阪.
李, Zhuohan, Zi Lin, Di He, Fei Tian, Tao Qin,
Liwei Wang, and Tie-Yan Liu. 2019.
Hint-based training for non-autoregressive
machine translation. 在诉讼程序中
2019 实证方法会议
Natural Language Processing and the 9th
International Joint Conference on Natural
语言处理 (EMNLP-IJCNLP),
pages 5708–5713, 香港.
https://doi.org/10.18653/v1/D19
-1573
Libovick ´y, Jindˇrich and Jindˇrich Helcl. 2018.
End-to-end non-autoregressive neural
machine translation with connectionist
temporal classification. 在诉讼程序中
2018 实证方法会议
自然语言处理,
pages 3016–3021, 布鲁塞尔. https://土井
.org/10.18653/v1/D18-1336
林, Chin Yew. 2004. ROUGE: A package for
automatic evaluation of summaries. In Text
Summarization Branches Out, pages 74–81,
巴塞罗那.
Ma, Shuming, Xu Sun, Yizhong Wang, 和
Junyang Lin. 2018. Bag-of-words as target
for neural machine translation. 在
Proceedings of the 56th Annual Meeting of the
计算语言学协会
(体积 2: Short Papers), pages 332–338,
墨尔本. https://doi.org/10
.18653/v1/P18-2053
Ma, Xuezhe, Chunting Zhou, Xian Li,
Graham Neubig, and Eduard Hovy. 2019.
FlowSeq: Non-autoregressive conditional
sequence generation with generative flow.
在诉讼程序中 2019 会议
自然语言的经验方法
Processing and the 9th International Joint
Conference on Natural Language Processing
(EMNLP-IJCNLP), pages 4282–4292,
香港. https://doi.org/10.18653
/v1/D19-1437
的, Andrew Y., Daishi Harada, and Stuart J.
拉塞尔. 1999. Policy invariance under
reward transformations: Theory and
application to reward shaping. 在
Proceedings of the Sixteenth International
Conference on Machine Learning, ICML ’99,
pages 278–287.
Norouzi, Mohammad, Samy Bengio, Zhifeng
陈, Navdeep Jaitly, Mike Schuster,
Yonghui Wu, and Dale Schuurmans. 2016.
Reward augmented maximum likelihood
for neural structured prediction. 在
Proceedings of the 30th International
Conference on Neural Information Processing
系统, NIPS’16, pages 1731–1739.
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
Pang, Bo, Lillian Lee, and Shivakumar
Vaithyanathan. 2002. Thumbs up?
Sentiment classification using machine
learning techniques. 在诉讼程序中
2002 实证方法会议
自然语言处理 (EMNLP 2002),
pages 79–86. https://doi.org/10.3115
/1118693.1118704
Papineni, Kishore, Salim Roukos, Todd
Ward, and Wei-Jing Zhu. 2002. 蓝线: A
method for automatic evaluation of
machine translation. 在诉讼程序中
40th Annual Meeting of the Association for
计算语言学, pages 311–318.
费城, PA. https://doi.org/10
.3115/1073083.1073135
Ran, Qiu, Yankai Lin, Peng Li, and Jie Zhou.
2020. Learning to recover from
multi-modality errors for
non-autoregressive neural machine
翻译. In Proceedings of the 58th
Annual Meeting of the Association for
计算语言学, pages 3059–3069,
在线的. https://doi.org/10.18653/v1
/2020.acl-main.277
Ran, Qiu, Yankai Lin, Peng Li, and Jie Zhou.
2021. Guiding non-autoregressive neural
machine translation decoding with
reordering information. In Thirty-Fifth
AAAI Conference on Artificial Intelligence,
AAAI 2021, Virtual Event, 二月 2-9,
2021, pages 13727–13735.
Ranzato, Marc’Aurelio, Sumit Chopra,
Michael Auli, and Wojciech Zaremba. 2016.
Sequence level training with recurrent
神经网络. In 4th International
Conference on Learning Representations, ICLR
2016, San Juan, Puerto Rico, 可能 2-4, 2016,
Conference Track Proceedings.
Saharia, Chitwan, William Chan, Saurabh
Saxena, and Mohammad Norouzi. 2020.
Non-autoregressive machine translation
with latent alignments. 在诉讼程序中
2020 实证方法会议
自然语言处理 (EMNLP),
pages 1098–1108. 在线的. https://土井
.org/10.18653/v1/2020.emnlp-main.83
Sennrich, Rico, Barry Haddow, 和
Alexandra Birch. 2016. Neural machine
translation of rare words with subword
units. In Proceedings of the 54th Annual
Meeting of the Association for Computational
语言学 (体积 1: Long Papers),
pages 1715–1725. 柏林. https://doi.org
/10.18653/v1/P16-1162
Shan, Yong, Yang Feng, and Chenze Shao.
2021. Modeling coverage for
non-autoregressive neural machine
翻译. arXiv 预印本 arXiv:2104.11897.
https://doi.org/10.18653/v1/D18
-1510
Shao, Chenze, Xilin Chen, and Yang Feng.
2018. Greedy search with probabilistic
n-gram matching for neural machine
翻译. 在诉讼程序中 2018
Conference on Empirical Methods in Natural
语言处理, pages 4778–4784,
布鲁塞尔. https://doi.org/10.18653
/v1/D18-1510
Shao, Chenze, Yang Feng, Jinchao Zhang,
Fandong Meng, Xilin Chen, and Jie Zhou.
2019. Retrieving sequential information for
non-autoregressive neural machine
翻译. In Proceedings of the 57th
Annual Meeting of the Association for
计算语言学, pages 3013–3024,
Florence. https://doi.org/10.18653
/v1/P19-1288
Shao, Chenze, Jinchao Zhang, Yang Feng,
Fandong Meng, and Jie Zhou. 2020.
Minimizing the bag-of-ngrams difference
for non-autoregressive neural machine
翻译. In The Thirty-Fourth AAAI
Conference on Artificial Intelligence, AAAI
2020, 纽约, 纽约, 美国, 二月 7-12,
2020, pages 198–205. https://doi.org
/10.1609/aaai.v34i01.5351
沉, Shiqi, Yong Cheng, Zhongjun He, Wei
他, Hua Wu, Maosong Sun, and Yang Liu.
2016. Minimum risk training for neural
machine translation. 在诉讼程序中
54th Annual Meeting of the Association for
计算语言学 (体积 1: 长的
文件), pages 1683–1692, 柏林. https://
doi.org/10.18653/v1/P16-1159
Shu, 拉斐尔, Jason Lee, Hideki Nakayama,
and Kyunghyun Cho. 2020.
Latent-variable non-autoregressive neural
machine translation with deterministic
inference using a delta posterior. In The
Thirty-Fourth AAAI Conference on Artificial
智力, AAAI 2020, 纽约, 纽约,
美国, 二月 7-12, 2020, pages 8846–8853.
https://doi.org/10.1609/aaai.v34i05
Sun, Zhiqing, Zhuohan Li, Haoqing Wang,
Di He, Zi Lin, and Zhihong Deng. 2019,
Fast structured decoding for sequence
型号. In Advances in Neural Information
Processing Systems 32, pages 3016–3026.
Sun, Zhiqing and Yiming Yang. 2020. 一个
EM approach to non-autoregressive
conditional sequence generation. 在
Proceedings of the 37th International
Conference on Machine Learning, 体积 119
of Proceedings of Machine Learning Research,
pages 9249–9258.
吸勺, 伊利亚, Oriol Vinyals, and Quoc V.
Le. 2014. Sequence to sequence learning
923
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
计算语言学
体积 47, 数字 4
with neural networks. 在诉讼程序中
27th International Conference on Neural
Information Processing Systems – 体积 2,
NIPS’14, pages 3104–3112.
Sutton, Richard S., David McAllester,
Satinder Singh, and Yishay Mansour. 1999.
Policy gradient methods for reinforcement
learning with function approximation.
In Proceedings of the 12th International
Conference on Neural Information
Processing Systems, NIPS’99,
pages 1057–1063.
Sutton, Richard Stuart. 1984. Temporal Credit
Assignment in Reinforcement Learning. 博士.
论文. AAI8410337.
Tu, Lifu, Richard Yuanzhe Pang, 山姆
Wiseman, and Kevin Gimpel. 2020.
ENGINE: Energy-based inference
networks for non-autoregressive machine
翻译. In Proceedings of the 58th
Annual Meeting of the Association for
计算语言学, pages 2819–2826,
在线的.
Vaswani, Ashish, Noam Shazeer, Niki
Parmar, Jakob Uszkoreit, Llion Jones,
Aidan N. Gomez, Łukasz Kaiser, 和
Illia Polosukhin. 2017. Attention is
all you need. In Proceedings of the 31st
International Conference on Neural
Information Processing Systems, NIPS’17,
pages 6000–6010.
Venkatraman, Arun, Martial Hebert, 和
J. Andrew Bagnell. 2015. Improving
multi-step prediction of learned time series
型号. In Proceedings of the Twenty-Ninth
AAAI Conference on Artificial Intelligence,
AAAI’15, pages 3024–3030.
王, Chunqi, Ji Zhang, and Haiqing Chen.
2018. Semi-autoregressive neural machine
翻译. 在诉讼程序中 2018
Conference on Empirical Methods in Natural
语言处理, pages 479–488,
布鲁塞尔. https://doi.org/10.18653
/v1/D18-1044
王, Yiren, Fei Tian, Di He, Tao Qin,
ChengXiang Zhai, and Tie-Yan Liu. 2019.
Non-autoregressive machine translation
with auxiliary regularization. In The
Thirty-Third AAAI Conference on Artificial
智力, AAAI 2019, 檀香山, Hawaii,
美国, 一月 27 – 二月 1, 2019,
pages 5377–5384. https://doi.org/10
.1609/aaai.v33i01.33015377
Weaver, Lex and Nigel Tao. 2001. 这
optimal reward baseline for
gradient-based reinforcement learning. 在
Proceedings of the 17th Conference in
Uncertainty in Artificial Intelligence,
UAI ’01, pages 538–545.
924
Wei, Bingzhen, Mingxuan Wang, Hao Zhou,
Junyang Lin, and Xu Sun. 2019. Imitation
learning for non-autoregressive neural
machine translation. 在诉讼程序中
57th Annual Meeting of the Association for
计算语言学, pages 1304–1312,
Florence. https://doi.org/10.18653
/v1/P19-1125
威廉姆斯, Ronald J. 1992. Simple statistical
gradient-following algorithms for
connectionist reinforcement learning.
Machine Learning, 8(3–4):229–256.
https://doi.org/10.1007/BF00992696
威廉姆斯, Ronald J. and David Zipser. 1989.
A learning algorithm for continually
running fully recurrent neural networks.
神经计算, 1(2):270–280.
https://doi.org/10.1162/neco.1989
.1.2.270
吴, Lijun, Fei Tian, Tao Qin, Jianhuang Lai,
and Tie-Yan Liu. 2018A. A study of
reinforcement learning for neural machine
翻译. 在诉讼程序中 2018
Conference on Empirical Methods in Natural
语言处理, pages 3612–3621,
布鲁塞尔. https://doi.org/10.18653
/v1/D18-1397
吴, Lijun, Yingce Xia, Fei Tian, Li Zhao, Tao
Qin, Jianhuang Lai, and Tie-Yan Liu.
2018乙. Adversarial neural machine
翻译. In Proceedings of the 10th Asian
Conference on Machine Learning, 体积 95
of Proceedings of Machine Learning Research,
pages 534–549.
吴, Yonghui, Mike Schuster, Zhifeng Chen,
Quoc V. Le, Mohammad Norouzi,
Wolfgang Macherey, Maxim Krikun, Yuan
曹, Qin Gao, Klaus Macherey, 和别的.
2016. Google’s neural machine translation
系统: Bridging the gap between human
and machine translation. arXiv 预印本
arXiv:1609.08144.
哪个, Zhen, Wei Chen, Feng Wang, 和
Bo Xu. 2018. Improving neural machine
translation with conditional sequence
generative adversarial nets. In Proceedings
的 2018 Conference of the North
American Chapter of the Association for
计算语言学: 人类
语言技术, 体积 1 (长的
文件), pages 1346–1355, New Orleans,
这. https://doi.org/10.18653/v1/N18
-1122
于, Lantao, Weinan Zhang, Jun Wang, 和
Yong Yu. 2017. SeqGAN: Sequence
generative adversarial nets with policy
gradient. In Proceedings of the Thirty-First
AAAI Conference on Artificial Intelligence,
AAAI’17, pages 2852–2858.
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
Shao et al.
Sequence-Level Training
张, Wen, Yang Feng, Fandong Meng,
Di You, and Qun Liu. 2019. Bridging the
gap between training and inference for
neural machine translation. In Proceedings
of the 57th Annual Meeting of the Association
for Computational Linguistics,
pages 4334–4343, Florence. https://土井
.org/10.18653/v1/P19-1426
周, Chunting, Jiatao Gu, and Graham
Neubig. 2020. Understanding knowledge
distillation in non-autoregressive machine
翻译. In 8th International Conference
on Learning Representations, ICLR 2020,
亚的斯亚贝巴(埃塞俄比亚首都, 埃塞俄比亚, 四月 26-30, 2020.
周, Jiawei and Phillip Keung. 2020.
Improving non-autoregressive neural
machine translation with monolingual
数据. In Proceedings of the 58th Annual
Meeting of the Association for Computational
语言学, pages 1893–1898, 在线的.
https://doi.org/10.18653/v1/2020
.acl-main.171
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
925
我
D
哦
w
n
哦
A
d
e
d
F
r
哦
米
H
t
t
p
:
/
/
d
我
r
e
C
t
.
米
我
t
.
e
d
你
/
C
哦
我
我
/
我
A
r
t
我
C
e
–
p
d
F
/
/
/
/
4
7
4
8
9
1
1
9
7
9
3
9
3
/
C
哦
我
我
_
A
_
0
0
4
2
1
p
d
.
F
乙
y
G
你
e
s
t
t
哦
n
0
7
S
e
p
e
米
乙
e
r
2
0
2
3
926
下载pdf