Learn Before
Concept
Literal Translation Errors
All models produce fewer errors when trained on joint split compared to zero split. Pretraining and upsampling idiom-train data helps all models. Masking increases errors on the joint split, and decoder-side word replacements yield a similar behavior in terms of LitTer. Adding word replacements in the encoder reduces LitTER.
0
1
Updated 2023-02-17
Tags
Data Science