Learn Before
Concept

Literal Translation Errors

All models produce fewer errors when trained on joint split compared to zero split. Pretraining and upsampling idiom-train data helps all models. Masking increases errors on the joint split, and decoder-side word replacements yield a similar behavior in terms of LitTer. Adding word replacements in the encoder reduces LitTER.

0

1

Updated 2023-02-17

Tags

Data Science