Concept

Increasing Idiom Context

Using more context and pretraining improves (lowers) perplexity. Training on joint split consistently improves random models, and upsampling idiom-train data yields bad results due to overfitting.

0

1

Updated 2023-02-17

Tags

Data Science