Learn Before
Concept
Increasing Idiom Context
Using more context and pretraining improves (lowers) perplexity. Training on joint split consistently improves random models, and upsampling idiom-train data yields bad results due to overfitting.
0
1
Updated 2023-02-17
Tags
Data Science