Learn Before
Concept
LAMBADA
It learns a label-conditioned generator by finetuning GPT-2 on the training data, using this to generate candidate examples per class. A classifier trained on the original training set is then used to select top k candidate examples that confidently belong to the respective class for augmentation.
0
1
Updated 2022-05-20
Tags
Data Science
Related
BackTranslation
Improving Neural Machine Translation Models with Monolingual Data
Generative Data Augmentation for Commonsense Reasoning
G-DAUG
Keep Calm and Switch On! Preserving Sentiment and Fluency in Semantic Text Exchange
SEMANTIC TEXT EXCHANGE (STE)
LAMBADA
Not Enough Data? Deep Learning to the Rescue!
Trade-off of Model-Based Data Augmentation Techniques in NLP