Sentence Reordering as an Input Corruption Method
Sentence reordering is a data corruption technique applied to multi-sentence inputs where the original order of sentences within a document is randomly shuffled. The model is then trained on this permuted sequence with the objective of learning to restore the sentences to their correct, logical order.
0
1
References
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Tags
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Sentence Reordering as an Input Corruption Method
Document Rotation as an Input Corruption Method
A research team is training a model on multi-paragraph documents. Their primary goal is to ensure the model learns the logical flow and coherence between sentences, not just the relationships between words within a single sentence. Which of the following input corruption strategies is specifically designed to target this higher-level, inter-sentence understanding?
Rationale for Sentence-Level Corruption
A language model is being trained using a denoising objective, where it learns to reconstruct original text from a corrupted version. Match each type of input corruption with the primary linguistic feature it forces the model to learn.
Sentence Reordering as an Input Corruption Method
A pre-training process is applied to the following two-sentence input: 'The team celebrated their victory. They had trained hard for months.' The input is transformed into: 'The team [MASK] hard for months.' The model is then tasked with reconstructing the original, complete text from this transformed input. Which specific data corruption technique, designed for handling sequences of text, does this process exemplify?
Choosing a Pre-training Strategy
A model is pre-trained using corruption techniques that operate on the structure of multi-sentence documents. Match each technique with its correct description.
Learn After
Example of a Sentence Reordering Task
A language model developer observes that their model generates fluent, grammatically correct individual sentences. However, when generating multi-sentence paragraphs, the model's output often lacks logical coherence, with the order of events or ideas appearing random and jumbled. Which of the following training objectives would be most effective at directly addressing this specific weakness?
A language model is being trained using a method where it receives a modified version of a text and must learn to reconstruct the original. Consider the following example:
Original Text: 'The team gathered in the lab to review the experiment's results. Initial readings showed a significant breakthrough. They decided to run a second verification test immediately.'
Training Input: 'Initial readings showed a significant breakthrough. They decided to run a second verification test immediately. The team gathered in the lab to review the experiment's results.'
Based on this example, what is the primary skill this specific training method is designed to teach the model?
Evaluating a Training Strategy
Example of Sentence Reordering in Denoising Autoencoding