Heuristics-Based Data Filtering for Fine-Tuning
One common strategy for data selection in fine-tuning is the application of heuristics. This involves using predefined rules or guidelines to systematically filter a dataset, aiming to remove low-quality or redundant samples to improve the overall training data.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Small Model-Based Data Selection
Heuristics-Based Data Filtering for Fine-Tuning
Prioritizing Influential Data for Fine-Tuning
A development team fine-tunes a large language model on a massive, newly-generated dataset of 1 million instruction-response pairs. After training, they find the model's performance is poor, often generating repetitive, nonsensical, or factually incorrect answers. Which of the following is the most likely root cause of this issue and the best initial strategy to address it?
Evaluating a Data Filtering Strategy
A team is preparing a large, synthetically-generated dataset for fine-tuning a language model. They suspect the dataset has several quality issues. Match each potential data quality problem with the primary goal of a filtering method designed to address it.
Your company is rolling out an instruction-tuned L...
You lead an LLM enablement team building an instru...
You’re leading an LLM platform team building an in...
Your company is building an internal IT helpdesk a...
Deciding Whether (and How) to Use Weak-Model Synthetic Data for Instruction Fine-Tuning
Diagnosing and Fixing a Synthetic Instruction-Tuning Data Flywheel That Degrades Model Behavior
Designing a Synthetic Instruction Fine-Tuning Pipeline Under Budget and Quality Constraints
Stabilizing an Instruction-Tuned Support Assistant When Synthetic Data Conflicts with Human Policy
Selecting and Filtering Self-Generated Instruction Data When Bootstrapping a Strong Model from a Weak Supervisor
Choosing a Weak-Model + Self-Instruct Data Strategy for Instruction Fine-Tuning Without Regressions
Efficiency Benefits of Data Selection in Fine-Tuning
Alpagasus Data Selection System
Learn After
A team is preparing a dataset to fine-tune a large language model for a customer service chatbot. The raw data, collected from public online forums, contains many instances of toxic language, messages shorter than five words, and conversations that are not in English. The primary goal is to improve the model's ability to provide helpful, safe, and coherent responses in English. Which of the following filtering rules would be the most effective first step to improve the quality of this specific dataset for the intended task?
Unintended Consequences of Data Filtering
Designing Filtering Rules for a Specialized AI