
Quality vs. Quantity: The 100-Example Rule
Master the fundamental law of modern fine-tuning. Learn why a small 'Golden Dataset' of 100 examples outperforms a messy million-row log.
Quality vs. Quantity: The 100-Example Rule
There is an old saying in Computer Science: "Garbage In, Garbage Out" (GIGO). In the world of fine-tuning, this isn't just a catchy phrase—it is a law of physics.
In the early days of NLP (2018-2020), engineering teams were obsessed with Quantity. They felt that to train a model, you needed millions of text samples. But something changed when we moved to Foundation Models (LLMs). These models are already "pre-smart." They already know language.
Because of this, modern fine-tuning is about Quality. In this lesson, we will explore the "100-Example Rule" and why your focus should be on curation, not collection.
The Diminishing Returns of Data
If you increase your dataset from 10 examples to 100 examples, you will see a massive leap in performance. If you increase it from 1,000 to 10,000, the improvement becomes much smaller.
The "SFT S-Curve"
The relationship between data volume and performance in Supervised Fine-Tuning is an S-Curve.
- Phase 1 (The Initial Boost): 0 to 100 examples. The model "gets the idea." It understands the format and the tone.
- Phase 2 (The Plateau): 100 to 1,000 examples. The model refines its behavior and starts handling edge cases.
- Phase 3 (The Overfitting Risk): > 10,000 examples. Unless your task is extremely diverse, the model starts to "memorize" the training data rather than "learning" the skill.
graph LR
A["Data Volume"] --> B["Low Accuracy (0-10)"]
B --> C["The 'Golden' Breakout (100)"]
C --> D["Diminishing Returns (1,000)"]
D --> E["Saturation & Overfitting (10k+)"]
style C fill:#f9f,stroke:#333,stroke-width:4px
Why 100 Examples is the "Magic Number"
For most industrial fine-tuning tasks (Style, Formatting, or Simple Classification), 100 "Golden Examples" are often sufficient for a production-ready model.
What is a "Golden Example"?
A Golden Example is a (Prompt, Response) pair that has been:
- Human-Curated: Written or verified by a domain expert.
- Statistically Representative: Covers a specific common scenario in your application.
- Syntactically Perfect: Zero spelling errors, perfect JSON formatting, and correct brand tone.
The Math of Developer Time
- Writing 100 Golden Examples takes a single engineer about 4 - 8 hours.
- Writing 1,000 Golden Examples takes a team one week.
- Cleaning 1,000,000 rows of noisy chat logs takes months and usually results in a worse model.
Case Study: The "LIMA" Insight
In 2023, Meta researchers published a paper called "LIMA: Less Is More for Alignment." They showed that fine-tuning a model on only 1,000 extremely high-quality examples allowed it to compete with (and often beat) models trained on hundreds of thousands of instructions. The conclusion? A model’s knowledge comes almost entirely from pretraining; fine-tuning is purely for learning the format to output that knowledge.
Implementation: The "Quality Audit" Script
Before you start training, you should run a "Self-Audit" on your data. Here is a Python pattern to detect "Low Quality" examples in your training set.
def quality_audit(dataset):
"""
Checks for common 'Quality Killers' in a fine-tuning dataset.
"""
audit_results = []
for idx, sample in enumerate(dataset):
errors = []
user_input = sample['messages'][0]['content']
assistant_output = sample['messages'][1]['content']
# 1. Length Check (Is it too short to be useful?)
if len(assistant_output) < 20:
errors.append("Output too short")
# 2. Pattern Check (Does it contain 'As an AI language model'?)
# We don't want to bake in the base model's default refusals
if "as an AI language model" in assistant_output.lower():
errors.append("Default refusal found")
# 3. Syntax Check (Is the JSON valid?)
if "{" in assistant_output:
try:
import json
json.loads(assistant_output)
except:
errors.append("Invalid JSON syntax")
if errors:
audit_results.append({"index": idx, "errors": errors})
return audit_results
# Before training, fix or delete any samples identified in the audit!
The "Diversity" Requirement
While we want high quality, we also need diversity. If your 100 examples only cover "Login problems," your model will be terrible at "Password reset" problems. The Pro Rule: Use 100 examples total, but ensure those 100 represent 10-15 different intent types (e.g., 10 for billing, 10 for shipping, 10 for returns).
Summary and Key Takeaways
- Modern Fine-Tuning is defined by "The Quality over Quantity" paradigm.
- 100 Examples is the threshold for noticeable behavioral changes.
- Golden Data must be perfect; otherwise, the model will learn your mistakes as "Rules."
- LIMA Principle: Almost all alignment happens within the first 1,000 high-quality samples.
In the next lesson, we will look at Data Source Identification, helping you find where those 100 examples are hiding in your organization.
Reflection Exercise
- If you had to choice between 100 pages of high-school essays or 5 pages of Nobel-prize-winning literature to teach a model "How to write a masterpiece," which would you choose?
- Why is "Memorization" bad in fine-tuning? (Hint: What happens if the user asks a question that is almost like a training example but has one key difference?)
SEO Metadata & Keywords
Focus Keywords: Quality vs Quantity Fine-Tuning, LIMA Paper Insight, 100 Example Rule LLM, Dataset Curation for AI, Golden Dataset Strategy. Meta Description: Discover the fundamental law of modern fine-tuning. Learn why a 'Golden Dataset' of 100 high-quality examples outperforms massive, noisy datasets and how to audit your data for success.