Technique · training
Instruction Tuning (FLAN)
Fine-tuning a pretrained LM on a mixture of tasks phrased as natural-language instructions, enabling strong zero-shot generalization.
Deployment timeline
- GPT-4ohigh
Deployed 2026-02-16 · Velocity 4y
“GPT-4o follows instruction-tuned models like InstructGPT, enabling strong zero-shot task following.”
- GPT-5high
Deployed 2026-02-16 · Velocity 4y
“GPT-5 follows the instruction-following paradigm established by its predecessors, which were instruction-tuned.”
- GPT-5.2 Prohigh
Deployed 2026-02-17 · Velocity 4y
“As a flagship OpenAI model, it follows the established instruction-following paradigm refined from InstructGPT and GPT-4.”
- Claude 3high
Deployed 2026-02-18 · Velocity 4y
“Claude 3 is instruction-tuned to follow diverse instructions, a core capability mentioned in documentation.”
- Gemini 3 Prohigh
Deployed 2026-02-19 · Velocity 4y
“Gemini models are instruction-tuned, building on the FLAN line of work.”
- Gemini 3.1medium
Deployed 2026-02-20 · Velocity 4y
“Gemini models are instruction-tuned, building on Google's FLAN tradition.”
- Claude 3.5 Sonnethigh
Deployed 2026-02-23 · Velocity 4y
“Claude models are instruction-tuned on diverse tasks to follow instructions and generalize to new tasks.”
- Claude Sonnet 4.6medium
Deployed 2026-02-25 · Velocity 4y
“Claude models are instruction-tuned to follow user instructions, a standard practice for Anthropic.”
- GPT-5.3medium
Deployed 2026-02-26 · Velocity 4y
“GPT-5.3 follows OpenAI's instruction-tuned lineage (InstructGPT) for strong zero-shot generalization.”
- Claude 4.5high
Deployed 2026-02-26 · Velocity 4y
“Claude models are instruction-tuned to follow natural language instructions effectively.”
- Gemini 3 Flashhigh
Deployed 2026-02-27 · Velocity 4y
“Gemini models are instruction-tuned, building upon the FLAN instruction-tuning methodology developed by Google.”
- Kimi K2.5medium
Deployed 2026-03-04 · Velocity 5y
“Kimi models are instruction-tuned for conversational ability, aligning with FLAN-style training.”
- Gemini 3.1 Flash-Litehigh
Deployed 2026-03-05 · Velocity 5y
“Gemini models are instruction-tuned, building on the FLAN instruction-tuning methodology.”
- Claude 3.5 Opushigh
Deployed 2026-03-18 · Velocity 5y
“Claude models are instruction-tuned; Claude 3.5 Opus shows strong instruction following.”
- GLM-5.1medium
Deployed 2026-03-21 · Velocity 5y
“GLM-5.1 is instruction-tuned on diverse tasks following FLAN methodology.”
- Qwen 3.6high
Deployed 2026-03-31 · Velocity 5y
“Qwen models are instruction-tuned on a large collection of datasets.”
- GPT-5.4-Cyberhigh
Deployed 2026-04-16 · Velocity 5y
“GPT models are instruction-tuned to follow diverse tasks, aligning with the FLAN paradigm.”