Recipe ·
GPT-4o
OpenAI flagship multimodal model. Text, images, audio natively. Faster/cheaper than GPT-4. Powers ChatGPT free tier. O-series expanded with o1, o3, o4-mini for reasoning.
Ingredient list
Invented by Google · 2022-11 · Velocity 3y
“GPT-4o's speed improvements align with inference optimizations like speculative decoding, though not explicitly confirmed.”
inferencelowInvented by Stanford · 2022-05 · Velocity 4y
“OpenAI's Triton kernels, used for GPT-4, are predecessors to FlashAttention. GPT-4o's speed improvements suggest optimized attention.”
inferencemediumInvented by University of Tokyo · 2022-05 · Velocity 4y
“GPT-4o exhibits zero-shot reasoning when prompted with 'think step by step', a hallmark of zero-shot CoT.”
reasoninghighInvented by Google · 2022-03 · Velocity 4y
“System Card mentions sampling multiple outputs for evaluation, a core self-consistency technique.”
reasoningmediumInvented by Google · 2022-01 · Velocity 4y
“GPT-4o can be prompted to show step-by-step reasoning, a core CoT capability.”
reasoninghighInvented by Google · 2021-09 · Velocity 4y
“GPT-4o follows instruction-tuned models like InstructGPT, enabling strong zero-shot task following.”
traininghighInvented by Zhuiyi Technology · 2021-04 · Velocity 5y
“GPT models since GPT-3 use rotary position embeddings (RoPE). GPT-4o's architecture is a direct evolution.”
architecturehighInvented by Google · 2017-06 · Velocity 9y
“GPT-4o is a Transformer-based model, the core architecture of all GPT models.”
architecturehighInvented by Google · 2017-01 · Velocity 9y
“GPT-4 is widely reported as a MoE model. GPT-4o is its successor, implying architectural continuity.”
architecturemedium
This recipe is part of the gentic.news Deployment Atlas. Every ingredient has an origin paper + evidence. Methodology is public. Dataset is CC BY 4.0.