AI Model Timeline
Key model releases from Transformer (2017) to today β filterable by category (LLMs, DLLMs, VLMs, Agents) and organization. β = PaperTrace deep-dive available. π = open weights.
Extended thinking, improved agentic capabilities β powers Claude Code. Released Feb 5, 2026.
Frontier performance across coding, agents, and professional work. Released Feb 17, 2026.
Native multimodal reasoning + agentic tool use β Google's strongest model at launch
Efficient MoE with novel load-balancing and multi-token prediction β strong coding and math at low training cost
o1-level reasoning via RL, fully open-weights MIT license β shocked markets
Best coding model at launch β outperforms GPT-4o on most benchmarks
Best open-weight model β competitive with GPT-4 Turbo on most tasks
Group Relative Policy Optimization β simpler PPO without critic network
Direct Preference Optimization β eliminates reward model and PPO from RLHF
RLHF alignment via PPO on human preferences β precursor to ChatGPT
Few-shot in-context learning emerges at scale β no gradient update needed
Zero-shot multitask LM β "too dangerous to release" at launch
Bidirectional Transformer pre-training with MLM β GLUE SOTA across all tasks
Attention Is All You Need β self-attention replaces RNNs entirely
Dates are approximate. Parameters are estimates where not officially confirmed.