OpenAI Β· 2023-03
Rumored MoE ~1.8T total params (8Γ220B experts). First multimodal GPT-4 variant, launched with Vision. Architecture never officially disclosed.
MoE (rumored)
Multimodal
~1.8T params
OpenAI Β· 2024-05
Dense or small MoE. Natively multimodal β text, vision, audio in a single end-to-end model. Faster inference than GPT-4. Exact architecture undisclosed.
MoE (possible)
Natively Multimodal
128K context
Anthropic Β· 2024-03
Constitutional AI training. Exact architecture undisclosed, likely dense decoder. 200K token context window. Top benchmarks on release, surpassed GPT-4.
Likely Dense
200K context
Constitutional AI
Google Β· 2024-02
Confirmed MoE. Multimodal from ground up β handles text, image, audio, video natively. 1M token context window. Backbone of the Gemini 1.5 family.
MoE (confirmed)
Multimodal
1M context
Google Β· 2025-06
Confirmed MoE. Thinking mode (extended reasoning). #1 on most benchmarks mid-2025. Deep Research and agentic task support built-in.
MoE (confirmed)
Thinking mode
#1 benchmarks
xAI Β· 2025-02
Confirmed MoE. Trained on X (Twitter) data at massive scale. 128K context. Think mode for extended reasoning. Competes directly with GPT-4o and Claude Opus 4.
MoE (confirmed)
Think mode
128K context
OpenAI Β· 2025-05
Architecture undisclosed, likely MoE. Strong agentic reasoning, improved tool use. Integrates extended thinking. Positioned as OpenAI's flagship 2025 model.
Likely MoE
Agentic
1M context