Llama 4 Behemoth
Meta's flagship multimodal model with massive MoE architecture (288B active parameters)
Llama 4 Behemoth
Meta • April 2025
Training Data
Up to August 2024
Llama 4 Behemoth
April 2025
Parameters
~2 trillion (288B active)
Training Method
Mixture of Experts
Context Window
1,000,000 tokens
Knowledge Cutoff
August 2024
Key Features
Open Source • Massive MoE Architecture • Multimodal
Capabilities
Reasoning: Outstanding
STEM: Outstanding
Complex Tasks: Outstanding
What's New in This Version
Massive MoE model with 16 experts designed for complex reasoning tasks
Meta's flagship multimodal model with massive MoE architecture (288B active parameters)
What's New in This Version
Massive MoE model with 16 experts designed for complex reasoning tasks
Technical Specifications
Key Features
Capabilities
Other Meta Models
Explore more models from Meta
Llama 4 Maverick
Meta's balanced multimodal MoE model with 128 experts for general use
Llama 4 Scout
Meta's efficient multimodal model with industry-leading 10M token context
Llama 3.1 405B
Meta's largest and most capable open-source model