fireworks/models/deepseek-v3-0324
Display Name: Deepseek V3 03-24
A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Updated checkpoint.
Specifications
Context160,000
Inputtext
Outputtext, json
Performance (7-day Average)
Uptime
TPS
RURT
API Paths
/v1/chat/completions
Pricing
Input$0.90× 1.1/ MTokens
Output$0.90× 1.1/ MTokens