← Back to Models
DeepSeek V3
by DeepSeekDeepSeek's largest model with 685B total parameters (37B active). A Mixture of Experts architecture that achieves state-of-the-art performance on reasoning and coding benchmarks. The model that shocked the industry.
chatcodereasoningmultilingual
Choose a size:
Download 685B MoE in Saga
Don't have Saga? Download it first
Or use the CLI:
ollama pull deepseek-v3:685b moeDetails
- License
- DeepSeek License
- Released
- December 2024
- Context
- 128K tokens
- Downloads
- 2.1M