by DeepSeek
DeepSeek R1 has 52.8M downloads on HuggingFace. 671B MoE architecture with 37B active parameters. Features enhanced chain-of-thought reasoning. Scores 52.8% on SWE-bench Verified.
Parameters
671B (MoE)
Architecture
Mixture of Experts
Context
256K
Provider
DeepSeek
Drop-in replacement for OpenAI API. Just change the base URL.
Only pay for actual GPU compute time. No idle costs.
99.9% uptime SLA, SOC 2 compliant, dedicated support.
Scales from zero to thousands of requests automatically.
| Fleek | Fireworks | Together | Baseten | |
|---|---|---|---|---|
| Input | $0.25 | $1.35 | $3.00 | — |
| Output | $1.00 | $5.40 | $7.00 | — |
| Savings | 70% | 70% | — |
Prices are per million tokens. Fleek pricing based on $0.0025/GPU-second.
See how much you'd save running DeepSeek R1 on Fleek
| Model Name | DeepSeek R1 |
| Total Parameters | 671B (MoE) |
| Active Parameters | 37B |
| Architecture | Mixture of Experts |
| Context Length | 256K tokens |
| Inference Speed | 16,000 tokens/sec |
| Provider | DeepSeek |
| Release Date | Jan 15, 2026 |
| License | MIT |
| HuggingFace | https://huggingface.co/deepseek-ai/DeepSeek-R1 |
Massive Multitask Language Understanding
OpenAI code generation benchmark
Mathematical problem solving
Software Engineering benchmark - real GitHub issues
Click any benchmark to view the official leaderboard. Rankings among open-source models.
Join the waitlist for early access. Start free with $5 in credits.