AI Models
52 models · 0 new in 60d
Live
- ▾DeepSeek R1Open
DeepSeek · 128K tokens · self-host
Best for: Budget reasoning, self-hosted chain-of-thought, research
How: API is OpenAI-compatible. Self-host the 70B distill on 2x A100. MIT license = no restrictions.
Example: Run the 14B distill locally for debugging complex distributed system issues.
AIME 2024 79.8%SWE-bench 49.2%GPQA Diamond 71.5%reasoningmathcodingMIT licensedistillableHardware to self-hostVRAM: 10GB (14B distill) / 48GB (70B distill) / 1TB+ (full 671B)GPU: RTX 4090 (14B) · 2× A100 (70B) · 8× H100 (full)RAM: Full model needs 256GB+ system RAMFull 671B MoE is massive. Distilled versions (14B, 32B, 70B) are far more practical.
API: api.deepseek.com ($0.55/M in, $2.19/M out) OR self-host via vLLM/Ollama