AI Models

52 models · 0 new in 60d

Compare →
  • DeepSeek R1Open

    DeepSeek · 128K tokens · self-host

    Best for: Budget reasoning, self-hosted chain-of-thought, research

    How: API is OpenAI-compatible. Self-host the 70B distill on 2x A100. MIT license = no restrictions.

    Example: Run the 14B distill locally for debugging complex distributed system issues.

    AIME 2024 79.8%SWE-bench 49.2%GPQA Diamond 71.5%
    reasoningmathcodingMIT licensedistillable
    Hardware to self-host
    VRAM: 10GB (14B distill) / 48GB (70B distill) / 1TB+ (full 671B)
    GPU: RTX 4090 (14B) · 2× A100 (70B) · 8× H100 (full)
    RAM: Full model needs 256GB+ system RAM

    Full 671B MoE is massive. Distilled versions (14B, 32B, 70B) are far more practical.

    API: api.deepseek.com ($0.55/M in, $2.19/M out) OR self-host via vLLM/Ollama