AI Models
52 models · 2 new in 60d
- ▾Gemma 4 27B MoENewOpen
Google · 128K tokens · self-host
Best for: Faster self-hosted inference, cost-efficient multimodal
How: MoE variant — faster inference than the 31B dense. Same multimodal capabilities.
Example: Process image-based monitoring alerts faster than the dense variant at the same quality.
LMSYS Arena #6 textMoE efficiencymultimodalimages + videoApache 2.0Hardware to self-hostVRAM: 18GB (quantized) / 54GB (FP16)GPU: RTX 4090 24GB or 1× A100 40GBRAM: 32GB+ system RAM27B total MoE — faster inference than the 31B dense thanks to sparse activations.
API: Ollama, vLLM, Hugging Face. ollama run gemma4:27b-moe
- ▾Gemma 4 E4BNewOpen
Google · 128K tokens · self-host
Best for: Edge, mobile, IoT, on-device AI with multimodal input
How: 4B params — runs on any hardware. Supports images, video, AND native audio input.
Example: Run on a Raspberry Pi to process security camera feeds with voice commands.
tinyon-devicemultimodal + audioApache 2.0Hardware to self-hostVRAM: 3GB (quantized) / 8GB (FP16)GPU: Any — CPU, phone, Jetson, Raspberry Pi 5, integrated GPURAM: 4-8GB system RAM4B params. Edge-first design: runs on phones, SBCs, IoT devices.
API: Ollama, Hugging Face. Runs on phones and Raspberry Pi.
- ▾Ministral 3 (3B/8B/14B)Open
Mistral · 128K tokens · self-host
Best for: Edge deployment, on-device AI, lightweight vision tasks
How: 3B fits on phones, 8B on laptops, 14B on dev GPUs. All have vision support.
Example: Run 8B on a Jetson to classify manufacturing defects from camera feeds.
edge-friendlyvisiondense3 sizesHardware to self-hostVRAM: 2GB (3B) / 6GB (8B) / 10GB (14B quantized)GPU: Phone/CPU (3B) · Laptop GPU (8B) · RTX 3060+ (14B)RAM: 8-16GB system RAMAll three sizes are dense with vision. 3B runs on phones, 8B on laptops, 14B on dev GPUs.
API: Ollama, vLLM, Hugging Face. Also on Mistral API.
- ▾Qwen 3 30BOpen
Alibaba · 128K tokens · self-host
Best for: Local development, laptop-friendly reasoning, privacy
How: Excellent for local dev. MoE means only 3B params active — fast on consumer hardware.
Example: Run on your dev machine as a private coding assistant with reasoning.
AIME 2024 66.7%MoE 3B active / 30B totalruns on consumer GPUhybrid thinkingHardware to self-hostVRAM: 20GB (quantized) / 60GB (FP16)GPU: RTX 4090 24GB (quantized) or 1× A100RAM: 32GB+ system RAM30B total (3B active). The 3B active params make inference fast on consumer hardware.
API: ollama run qwen3:30b — fits on RTX 4090 (24GB)
- ▾Gemma 3 27BOpen
Google · 128K tokens · self-host
Best for: On-device/edge deployment, multimodal at small scale
How: ollama run gemma3:27b. Fits on RTX 3090/4090. Good multimodal + tool use at small size.
Example: Run on a dev server to process screenshots and generate bug reports.
MMLU 75.6%HumanEval 78.0%compactmultimodalruns on single GPUfunction callingHardware to self-hostVRAM: 18GB (quantized) / 54GB (FP16)GPU: RTX 3090/4090 24GB or 1× A100 40GBRAM: 32GB+ system RAM27B dense. Fits on a single high-end consumer GPU with quantization.
API: Ollama, vLLM, Hugging Face. Also on Vertex AI.
- ▾Phi-4Open
Microsoft · 16K tokens · self-host
Best for: Edge deployment, STEM tasks, embedded AI in products
How: ollama run phi4. MIT license — embed in commercial products freely.
Example: Embed in a CI pipeline to validate config files and Terraform plans.
GPQA Diamond 56.2%MATH 80.4%14B paramsSTEM reasoningMIT licenseruns on laptopHardware to self-hostVRAM: 9GB (quantized) / 28GB (FP16)GPU: Any 8GB+ GPU (RTX 3060, laptop 4050, etc.)RAM: 16GB system RAM14B dense. Runs locally on most developer laptops with quantization.
API: Ollama, Hugging Face, Azure AI