Skip to content

Providers

ProviderFree Tier (per key)ModelsGet a Key
Groq~30 req/minLlama 3.3 70B, Llama 3.1 8B, Llama 4 Scout, Qwen3 32Bconsole.groq.com
Gemini~15 req/minGemini 2.5 Flash, 2.5 Pro, 2.0 Flash, 2.0 Flash Liteaistudio.google.com
Mistral~5 req/minMistral Small, Medium, Nemoconsole.mistral.ai
Cerebras~30 req/minLlama 3.1 8B, Qwen3 235B, GPT-OSS 120Bcloud.cerebras.ai
NVIDIA NIM~40 req/minLlama 3.3 70B, Llama 3.1 405B, Nemotron 70B, DeepSeek R1build.nvidia.com
OllamaUnlimited (local)Any local modelollama.com

Combined free capacity: ~120 req/min with one key each, ~360 req/min with three keys per provider via Multi-Key Rotation. All $0.

Adding API keys

Set the corresponding env var. Each accepts a single key OR a comma-separated list:

Terminal window
GROQ_API_KEY=gsk_...,gsk_another,gsk_third # 3× the Groq capacity
GEMINI_API_KEY=AI...
MISTRAL_API_KEY=...
CEREBRAS_API_KEY=csk_...
NVIDIA_NIM_API_KEY=nvapi-...
OLLAMA_BASE_URL=http://localhost:11434

Only the providers with valid keys are enabled at runtime. You don’t need to provide all six.