Skip to main content
Proliferate routes all LLM requests through its built-in proxy. You never need to bring your own API keys — every supported model works out of the box. On Proliferate Cloud, the latest models are available within 20 minutes of release. Self-hosted deployments work with any model provider. Bring your own API keys and route through Azure OpenAI, AWS Bedrock, Google Vertex AI, or connect directly to any provider. See the self-hosting guide and environment configuration for setup.

Supported Models

ModelProviderBest For
Claude Sonnet 4.6 (default)AnthropicAgents and coding tasks. The best balance of speed, cost, and quality.
Claude Opus 4.6AnthropicComplex multi-step tasks requiring deep reasoning.
GPT-5.2OpenAIOpenAI flagship thinking model.
Gemini 3 ProGoogleLarge-context tasks (1M token context window).
Gemini 3 FlashGoogleFast, lightweight tasks where speed matters most.
DeepSeek V3DeepSeekStrong open-weight coding model.
DeepSeek R1DeepSeekReasoning-heavy tasks with chain-of-thought.
Grok 4.1xAIxAI flagship model.
Grok 4.1 FastxAIQuick xAI tasks where latency is critical.
CodestralMistralCode-specialized tasks from Mistral.
Claude Sonnet 4.6 is the default model for all new sessions and automations. It offers the best combination of coding ability, speed, and cost efficiency.

Providers

Proliferate supports models from six providers:
  • Anthropic — Claude Sonnet 4.6, Claude Opus 4.6
  • OpenAI — GPT-5.2
  • Google — Gemini 3 Pro, Gemini 3 Flash
  • DeepSeek — DeepSeek V3, DeepSeek R1
  • xAI — Grok 4.1, Grok 4.1 Fast
  • Mistral — Codestral

No API Keys Required

Proliferate’s LLM proxy handles all authentication between the agent and the model provider. This means:
  • No key management — You never create, rotate, or store API keys for any model provider.
  • Sandboxes never see real keys — The proxy generates scoped, short-lived virtual keys for each session. The sandbox environment has no access to your actual provider credentials.
  • Per-org spend tracking — All model usage is tracked and attributed to your organization automatically. You can monitor spend across models and sessions from the dashboard.
Self-hosted deployments bring their own API keys and can route through Azure OpenAI, AWS Bedrock, Google Vertex AI, or any provider directly. See environment configuration for details.

Reasoning Effort

Some models support configurable reasoning effort, which controls how much “thinking” the model does before responding. This lets you trade speed for depth depending on the task.
LevelBehavior
QuickMinimal reasoning. Fastest responses. Best for simple, well-defined tasks.
Normal (default)Balanced reasoning. Good for most coding tasks.
DeepMaximum reasoning depth. Best for complex architecture decisions, debugging, and multi-step planning.
Models that support reasoning effort:
  • Claude Sonnet 4.6
  • Claude Opus 4.6
  • GPT-5.2
  • DeepSeek R1
  • Grok 4.1
Models without reasoning support (Gemini 3 Pro, Gemini 3 Flash, DeepSeek V3, Grok 4.1 Fast, Codestral) ignore the reasoning effort setting and always use their default behavior.

Choosing the Right Model

Use CaseRecommended ModelWhy
General coding tasksClaude Sonnet 4.6Best all-around coding agent. Fast, accurate, cost-effective.
Complex refactoringClaude Opus 4.6Deeper reasoning handles large-scale architectural changes.
Large codebase analysisGemini 3 Pro1M context window fits entire codebases in a single pass.
Quick fixes and typosGemini 3 FlashFastest response time for simple changes.
Debugging with chain-of-thoughtDeepSeek R1Explicit reasoning traces help with hard-to-find bugs.
Cost-sensitive workloadsDeepSeek V3Strong open-weight model at lower cost.
Code generation tasksCodestralPurpose-built for code by Mistral.

Configuring Models

Per session

When starting a session from the web dashboard or terminal, select the model from the model picker before sending your first message. You can also change the model mid-session.

Per automation

Each automation has a model setting in its configuration. Set it when creating or editing the automation:
  1. Go to the automation settings
  2. Under Model, select the model and optionally set reasoning effort
  3. Save the automation
All runs triggered by that automation will use the configured model.

Via the API

When creating sessions or automations through the API, pass the modelId field:
{
  "modelId": "claude-sonnet-4.6",
  "reasoningEffort": "normal"
}
Valid model IDs: claude-sonnet-4.6, claude-opus-4.6, gpt-5.2, gemini-3-pro, gemini-3-flash, deepseek-v3, deepseek-r1, grok-4.1, grok-4.1-fast, codestral.