Skip to main content

Available Models

Below is a list of all available models. Models are categorized by type:
  • Frontier: Best coding agent models
  • Research: Expensive and limited heavily
  • General: Cheaper but less intelligent models
IDNameTypeLimit
claude-opus-4-20250514Claude Opus 4Research200K
claude-opus-4-1-20250805Claude Opus 4.1Research200K
o3-proOpenAI-o3-proResearch200K
claude-haiku-4-5-20251001Claude Haiku 4.5Frontier200K
claude-sonnet-4-20250514Claude Sonnet 4Frontier200K
claude-sonnet-4-1mClaude Sonnet 4 (1M)Frontier1.0M
claude-sonnet-4-5-20250929Claude Sonnet 4.5Frontier200K
claude-sonnet-4-5-1mClaude Sonnet 4.5 (1M)Frontier1.0M
gpt-5GPT-5Frontier400K
gpt-5-highGPT-5 (High Reasoning)Frontier400K
gpt-5-high-fastGPT-5 (High Reasoning, Priority)Frontier400K
gpt-5-lowGPT-5 (Low Reasoning)Frontier400K
gpt-5-low-fastGPT-5 (Low Reasoning, Priority)Frontier400K
gpt-5-fastGPT-5 (Priority)Frontier400K
gpt-5-codexGPT-5 CodexFrontier400K
gpt-5-codex-highGPT-5 Codex (High Reasoning)Frontier400K
gpt-5-codex-lowGPT-5 Codex (Low Reasoning)Frontier400K
claude-3.5-sonnetClaude-3.5 SonnetGeneral200K
claude-3.7-sonnetClaude-3.7 SonnetGeneral200K
deepseekr1Deepseek r1General164K
deepseekv3Deepseek v3General128K
gpt-4.1GPT-4.1General1.0M
gpt-4oGPT-4oGeneral128K
gpt-5-miniGPT-5-miniGeneral400K
gpt-oss-120bGPT-OSS-120bGeneral131K
gemini-2.5-proGemini 2.5 ProGeneral1.0M
grok-4Grok 4General256K
grok-4-fastGrok 4 FastGeneral2.0M
grok-code-fast-1Grok Code Fast 1General256K
grok-3Grok-3General131K
o3OpenAI-o3General200K
o3-miniOpenAI-o3-miniGeneral200K
o4-miniOpenAI-o4-miniGeneral200K
qwen-3-coder-480bQwen-3-Coder-480bGeneral131K

Models Picker

You can see what model you are currently using by clicking the dropdown at the bottom left of the chat interface: Screenshot showing the model dropdown at the bottom left of the chat interface
Models can misreport their name/version at runtime. Rely on the model ID you selected in Firebender, not the model’s self-identification response.

Example: Deepseek v3 misidentifies itself as gpt-4

Screenshot showing a model reporting the wrong ID

Context Windows

A context window is the maximum span of tokens (text and code) a model can consider at once. The more prompts, files, and responses in a session, the more context is consumed. If you run into context window limit, its recommended to summarize or start a new chat.

Avoiding Window Limits

  • prune chats frequently if the coding agent does something wrong (instead of telling the agent to undo its changes, restart the prompt with approaches to avoid)
  • scope down changes that can be easily verifiable (e.g. running tests, reviewing UI change, etc.)
You can monitor how close you are to the context limit at the bottom of each chat: Screenshot showing context window monitoring at the bottom of the chat interface

Default Mode

Default Mode automatically selects the best model for each task based on request complexity and availability. When a rate limit is hit, Firebender automatically switches to the next best model.

Controls

You can specify models using one of the following methods:
  • Deep Links: jetbrains://idea/firebender/chat?model=claude-sonnet-4-20250514
  • Commands: Configure in firebender.json with "model": "claude-sonnet-4-20250514"
  • Plugin Interface: Select from the model dropdown in the Firebender chat interface

Restrict models for your team

Organizations can configure which models are available to team members. Learn more about Business organizations.