Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions docs/cli/configuration/mixed-models.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -131,8 +131,8 @@ Here are some effective model combinations for different scenarios:
**Best for:** Most development workflows, flexible reasoning control

- **Default model:** Haiku 4.5 or Sonnet 4.5 (select reasoning of your choice if supported)
- **Spec mode model:** Opus 4.6 (reasoning: high) or Sonnet 4.5 (reasoning: high)
- **Benefits:** Fast implementation with deep planning analysis. Opus 4.6 offers **Max** reasoning for the most thorough planning.
- **Spec mode model:** Opus 4.7 (reasoning: high) or Sonnet 4.5 (reasoning: high)
- **Benefits:** Fast implementation with deep planning analysis. Opus 4.7 offers **Max** reasoning for the most thorough planning.

### OpenAI Models

Expand Down
3 changes: 2 additions & 1 deletion docs/cli/configuration/settings.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,8 @@ Local overrides merge on top of the corresponding `settings.json` at the same le
Choose the default AI model that powers your droid:

- **`opus`** - Claude Opus 4.5 (current default)
- **`opus-4-6`** - Claude Opus 4.6, latest flagship with Max reasoning
- **`opus-4-7`** - Claude Opus 4.7, newest flagship with Max reasoning (2×; 1× promotional until April 30)
- **`opus-4-6`** - Claude Opus 4.6, previous flagship with Max reasoning
- **`opus-4-6-fast`** - Claude Opus 4.6 Fast, tuned for faster responses
- **`sonnet`** - Claude Sonnet 4.5, balanced cost and quality
- **`sonnet-4-6`** - Claude Sonnet 4.6, Max reasoning at the Sonnet price point
Expand Down
42 changes: 22 additions & 20 deletions docs/cli/user-guides/choosing-your-model.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -12,22 +12,23 @@ Model quality evolves quickly, and we tune the CLI defaults as the ecosystem shi

| Rank | Model | Why we reach for it |
| ---- | ----------------------------- | ------------------------------------------------------------------------------------------------------------------------------------------------ |
| 1 | **Claude Opus 4.6** | Latest Anthropic flagship with **Max** reasoning; best depth and safety for complex work. |
| 2 | **Claude Opus 4.6 Fast** | Opus 4.6 tuned for faster response times; 12× multiplier. |
| 3 | **Claude Opus 4.5** | Proven quality-and-safety balance; strong default for TUI and exec. |
| 4 | **Claude Sonnet 4.6** | **Max** reasoning at the Sonnet price point (1.2×); strong daily driver for planning and implementation. |
| 5 | **GPT-5.4** | Latest OpenAI model with 922K context, 128K output, verbosity support, and **Extra High** reasoning; excellent for large-context tasks. |
| 6 | **Claude Sonnet 4.5** | Strong daily driver with balanced cost/quality; great general-purpose choice when you don't need Opus-level depth. |
| 7 | **GPT-5.3-Codex** | Newest OpenAI coding model with **Extra High** reasoning and verbosity support; strong for implementation-heavy tasks. |
| 8 | **GPT-5.2-Codex** | Proven OpenAI coding model with **Extra High** reasoning; solid for implementation-heavy tasks. |
| 9 | **GPT-5.2** | OpenAI model with verbosity support and reasoning up to **Extra High**. |
| 10 | **Claude Haiku 4.5** | Fast, cost-efficient for routine tasks and high-volume automation. |
| 11 | **Gemini 3.1 Pro** | Newer Gemini Pro generation with strong structured outputs and mixed reasoning controls for research-heavy tasks. |
| 12 | **Gemini 3 Flash** | Fast, cheap (0.2× multiplier) with full reasoning support; great for high-volume tasks where speed matters. |
| 13 | **Droid Core (MiniMax M2.5)** | Open-source, 0.12× multiplier with reasoning support (Low/Medium/High); cheapest model available. No image support. |
| 14 | **Droid Core (GLM-5)** | Open-source, 0.4× multiplier with updated GLM capabilities for bulk automation and air-gapped environments; no image support. |
| 15 | **Droid Core (GLM-4.7)** | Open-source, 0.25× multiplier, stable choice for bulk automation or air-gapped environments; note: no image support. |
| 16 | **Droid Core (Kimi K2.5)** | Open-source, 0.25× multiplier with image support; good for cost-sensitive work. |
| 1 | **Claude Opus 4.7** | Newest Anthropic flagship with **Max** reasoning; top pick for the hardest work. Promotional 1× multiplier through April 30 (2× afterward). |
| 2 | **Claude Opus 4.6** | Previous Anthropic flagship with **Max** reasoning; still excellent depth and safety for complex work. |
| 3 | **Claude Opus 4.6 Fast** | Opus 4.6 tuned for faster response times; 12× multiplier. |
| 4 | **Claude Opus 4.5** | Proven quality-and-safety balance; strong default for TUI and exec. |
| 5 | **Claude Sonnet 4.6** | **Max** reasoning at the Sonnet price point (1.2×); strong daily driver for planning and implementation. |
| 6 | **GPT-5.4** | Latest OpenAI model with 922K context, 128K output, verbosity support, and **Extra High** reasoning; excellent for large-context tasks. |
| 7 | **Claude Sonnet 4.5** | Strong daily driver with balanced cost/quality; great general-purpose choice when you don't need Opus-level depth. |
| 8 | **GPT-5.3-Codex** | Newest OpenAI coding model with **Extra High** reasoning and verbosity support; strong for implementation-heavy tasks. |
| 9 | **GPT-5.2-Codex** | Proven OpenAI coding model with **Extra High** reasoning; solid for implementation-heavy tasks. |
| 10 | **GPT-5.2** | OpenAI model with verbosity support and reasoning up to **Extra High**. |
| 11 | **Claude Haiku 4.5** | Fast, cost-efficient for routine tasks and high-volume automation. |
| 12 | **Gemini 3.1 Pro** | Newer Gemini Pro generation with strong structured outputs and mixed reasoning controls for research-heavy tasks. |
| 13 | **Gemini 3 Flash** | Fast, cheap (0.2× multiplier) with full reasoning support; great for high-volume tasks where speed matters. |
| 14 | **Droid Core (MiniMax M2.5)** | Open-source, 0.12× multiplier with reasoning support (Low/Medium/High); cheapest model available. No image support. |
| 15 | **Droid Core (GLM-5)** | Open-source, 0.4× multiplier with updated GLM capabilities for bulk automation and air-gapped environments; no image support. |
| 16 | **Droid Core (GLM-4.7)** | Open-source, 0.25× multiplier, stable choice for bulk automation or air-gapped environments; note: no image support. |
| 17 | **Droid Core (Kimi K2.5)** | Open-source, 0.25× multiplier with image support; good for cost-sensitive work. |

<Note>
We ship model updates regularly. When a new release overtakes the list above,
Expand All @@ -40,14 +41,14 @@ Model quality evolves quickly, and we tune the CLI defaults as the ecosystem shi

| Scenario | Recommended model |
| ---------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------------------------------ |
| **Deep planning, architecture reviews, ambiguous product specs** | Start with **Opus 4.6** for best depth and safety, or **Opus 4.6 Fast** for faster turnaround. Use **Sonnet 4.6** or **Sonnet 4.5** when you want balanced cost/quality, or **GPT-5.4** for large-context reasoning. |
| **Full-feature development, large refactors** | **Opus 4.6** or **Opus 4.5** for depth and safety. **GPT-5.4**, **GPT-5.3-Codex**, or **GPT-5.2-Codex** when you need speed plus **Extra High** reasoning; **Sonnet 4.6** or **Sonnet 4.5** for balanced loops. |
| **Deep planning, architecture reviews, ambiguous product specs** | Start with **Opus 4.7** for best depth and safety (1× promotional multiplier through April 30), or fall back to **Opus 4.6** / **Opus 4.6 Fast** for faster turnaround. Use **Sonnet 4.6** or **Sonnet 4.5** when you want balanced cost/quality, or **GPT-5.4** for large-context reasoning. |
| **Full-feature development, large refactors** | **Opus 4.7** or **Opus 4.6** for depth and safety. **GPT-5.4**, **GPT-5.3-Codex**, or **GPT-5.2-Codex** when you need speed plus **Extra High** reasoning; **Sonnet 4.6** or **Sonnet 4.5** for balanced loops. |
| **Repeatable edits, summarization, boilerplate generation** | **Haiku 4.5** or **Droid Core** (including **MiniMax M2.5** at 0.12×) for speed and cost. **GPT-5.2** when you need higher quality or structured outputs. |
| **CI/CD or automation loops** | Favor **Haiku 4.5** or **Droid Core** for predictable, low-cost throughput. Use **GPT-5.3-Codex** or **GPT-5.4** when automation needs stronger reasoning. |
| **High-volume automation, frequent quick turns** | **Haiku 4.5** for speedy feedback. **Droid Core** (especially **MiniMax M2.5** at 0.12× with reasoning) when cost is critical or you need air-gapped deployment. |

<Tip>
**Claude Opus 4.6** is the top-tier option for extremely complex architecture decisions or critical work where you need maximum reasoning capability. **Opus 4.6 Fast** is tuned for faster responses at a higher cost. Most tasks don't require Opus-level power—start with Sonnet 4.6 or Sonnet 4.5 and escalate only if needed.
**Claude Opus 4.7** is the newest top-tier option for extremely complex architecture decisions or critical work where you need maximum reasoning capability—and it runs at a promotional 1× multiplier through April 30 (2× afterward). **Claude Opus 4.6** remains an excellent alternative, and **Opus 4.6 Fast** is tuned for faster responses at a higher cost. Most tasks don't require Opus-level power—start with Sonnet 4.6 or Sonnet 4.5 and escalate only if needed.
</Tip>

Tip: you can swap models mid-session with `/model` or by toggling in the settings panel (`Shift+Tab` → **Settings**).
Expand All @@ -65,6 +66,7 @@ Tip: you can swap models mid-session with `/model` or by toggling in the setting

## 4 · Reasoning effort settings

- **Opus 4.7**: Off / Low / Medium / High / **Max** (default: High)
- **Opus 4.6 / Opus 4.6 Fast**: Off / Low / Medium / High / **Max** (default: High)
- **Sonnet 4.6**: Off / Low / Medium / High / **Max** (default: High)
- **Opus 4.5 / Sonnet 4.5 / Haiku 4.5**: Off / Low / Medium / High (default: Off)
Expand All @@ -79,7 +81,7 @@ Tip: you can swap models mid-session with `/model` or by toggling in the setting
- **Droid Core (Kimi K2.5)**: None only (default: None)
- **Droid Core (MiniMax M2.5)**: Low / Medium / High (default: High)

Reasoning effort increases latency and cost—start low for simple work and escalate as needed. **Max** is available on Claude Opus 4.6 and Sonnet 4.6. **Extra High** is available on GPT-5.4, GPT-5.2, GPT-5.2-Codex, and GPT-5.3-Codex.
Reasoning effort increases latency and cost—start low for simple work and escalate as needed. **Max** is available on Claude Opus 4.7, the Opus 4.6 family (Opus 4.6 and Opus 4.6 Fast), and Sonnet 4.6. **Extra High** is available on GPT-5.4, GPT-5.2, GPT-5.2-Codex, and GPT-5.3-Codex.

<Tip>
Change reasoning effort from `/model` → **Reasoning effort**, or via the
Expand Down
2 changes: 1 addition & 1 deletion docs/guides/power-user/prompt-crafting.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -372,7 +372,7 @@ Match the model to the task:

| Task Type | Recommended Model | Reasoning Level |
|-----------|-------------------|-----------------|
| **Complex architecture** | Opus 4.6 or Opus 4.5 | High-Max |
| **Complex architecture** | Opus 4.7 or Opus 4.6 | High-Max |
| **Feature implementation** | Sonnet 4.5 or GPT-5.1-Codex | Medium |
| **Quick edits, formatting** | Haiku 4.5 | Off/Low |
| **Code review** | GPT-5.1-Codex-Max | High |
Expand Down
5 changes: 3 additions & 2 deletions docs/guides/power-user/token-efficiency.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -146,16 +146,17 @@ Different models have different cost multipliers and capabilities. Match the mod
| Gemini 3.1 Pro | 0.8× | Research, analysis with newer Gemini generation |
| Claude Sonnet 4.5 | 1.2× | Balanced quality/cost |
| Claude Opus 4.5 | 2× | Complex reasoning, architecture |
| Claude Opus 4.6 | 2× | Latest flagship, Max reasoning |
| Claude Opus 4.6 | 2× | Previous flagship, Max reasoning |
| Claude Opus 4.6 Fast | 12× | Opus 4.6 tuned for faster responses |
| Claude Opus 4.7 | 2× (1× until April 30) | Newest flagship, discounted during launch window |
Comment thread
shashank-factory marked this conversation as resolved.

### Task-Based Model Selection

```
Simple edit, formatting → Haiku 4.5 (0.4×)
Implement feature from spec → GPT-5.1-Codex (0.5×)
Debug complex issue → Sonnet 4.5 (1.2×)
Architecture planning → Opus 4.6 (2×) or Opus 4.5 (2×)
Architecture planning → Opus 4.7 (2×, 1× until April 30) or Opus 4.6 (2×)
Bulk file processing → Droid Core (GLM-4.7 at 0.25× or GLM-5 at 0.4×)
```

Expand Down
4 changes: 2 additions & 2 deletions docs/jp/cli/configuration/mixed-models.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -131,8 +131,8 @@ spec modeモデルの推論努力は、デフォルトモデルの推論設定
**最適な用途:** ほとんどの開発ワークフロー、柔軟な推論制御

- **デフォルトモデル:** Haiku 4.5またはSonnet 4.5(サポートされている場合は任意の推論を選択)
- **Spec modeモデル:** Opus 4.6(推論:高)またはSonnet 4.5(推論:高)
- **利点:** 深いプランニング分析による高速実装。Opus 4.6は最も徹底的なプランニングのために**Max**推論を提供
- **Spec modeモデル:** Opus 4.7(推論:高)またはSonnet 4.5(推論:高)
- **利点:** 深いプランニング分析による高速実装。Opus 4.7は最も徹底的なプランニングのために**Max**推論を提供

### OpenAI モデル

Expand Down
3 changes: 2 additions & 1 deletion docs/jp/cli/configuration/settings.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,8 @@ droidの設定を構成するには:
droidを動かすデフォルトのAIモデルを選択します:

- **`opus`** - Claude Opus 4.5(現在のデフォルト)
- **`opus-4-6`** - Claude Opus 4.6、Max推論機能を持つ最新フラグシップモデル
- **`opus-4-7`** - Claude Opus 4.7、Max推論機能を持つ最新フラグシップモデル(2×;4月30日まで1×のプロモーション価格)
- **`opus-4-6`** - Claude Opus 4.6、Max推論機能を持つ前世代フラグシップモデル
- **`opus-4-6-fast`** - Claude Opus 4.6 Fast、高速レスポンス用にチューニングされたモデル
- **`sonnet`** - Claude Sonnet 4.5、コストと品質のバランス型
- **`sonnet-4-6`** - Claude Sonnet 4.6、Sonnet価格帯でMax推論を搭載
Expand Down
Loading