185 lines
4.8 KiB
Markdown
185 lines
4.8 KiB
Markdown
|
|
# Choosing the Right Model
|
||
|
|
|
||
|
|
**Different models, different strengths. Here's how to pick.**
|
||
|
|
|
||
|
|
OpenRouter gives you access to 100+ models. That's overwhelming. Let me cut through the noise.
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## The Quick Answer
|
||
|
|
|
||
|
|
Just getting started? Use these:
|
||
|
|
|
||
|
|
| Use Case | Model | Why |
|
||
|
|
|----------|-------|-----|
|
||
|
|
| General coding | `x-ai/grok-code-fast-1` | Fast, cheap, capable |
|
||
|
|
| Complex problems | `google/gemini-3-pro-preview` | 1M context, solid reasoning |
|
||
|
|
| Code-specific | `openai/gpt-5.1-codex` | Trained specifically for code |
|
||
|
|
| Budget mode | `minimax/minimax-m2` | Cheapest that actually works |
|
||
|
|
|
||
|
|
Pick one. Start working. Switch later if needed.
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Discovering Models
|
||
|
|
|
||
|
|
**Top recommended (curated list):**
|
||
|
|
```bash
|
||
|
|
claudish --top-models
|
||
|
|
```
|
||
|
|
|
||
|
|
**All OpenRouter models (hundreds):**
|
||
|
|
```bash
|
||
|
|
claudish --models
|
||
|
|
```
|
||
|
|
|
||
|
|
**Search for specific models:**
|
||
|
|
```bash
|
||
|
|
claudish --models grok
|
||
|
|
claudish --models codex
|
||
|
|
claudish --models gemini
|
||
|
|
```
|
||
|
|
|
||
|
|
**JSON output (for scripts):**
|
||
|
|
```bash
|
||
|
|
claudish --top-models --json
|
||
|
|
claudish --models --json
|
||
|
|
```
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Understanding the Columns
|
||
|
|
|
||
|
|
When you see the model table:
|
||
|
|
|
||
|
|
```
|
||
|
|
Model Provider Pricing Context Caps
|
||
|
|
google/gemini-3-pro-preview Google $7.00/1M 1048K ✓ ✓ ✓
|
||
|
|
```
|
||
|
|
|
||
|
|
**Model** - The ID you pass to `--model`
|
||
|
|
|
||
|
|
**Provider** - Who made it (Google, OpenAI, xAI, etc.)
|
||
|
|
|
||
|
|
**Pricing** - Average cost per 1 million tokens. Input and output prices vary, this is the midpoint.
|
||
|
|
|
||
|
|
**Context** - Maximum tokens the model can handle (input + output combined)
|
||
|
|
|
||
|
|
**Caps (Capabilities):**
|
||
|
|
- First ✓ = **Tools** - Can use Claude Code's file/bash tools
|
||
|
|
- Second ✓ = **Reasoning** - Extended thinking mode
|
||
|
|
- Third ✓ = **Vision** - Can analyze images/screenshots
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## My Honest Model Breakdown
|
||
|
|
|
||
|
|
### Grok Code Fast 1 (`x-ai/grok-code-fast-1`)
|
||
|
|
**Price:** $0.85/1M | **Context:** 256K
|
||
|
|
|
||
|
|
My daily driver. Fast responses, good code quality, reasonable price. Handles most tasks without drama.
|
||
|
|
|
||
|
|
**Good for:** General coding, refactoring, quick fixes
|
||
|
|
**Bad for:** Very long files (256K limit), vision tasks
|
||
|
|
|
||
|
|
### Gemini 3 Pro (`google/gemini-3-pro-preview`)
|
||
|
|
**Price:** $7.00/1M | **Context:** 1M (!)
|
||
|
|
|
||
|
|
The context king. A million tokens means you can dump entire codebases into context. Reasoning is solid. Vision works.
|
||
|
|
|
||
|
|
**Good for:** Large codebase analysis, complex architecture, image-based tasks
|
||
|
|
**Bad for:** Quick tasks (overkill), budget-conscious work
|
||
|
|
|
||
|
|
### GPT-5.1 Codex (`openai/gpt-5.1-codex`)
|
||
|
|
**Price:** $5.63/1M | **Context:** 400K
|
||
|
|
|
||
|
|
OpenAI's coding specialist. Trained specifically for software engineering. Does code review really well.
|
||
|
|
|
||
|
|
**Good for:** Code review, debugging, complex refactoring
|
||
|
|
**Bad for:** General chat (waste of a specialist)
|
||
|
|
|
||
|
|
### MiniMax M2 (`minimax/minimax-m2`)
|
||
|
|
**Price:** $0.60/1M | **Context:** 204K
|
||
|
|
|
||
|
|
The budget champion. Cheapest model that doesn't suck. Surprisingly capable for simple tasks.
|
||
|
|
|
||
|
|
**Good for:** Quick fixes, simple generation, high-volume tasks
|
||
|
|
**Bad for:** Complex reasoning, architecture decisions
|
||
|
|
|
||
|
|
### GLM 4.6 (`z-ai/glm-4.6`)
|
||
|
|
**Price:** $1.07/1M | **Context:** 202K
|
||
|
|
|
||
|
|
Underrated. Good balance of price and capability. Handles long context well.
|
||
|
|
|
||
|
|
**Good for:** Documentation, explanations, medium complexity tasks
|
||
|
|
**Bad for:** Cutting-edge reasoning
|
||
|
|
|
||
|
|
### Qwen3 VL (`qwen/qwen3-vl-235b-a22b-instruct`)
|
||
|
|
**Price:** $1.06/1M | **Context:** 131K
|
||
|
|
|
||
|
|
Vision + code combo. Best for when you need to work with screenshots, designs, or diagrams.
|
||
|
|
|
||
|
|
**Good for:** UI work from screenshots, diagram understanding, visual debugging
|
||
|
|
**Bad for:** Extended reasoning (no reasoning capability)
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Pricing Reality Check
|
||
|
|
|
||
|
|
Let's do real math.
|
||
|
|
|
||
|
|
**Average coding session:** ~50K tokens (input + output)
|
||
|
|
|
||
|
|
| Model | Cost per 50K tokens |
|
||
|
|
|-------|---------------------|
|
||
|
|
| MiniMax M2 | $0.03 |
|
||
|
|
| Grok Code Fast | $0.04 |
|
||
|
|
| GLM 4.6 | $0.05 |
|
||
|
|
| Qwen3 VL | $0.05 |
|
||
|
|
| GPT-5.1 Codex | $0.28 |
|
||
|
|
| Gemini 3 Pro | $0.35 |
|
||
|
|
|
||
|
|
For most tasks, we're talking cents. Don't obsess over pricing unless you're doing high-volume automation.
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Model Selection Strategy
|
||
|
|
|
||
|
|
**For experiments:** Start cheap (MiniMax M2). See if it works.
|
||
|
|
|
||
|
|
**For important code:** Use a capable model (Grok, Codex). It's still cheap.
|
||
|
|
|
||
|
|
**For architecture decisions:** Go premium (Gemini 3 Pro). Context and reasoning matter.
|
||
|
|
|
||
|
|
**For automation:** Pick the cheapest that works reliably for your task.
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Custom Models
|
||
|
|
|
||
|
|
See a model on OpenRouter that's not in our list? Use it anyway:
|
||
|
|
|
||
|
|
```bash
|
||
|
|
claudish --model anthropic/claude-sonnet-4.5 "your prompt"
|
||
|
|
claudish --model mistralai/mistral-large-2411 "your prompt"
|
||
|
|
```
|
||
|
|
|
||
|
|
Any valid OpenRouter model ID works.
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Force Update Model List
|
||
|
|
|
||
|
|
The model cache updates automatically every 2 days. Force it:
|
||
|
|
|
||
|
|
```bash
|
||
|
|
claudish --top-models --force-update
|
||
|
|
```
|
||
|
|
|
||
|
|
---
|
||
|
|
|
||
|
|
## Next
|
||
|
|
|
||
|
|
- **[Model Mapping](model-mapping.md)** - Use different models for different Claude Code roles
|
||
|
|
- **[Cost Tracking](../advanced/cost-tracking.md)** - Monitor your spending
|