Peripheral tier: cheap background thinking ticks (~500 token prompt, ~200 token response) Research date: 2026-02-15
| Model | Built-in Provider | Custom Config Needed |
|---|---|---|
| Kimi (Moonshot) | β Not built-in | β
Yes β models.providers with OpenAI-compatible baseUrl |
| GLM (Zhipu/Z.AI) | β
Built-in (zai) |
No β just set ZAI_API_KEY |
| Qwen (Alibaba) | β
Built-in (qwen-portal plugin) |
No β OAuth device flow or DashScope API key |
| DeepSeek | β Via HuggingFace or OpenRouter | Custom provider also works (OpenAI-compatible) |
| Mistral | β
Built-in (mistral) |
No β just set MISTRAL_API_KEY |
| Ollama (local) | β
Built-in (ollama) |
No β auto-detected at localhost:11434 |
- Sign up: https://platform.moonshot.ai β email registration, $5 bonus on first $5 recharge
- API Endpoint:
https://api.moonshot.ai/v1(OpenAI-compatible) - Auth: Bearer token (
MOONSHOT_API_KEY) - OpenAI-compatible: β Yes β chat completions format
- Models:
kimi-k2.5,kimi-k2-0905-preview,kimi-k2-turbo-preview,kimi-k2-thinking - Context: up to 131Kβ256K tokens
| Model | Input | Output |
|---|---|---|
| Kimi K2 | $0.50β$0.60 | $2.40β$2.50 |
| Kimi K2 (cache hit) | $0.15 | $2.50 |
- Input: 500 tokens Γ $0.60/1M = $0.0003
- Output: 200 tokens Γ $2.50/1M = $0.0005
- Total: ~$0.0008/tick ($0.80 per 1000 ticks)
{
models: {
mode: "merge",
providers: {
moonshot: {
baseUrl: "https://api.moonshot.ai/v1",
apiKey: "${MOONSHOT_API_KEY}",
api: "openai-completions",
models: [{ id: "kimi-k2.5", name: "Kimi K2.5" }],
},
},
},
}- Sign up: https://open.bigmodel.cn or https://z.ai
- API Endpoint:
https://open.bigmodel.cn/api/paas/v4(or Z.AI endpoint) - Auth: Bearer token (
ZAI_API_KEY) - OpenAI-compatible: β Yes (also supports Anthropic format)
- Context: up to 200K tokens (GLM-4.7)
| Model | Input | Output |
|---|---|---|
| GLM-4.7 | $0.60 | $2.20 |
| GLM-4.7-FlashX | $0.07 | $0.40 |
| GLM-4.5-Air | $0.20 | $1.10 |
| GLM-4.7-Flash | FREE | FREE |
| GLM-4.5-Flash | FREE | FREE |
- GLM-4.7-Flash: $0.00/tick (FREE!) β Best for thinking ticks
- GLM-4.7-FlashX: ~$0.0001/tick
- GLM-4.7: ~$0.0007/tick
{
agents: { defaults: { model: { primary: "zai/glm-4.7-flash" } } },
}
// Just set env: ZAI_API_KEYβ RECOMMENDED for free thinking ticks β GLM-4.7-Flash and GLM-4.5-Flash are completely free.
- Sign up: https://dashscope.console.aliyun.com or https://qwen.ai/apiplatform
- API Endpoint:
https://dashscope.aliyuncs.com/compatible-mode/v1(OpenAI-compatible) - Auth: Bearer token (
DASHSCOPE_API_KEY) - OpenAI-compatible: β Yes β explicit OpenAI Chat Completion compatibility
- Also available via: OpenClaw
qwen-portalplugin (OAuth, free tier)
| Model | Input | Output |
|---|---|---|
| Qwen-Max | $0.38 | ~$1.20 |
| Qwen-Plus | ~$0.15 | ~$0.60 |
| Qwen-Turbo | ~$0.05 | ~$0.20 |
| Qwen3-32B (via Groq) | $0.29 | ~$0.39 |
- Qwen-Turbo: ~$0.0001/tick
- Qwen-Plus: ~$0.0002/tick
- Qwen-Max: ~$0.0004/tick
Option A β Built-in plugin (free OAuth tier):
openclaw plugins enable qwen-portal-auth
openclaw models auth login --provider qwen-portal --set-default
# Model: qwen-portal/coder-modelOption B β DashScope API key (custom provider):
{
models: {
mode: "merge",
providers: {
qwen: {
baseUrl: "https://dashscope.aliyuncs.com/compatible-mode/v1",
apiKey: "${DASHSCOPE_API_KEY}",
api: "openai-completions",
models: [{ id: "qwen-turbo", name: "Qwen Turbo" }],
},
},
},
}- Sign up: https://platform.deepseek.com
- API Endpoint:
https://api.deepseek.com/v1(OpenAI-compatible) - Auth: Bearer token (
DEEPSEEK_API_KEY) - OpenAI-compatible: β Yes
- Context: 64K tokens
| Model | Input (cache miss) | Input (cache hit) | Output |
|---|---|---|---|
| deepseek-chat (V3) | $0.27 | $0.07 | $1.10 |
| deepseek-reasoner (R1) | $0.55 | $0.14 | $2.19 |
- DeepSeek V3: 500Γ$0.27/1M + 200Γ$1.10/1M = ~$0.0004/tick
- DeepSeek V3 (cached): ~$0.0003/tick
- DeepSeek R1: ~$0.0007/tick (plus CoT tokens add cost)
{
models: {
mode: "merge",
providers: {
deepseek: {
baseUrl: "https://api.deepseek.com/v1",
apiKey: "${DEEPSEEK_API_KEY}",
api: "openai-completions",
models: [
{ id: "deepseek-chat", name: "DeepSeek V3" },
{ id: "deepseek-reasoner", name: "DeepSeek R1" },
],
},
},
},
}Also available via: huggingface/deepseek-ai/DeepSeek-R1 (built-in) or OpenRouter.
- Sign up: https://console.mistral.ai
- API Endpoint:
https://api.mistral.ai/v1(OpenAI-compatible) - Auth: Bearer token (
MISTRAL_API_KEY) - OpenAI-compatible: β Yes
- OpenClaw built-in: β
Yes (
mistralprovider)
| Model | Input | Output |
|---|---|---|
| Mistral Small | $1.00 | $3.00 |
| Mistral Large | $2.00 | $6.00 |
- Mistral Small: 500Γ$1.00/1M + 200Γ$3.00/1M = ~$0.0011/tick
- Mistral Large: ~$0.0022/tick
export MISTRAL_API_KEY="your-key"
# Model: mistral/mistral-small-latest or mistral/mistral-large-latest- Install: https://ollama.ai β
curl -fsSL https://ollama.ai/install.sh | sh - API Endpoint:
http://127.0.0.1:11434/v1(OpenAI-compatible) - Auth: None
- OpenAI-compatible: β Yes
- Cost: $0.00 (local compute only β electricity + hardware)
- RAM needed: ~5GB for Llama 3 8B (quantized)
ollama pull llama3:8b
# Or for newer: ollama pull llama3.3- $0.00/tick β Zero marginal cost
{
agents: { defaults: { model: { primary: "ollama/llama3:8b" } } },
}Our server (2-core Xeon, 8GB RAM) can run Llama 3 8B quantized (Q4_K_M) but expect ~2-5 tokens/sec. Adequate for background thinking ticks, not for interactive use.
| Provider | Cost/Tick | OpenAI Compat | Built-in OC | Quality | Latency |
|---|---|---|---|---|---|
| GLM-4.7-Flash | $0.000 | β | β | Medium | Fast |
| Ollama Llama3 8B | $0.000 | β | β | Low-Med | Slow (local) |
| Qwen-Turbo | $0.0001 | β | β plugin | Medium | Fast |
| DeepSeek V3 | $0.0004 | β | Via proxy | High | Fast |
| Kimi K2 | $0.0008 | β | Custom | High | Medium |
| Mistral Small | $0.0011 | β | β | Medium | Fast |
| GLM-4.7 | $0.0007 | β | β | High | Fast |
| DeepSeek R1 | $0.0007+ | β | Via proxy | Very High | Slow (CoT) |
| Mistral Large | $0.0022 | β | β | High | Fast |
Tier 1 β Free thinking ticks:
- GLM-4.7-Flash / GLM-4.5-Flash (Z.AI) β Free, cloud-hosted, built-in OpenClaw support. Best option.
- Ollama Llama 3 8B β Free, local, but slow on our hardware.
Tier 2 β Near-free (<$1/1000 ticks): 3. Qwen-Turbo β $0.10/1000 ticks, OpenClaw plugin available 4. DeepSeek V3 β $0.40/1000 ticks, excellent quality/price ratio
Tier 3 β Budget ($1-2/1000 ticks): 5. Kimi K2 β $0.80/1000 ticks, strong reasoning 6. Mistral Small β $1.10/1000 ticks, built-in support
All providers use OpenAI-compatible API format. Integration with OpenClaw is straightforward for all of them.