Price Per TokenPrice Per Token
OpenClaw

Best Local & Open-Source Models for OpenClaw (2026)

Community-voted LLM rankings for running OpenClaw with local and open-source models

Discussion · OpenClaw
3
ellmanalex·30d ago

After Claude ban I found my new main model

I've been using OpenClaw for months with only Opus 4.6, Sonnet 4.6, and GPT 5.3/5.4. I'm the kind of person who needs the flagship model as long as budgets are reasonable. Claude is dead. OpenAI made business plan quotas unusable. So I went shopping for alternatives. GLM 5.1 and 5 Turbo: absolute garbage for agentic tasks and automation. Couldn't even write a simple Reddit reply without flooding Telegram with code dumps. Felt like talking to a drunk model. Cancelled. (They said "we'll refund" — still waiting 3 weeks later.) MiMo V2 Pro: using it since launch, really liked it. Honestly got Opus/GPT vibes in many ways. After Claude banned OpenClaw yesterday I got the Token Plan (standard $16). Terrible credit system. Everything in OpenClaw deducts from credits. Session history, bootstrap MD content, tool outputs, cache — literally everything. One month's quota gone in 1 day after filling just 2 session contexts. Horribly inefficient. I will never pay again until they fix the credit logic. Kimi: reviews were bad, never tried. Grok: community feedback looked bad, skipped it. Gemini: no monthly payment option. If there was I'd probably use it, but it's too expensive. So I went with the most popular alternative: Minimax 2.7. When MiMo and GPT failed to handle my nit cron task and Minimax M2.7 solved it in 5 minutes. And the quota on Minimax is impossible to exhaust. — I was shocked. How are they this generous? If anyone knows please explain because it really feels like it won't run out. Tested browser automations. It's not as smart as Opus, but for my automation tasks, light coding work, and being a personal agent — it's enough. If it falls short, I might rotate between a few GPT Plus subscriptions for GPT 5.4 access. Right now, price/performance-wise, Minimax and GPT Plus x 2 accounts are the only efficient OpenClaw model options I could find.

6 comments·Join the discussion →
5 Tips to Save Money on OpenClaw
Provider
Model
Input $/M
Output $/M
Vote
Score
$0.140
$0.420
71+6
$0.140
$0.200
40+4
$0.500
$0.500
20+2
$0.200
$0.200
10+1
$0.000
$0.000
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.010
$0.020
000
$0.010
$0.020
000
$0.100
$0.100
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.100
$0.200
000
-
$0.000
$0.000
000
$0.800
$1.200
000
$0.037
$0.150
000
$0.150
$0.150
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.100
$0.100
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.100
$0.100
000
$0.200
$0.200
000
$0.170
$0.170
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.100
$0.100
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.045
$0.150
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.180
$0.180
000
$0.100
$0.100
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.100
$0.100
000
$0.100
$0.100
000
$0.000
$0.000
000
$0.150
$0.150
000
$0.100
$0.100
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.100
$0.100
000
$0.100
$0.100
000
$0.050
$0.080
000
$0.200
$0.200
000
$0.500
$0.500
000
$0.170
$0.170
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.500
$0.500
000
$0.070
$0.070
000
$0.200
$0.200
000
$0.240
$0.240
000
$0.200
$0.200
000
$0.300
$0.300
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.040
$0.130
000
$0.000
$0.000
000
$0.200
$0.200
000
$0.300
$0.300
000
$0.100
$0.100
000
IB
Ibm
$0.000
$0.000
000
$0.300
$0.300
000
$0.060
$0.060
000
$0.100
$0.100
000
$0.100
$0.100
000
$0.400
$0.400
000
$0.100
$0.100
000
$0.040
$0.040
000
$0.040
$0.100
000
$0.170
$0.430
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.080
$0.200
000
$0.000
$0.000
000
$0.100
$0.100
000
$0.200
$0.200
000
$0.000
$0.000
000
$0.040
$0.050
000
$0.000
$0.000
000
$0.140
$0.140
000
$0.000
$0.000
000
$0.059
$0.059
000
$0.140
$0.200
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.140
$0.420
000
$0.030
$0.040
000
$0.300
$0.300
000
$0.060
$0.060
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.020
$0.020
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.000
$0.000
000
$0.200
$0.200
000
$0.200
$0.200
000
$0.060
$0.120
01-1
$0.100
$0.150
01-1
$0.020
$0.050
01-1

Vote for open-source models that work well (or don't) with OpenClaw.

Pricing from OpenRouter.

Running OpenClaw with a Local Model or LLM

OpenClaw supports local models via OpenAI-compatible APIs. The easiest way to run a model locally is with Ollama or llama.cpp, both of which expose a local API endpoint that OpenClaw can connect to.

1

VRAM Requirements

7B models need ~6GB VRAM (4-bit quantized). 13B models need ~10GB. 70B models need ~40GB or multi-GPU. If you're on CPU-only, smaller quantized models (Q4) are usable but slow.

2

Recommended: Ollama

Install Ollama, run ollama pull llama3.3, then point OpenClaw to http://localhost:11434/v1.

3

Alternative: llama.cpp server

For more control over quantization and performance, llama.cpp's server mode gives you a full OpenAI-compatible API with fine-grained settings.

Compare all local LLM runners →

About This Leaderboard

This leaderboard shows community votes specifically for open-source and locally-runnable models and LLMs used with OpenClaw. Models are filtered to include those from open-source providers: Meta (Llama), Mistral, Qwen (Alibaba), Google (Gemma), DeepSeek, Microsoft (Phi), and similar open-weight families.

Running local models with OpenClaw gives you full privacy, zero API costs, and offline capability — at the cost of needing hardware and accepting some quality trade-offs versus frontier API models.

Frequently Asked Questions

Based on community votes, Llama 3.3 70B is the top-rated open-source model for OpenClaw. It offers strong instruction-following and tool use while being free to run locally with sufficient VRAM.
Yes. OpenClaw supports any OpenAI-compatible API endpoint, so you can point it at a local Ollama or llama.cpp server running any open-source model. Performance depends on model size and your hardware.
For 7B models you need ~6GB VRAM (4-bit). For 70B models like Llama 3.3 70B, plan for 40GB+ or multiple GPUs. CPU-only inference is possible but very slow for agentic coding tasks.
Models with strong instruction-following and tool use matter most for agentic tasks. Llama 3.3 70B, DeepSeek Coder V2, and Qwen2.5 Coder consistently rank well for agentic coding workflows.