Show HN: First Claude Code client for Ollama local models
44 points by SerafimKorablev
by oceanplexian
2 subcomments
The Anthropic API was already supported by llama.cpp (The project Ollama ripped off and typically lags in features by 3-6 months), which already works perfectly fine with Claude Code by setting a simple environment variable.
by d4rkp4ttern
0 subcomment
As others said this was possible for months already with llama-cop’s support for Anthropic messages API. You just need to set the ANTHROPIC_BASE_URL. The specific llama-server settings/flags were a pain to figure out and required some hunting, so I collected them in this guide to using CC with local models:
One tricky thing that took me a whole day to figure out is that using Claude Code in this setup was causing total network failures due to telemetry pings, so I had to set this env var to 1: CLAUDE_CODE_DISABLE_NONESSENTIAL_TRAFFIC
by eli
1 subcomments
There are already various proxies to translate between OpenAI-style models (local or otherwise) and an Anthropic endpoint that Claude Code can talk to. Is the advantage here just one less piece of infrastructure to worry about?
by dsrtslnd23
2 subcomments
What hardware are you running the 30b model on? I guess it needs at least 24GB VRAM for decent inference speeds.
by horacemorace
2 subcomments
I was trying to get Claude code to work with llama.cpp but could never figure out anything functional. It always insisted on a phone home login for first time setup. In cline I’m getting better results with glm-4.7-flash than with qwen3-coder:30b
by dosinga
0 subcomment
this is cool. not sure it is the first claude code style coding agent that runs against Ollama models though. goose, opencode and others have been able to do that a while no?
by d0100
1 subcomments
Does this UI work with Open Code?
by mchiang
1 subcomments
hey, thanks for sharing. I had to go to the Twitter feed to find the GitHub link: