Local Claude Code with Qwen3.5 27B
after long research, finding best alternative for Using a local LLM in OpenCode with llama.cpp to use totally local environment for coding tasks I found this article How to connect Claude Code CLI to a local llama.cpp server how to disable telemetry and make claude code totally offline. model used - Qwen3.5 27B Quant used - unsloth/UD-Q4_K_XL inference engine - llama.cpp Operating Systems - Arch Linux Hardware - Strix Halo I have separated my setups into sessions to run iterative cycle how I managed to improve CC (claude code) and llama.cpp model parameters. First Session as guide stated, I used option 1 to disable telemetry ~/.bashrc config; export ANTHROPIC_BASE_URL="http://127.0.0.1:8001" export ANTHROPIC_API_KEY="not-set" export ANTHROPIC_AUTH_TOKEN="not-set" export CLAUDE_CODE_DISABLE
Could not retrieve the full article text.
Read on Reddit r/LocalLLaMA →Reddit r/LocalLLaMA
https://www.reddit.com/r/LocalLLaMA/comments/1scrnzm/local_claude_code_with_qwen35_27b/Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.






Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!