Gemma-4 26B-A4B + Opencode on M5 MacBook is *actually good*
TL;DR, 32gb M5 MacBook Air can run gemma-4-26B-A4B-it-UD-IQ4_XS at 300t/s PP and 12t/s generation (running in low power mode, uses 8W , making it the first laptop I've used to not get warm and noisy whilst running LLMs). Fast prompt processing + short thinking traces + can actually handle agentic behaviour = Opencode is actually usable from my laptop! -- Previously I've been running LLMs off my M1 Max 64gb. And whilst it's been good enough for tinkering and toy use cases, it's never really been great for running anything that requires longer context... i.e. it could be useful as a simple chatbot but not much else. Making a single Snake game in Python was fine, but anything where I might want to do agentic coding / contribute to a larger codebase has always been a bit janky. And unless I ar
Could not retrieve the full article text.
Read on Reddit r/LocalLLaMA →Reddit r/LocalLLaMA
https://www.reddit.com/r/LocalLLaMA/comments/1sbaack/gemma4_26ba4b_opencode_on_m5_macbook_is_actually/Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!