Hacker Newsnew | past | comments | ask | show | jobs | submit | dial9-1's commentslogin

still waiting for the day I can comfortably run Claude Code with local llm's on MacOS with only 16gb of ram


My super uninformed theory is that local LLM will trail foundation models by about 2 years for practical use.

For example right now a lot of work is being done on improving tool calling and agentic workflows, which tool calling was first popping up around end of 2023 for local LLMs.

This is putting aside the standard benchmarks which get "benchmaxxed" by local LLMs and show impressive numbers, but when used with OpenCode rarely meet expectations. In theory Qwen3.5-397B-A17B should be nearly a Sonnet 4.6 model but it is not.


How close is this? It says it needs 32GB min?


You can run Qwen3.5-35B-A3B on 32GB of RAM sure, although to get 'Claude Code' performance, which I assume he means Sonnet or Opus level models in 2026, this will likely be a few years away before its runnable locally (with reasonable hardware).


I fully agree, I run that one with Q4 on my MBP, and the performance (including quality of response) is a let down.

I am wondering how people rave so much about local "small devices" LLM vs what codex or Claude code are capable of.

Sadly there are too much hype on local LLM, they look great for 5min tests and that's it.


Just train it better with AGENTS.md


I'm reading "more than 32GB of unified memory" to mean at least a 36 GB model.


Doesn't OpenCode supports local models?


You can, but the quality sucks.

Local LLMs don't make sense for most people compared to "cloud" services, even more so for coding.


the OP probably is not telling the whole story and must have some kind of drug addiction going on that sucks up all of his wealth, because how do you even end up on a van when you split rent with a girlfriend/roommates? also the part where he refuses non-vegan food. yikes. and skateboarding as a hobby sounds great when you are uninsured


> and skateboarding as a hobby sounds great when you are uninsured

OK, THAT is clearly a failure of the system. Broken and sick people deserve top-quality medical care; it's what a high-functioning society would do.


scale of complexity


That's a tempting answer. I see why you proffer it. But I have to say no.

Complexity is neither an immanent feature nor inevitability. Behind unruly complexity is our failure to manage it. And indeed, a love of complexity, a fetish for it that seduces us into ever more.

To defeat complexity we have to embrace, and engage with it. We have to see what parts of technology that got us to where we are, must now be justifiably rejected.

All I see right now, especially with regards to "AI" and the new wave of techno-populism, is a retreat from complexity and more embrace of "magic".


modules just do it better


buy an ad


I don’t own lunch money so it’s just a user’s opinion. But your cynicism has added real value to this thread. Thanks!


simply because your wage will not keep up with inflation and would force most people to take a second job


it's catalá in catalá


It's actually "català" not "catalá". Source: I'm català.


diversity is good


you can accomplish that with just modules and functions


I'd say that all that (modern style) OOP does for organizing code comes from its copying earlier module systems. There is really nothing else there.


so it's just the "drag & drop .jar file into production" reinvented


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: