this post was submitted on 17 Apr 2026
58 points (98.3% liked)
Privacy
48080 readers
560 users here now
A place to discuss privacy and freedom in the digital world.
Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.
In this community everyone is welcome to post links and discuss topics related to privacy.
Some Rules
- Posting a link to a website containing tracking isn't great, if contents of the website are behind a paywall maybe copy them into the post
- Don't promote proprietary software
- Try to keep things on topic
- If you have a question, please try searching for previous discussions, maybe it has already been answered
- Reposts are fine, but should have at least a couple of weeks in between so that the post can reach a new audience
- Be nice :)
Related communities
much thanks to @gary_host_laptop for the logo design :)
founded 6 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Say I have a GPU with 32GB VRAM and I am on Linux, what local LLM would be good for coding?
Currently I just have an iGPU ;) but that's always an option, albeit a very expensive one.
Get llama.cpp and try Qwen3.6-35B-A3B. Just came out and looks good. You'll have to look into optimal settings, as it's a Mixture of Experts (MoE) model with only 3B parameters active. That means that the rest can stay in RAM for quick inference.
You could also try the dense model (Qwen3.5-27B), but that will be significantly slower. Put these in a coding harness like Oh-My-Pi, OpenCode, etc. and see how it fares for your tasks. Should be ok for small tasks, but don't expect Opus / Sonnet 4.6 quality, more like better than Haiku.
Sadly...none. Well, I mean...it depends what you mean by "coding". If you mean "replace Claude with local?". Then...none. Sorry.
If you mean "actually, if I use ECA to call a cloud model from OpenRouter for planning, then have it direct a local LLM to do the scutt work", then the Qwen series of models (like Qwen 3 Next) are pretty awesome.
The iGPU will make you want to kill yourself though. Get a GPU :) Even a 4-16GB one can make a difference.
PS: You said GPU and iGPU, so I'm not sure which one has the 32GB or what rig your running. I have suspicion though you're running on a i5 or i7 with something like a intel 630 igpu inbuilt? In which case, the iGPU is pretty slow and depending on the exact chip, you likely won't be able to use CUDA or Vulkan acceleration.
So, the "get a GPU" thing still holds :)
I meant that I can buy one of those Radeons dedicated to AI work, like the ASRock Radeon AI PRO R9700 Creator 32GB GDDR6. If I need to.
Currently my Ryzen iGPU is all I need, because all I need is to see the graphical desktop environment on my screen ;) It does the job well.
I use Claude Code as well and I am slightly concerned with that ID verification news, even more so because of the technology partner that they chose.
Hmm. The R9700 is RDNA4 - ROCm support for that architecture may be patchy in linux? Dunno. Check that before you commit your hard earned dollary-doos.
If all good
Qwen 3.6 is the latest hotness. I'd start from there and work backwards
https://inv.nadeko.net/embed/YKNvkBbRJIE?
https://www.youtube.com/watch?v=YKNvkBbRJIE