this post was submitted on 17 Apr 2026
60 points (98.4% liked)
Privacy
48109 readers
1394 users here now
A place to discuss privacy and freedom in the digital world.
Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.
In this community everyone is welcome to post links and discuss topics related to privacy.
Some Rules
- Posting a link to a website containing tracking isn't great, if contents of the website are behind a paywall maybe copy them into the post
- Don't promote proprietary software
- Try to keep things on topic
- If you have a question, please try searching for previous discussions, maybe it has already been answered
- Reposts are fine, but should have at least a couple of weeks in between so that the post can reach a new audience
- Be nice :)
Related communities
much thanks to @gary_host_laptop for the logo design :)
founded 6 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Personally, I would like to use AI, but I don't due to it being non local. I know there are local AI that could do things, but I don't know which models are the good one for each task. If someone can give me pointers for it, I'd be grateful, for exemple a good model for local coding :)
depends on your hardware and your preferred language. i think wizardcoder is a pretty common choice but the smallest useful version is around 14GB so you need the vram to accommodate it.
Thanks, I'll dig into this. BTW, I have a 9070 XT, with 16 Gb VRAM, so it should do the job I guess.
you need space for the context and runtime parameters too, but i think it should work. worst case there are some offloading settings you can do depending on the server you use. only way to knew is to try, really.
No point in telling you "yo, dude, just grab MinMax 2.7 or GLM5.1"...unless you happen to have several GPUs running concurrently with a total combined VRAM pool of 500GB or more.
There are strong local contenders... (Like Qwen3-Coder-Next but as you can see, the table ante is probably in the 45GB vram range just to load them up. Actually running them with a decent context length is likely to mean you need to be in the 80-100GB range.
Do-able...but maybe pay $10 on OpenRouter first to test drive them before committing to $2000+ worth of hardware upgrades.
There are other, more reasonable, less hardware dependent uses for local LLMs, but if you want fully local coders, it's the same old story: pay to play (and that's even if you don't mind slow speed / overnight batch jobs).
Right now, cloud-based providers are hemorrhaging money because they know it will lead to lock-in (ie: people will get use to what can be achieved with SOTA models, forgetting the multi-million dollar infrastructure required to run them). Then, when they realize you can't quite do the same with local gear (at least, without spending $$$), they can ratchet the prices up.
Codex pro-plan just went to $300/month.
We've seen this playbook before, right?
Thanks for the pointers. For the hardware, I have a 9070 XT with 16 Gb of VRAM. It's sure that it can be very expensive. As I only do this as a hobby, I don't want to pay that amount of money. I'm okay with having a slow llm as it wouldn't be a tool I'd use often. I prefer to try doing things on my own and use the ai to help for little tasks first, such as checking why this one line of code didn't want to work correctly or things like that.