Skip to main content


Claude: papers please?


This entry was edited (2 weeks ago)
in reply to SuspciousCarrot78

I hope people in the us wake the fuck up and these companies start losing users to foreign ones by the millions
in reply to Eager Eagle

We spread word to migrate to federated and non US tech platforms to start taking their financial leg, then ad revenue drops and slowly they weaken. The weaker they are, the less political influence (aka purchasing power) they have!
in reply to Eager Eagle

I'm not sure that solves the issue or just changes the actors. Still, I'm all for "fight the power".

I'm just a silly man with a box of scraps. But I hope enough silly men with boxes can come together to form some sort of co-op. Maybe. I don't know. But...I hope, people smarter and better resourced than I can find a way forward.

The writing is on the wall here.

in reply to SuspciousCarrot78

Yucky.. The identity partner looks nasty. I bet all these corps are linked to the Epstein class and their think tanks. The 1% are taking control of their cattle..
in reply to SuspciousCarrot78

Even if the tools are not yet there, “they” want to know exactly who asks for code to things like a DIY radar station or autonomous drone control. We’re well into “first they came” territory.
in reply to trailee

I hope you're wrong. I'm worried that you're probably not.

Still time. Just barely.

This entry was edited (2 weeks ago)
in reply to SuspciousCarrot78

IMHO LLM usage isn't coherent with independence. That being said I wrote quite a bit on self-hosting LLMs. There are quite a few tools available, like ollama itself relying on llama.cpp that can both work locally and provide an API compatible replacement to cloud services. As you suggested though typically at home one doesn't have the hardware, GPUs with 100+GB of VRAM, to run the state of the art. There is a middle ground though between full cloud, API key, closed source vs open source at home on low-end hardware : running STOA open models on cloud. It can be done on any cloud but it's much easier to start with dedicated hardware and tooling, for that HuggingFace is great but there are multiples.

TL;DR: closed cloud -> models on clouds -> self-hosted provide a better path to independence, including training.

This entry was edited (2 weeks ago)
in reply to utopiah

Yeah, me too :)

bobbyllm.github.io/llama-condu…

codeberg.org/BobbyLLM/llama-co…

I'm thinking about coding a >>cloud side car at the moment, with the exact feature you mentioned...but...that's scope creep for what I have in mind.

Irrespective of all that, I agree: an open cloud co-op could be a good way to have SOTA (or near SOTA - GLM 5.1 is about as close as we have right now) access for when needed.

(Not teaching you to suck eggs, so this comment is for the lay-reader):

For coding, you can do some interesting stuff where the cloud model is the "general" and the locally hosted LLM is the "soldier" that does the grunt work. We have some pretty decent, consumer-level-hardware runnable "soldiers" now (I still like Qwen 3 coder)...they just don't quite have the brains to see the full/big picture for coding.

This entry was edited (2 weeks ago)
in reply to SuspciousCarrot78

Say I have a GPU with 32GB VRAM and I am on Linux, what local LLM would be good for coding?

Currently I just have an iGPU ;) but that's always an option, albeit a very expensive one.

in reply to steel_for_humans

Sadly...none. Well, I mean...it depends what you mean by "coding". If you mean "replace Claude with local?". Then...none. Sorry.

If you mean "actually, if I use ECA to call a cloud model from OpenRouter for planning, then have it direct a local LLM to do the scutt work", then the Qwen series of models (like Qwen 3 Next) are pretty awesome.

The iGPU will make you want to kill yourself though. Get a GPU :) Even a 4-16GB one can make a difference.

PS: You said GPU and iGPU, so I'm not sure which one has the 32GB or what rig your running. I have suspicion though you're running on a i5 or i7 with something like a intel 630 igpu inbuilt? In which case, the iGPU is pretty slow and depending on the exact chip, you likely won't be able to use CUDA or Vulkan acceleration.

So, the "get a GPU" thing still holds :)

This entry was edited (2 weeks ago)
in reply to SuspciousCarrot78

I meant that I can buy one of those Radeons dedicated to AI work, like the ASRock Radeon AI PRO R9700 Creator 32GB GDDR6. If I need to.

Currently my Ryzen iGPU is all I need, because all I need is to see the graphical desktop environment on my screen ;) It does the job well.

I use Claude Code as well and I am slightly concerned with that ID verification news, even more so because of the technology partner that they chose.

in reply to steel_for_humans

Hmm. The R9700 is RDNA4 - ROCm support for that architecture may be patchy in linux? Dunno. Check that before you commit your hard earned dollary-doos.

If all good

  • Qwen2.5-Coder-32B fits comfortably and is genuinely capable.
  • Qwen3.5-27B (dense)
  • Qwen3.5-35B-A3B (MoE, only 3B active parameters)
  • Qwen3.6-35B-A3B just dropped

Qwen 3.6 is the latest hotness. I'd start from there and work backwards

inv.nadeko.net/embed/YKNvkBbRJ…

in reply to steel_for_humans

Get llama.cpp and try Qwen3.6-35B-A3B. Just came out and looks good. You'll have to look into optimal settings, as it's a Mixture of Experts (MoE) model with only 3B parameters active. That means that the rest can stay in RAM for quick inference.

You could also try the dense model (Qwen3.5-27B), but that will be significantly slower. Put these in a coding harness like Oh-My-Pi, OpenCode, etc. and see how it fares for your tasks. Should be ok for small tasks, but don't expect Opus / Sonnet 4.6 quality, more like better than Haiku.

in reply to SuspciousCarrot78

Personally, I would like to use AI, but I don't due to it being non local. I know there are local AI that could do things, but I don't know which models are the good one for each task. If someone can give me pointers for it, I'd be grateful, for exemple a good model for local coding :)
in reply to lsjw96kxs

This entry was edited (2 weeks ago)
in reply to SuspciousCarrot78

Thanks for the pointers. For the hardware, I have a 9070 XT with 16 Gb of VRAM. It's sure that it can be very expensive. As I only do this as a hobby, I don't want to pay that amount of money. I'm okay with having a slow llm as it wouldn't be a tool I'd use often. I prefer to try doing things on my own and use the ai to help for little tasks first, such as checking why this one line of code didn't want to work correctly or things like that.
in reply to lsjw96kxs

depends on your hardware and your preferred language. i think wizardcoder is a pretty common choice but the smallest useful version is around 14GB so you need the vram to accommodate it.
in reply to lime!

Thanks, I'll dig into this. BTW, I have a 9070 XT, with 16 Gb VRAM, so it should do the job I guess.
in reply to lsjw96kxs

you need space for the context and runtime parameters too, but i think it should work. worst case there are some offloading settings you can do depending on the server you use. only way to knew is to try, really.
in reply to SuspciousCarrot78

I love the sound of this but can I ask, if the net goes down and you hardly notice, where do you get your 'net' from? Or is it that your intranet doesn't need internet as such and everything is just local?

I might have answered my own question there but I'm interested to understand it a bit more.

Thanks!

in reply to ropatrick

This entry was edited (1 week ago)
in reply to SuspciousCarrot78

OK I have you. You dont need the internet because you have the internet in your terabyte farm. Pretty cool.

Thanks for the detailed reply.

One final question, I'm sure its dark at the bottom of the deep rabbit hole you are in, what do you do for batteries for your head torch?! 😀

in reply to ropatrick

Exactly so. Mom - can we get the internet? Mom: we have the internet at home.

Batteries? I don't need batteries. I have the never-ending warm glow of weaponized autism. And that's not even a joke.

I tend to hyper-fixate on something until either it breaks or I do. It's usually 70/30 in my favour :)

This entry was edited (1 week ago)
in reply to SuspciousCarrot78

Thats weird, i don't remember having an alt account called SuspiciousCarrot78 but surely you must be me, same project, same neurology.. same fixation pattern.
This entry was edited (1 week ago)
in reply to webghost0101

Well, there's a quick check. Take a shot of whisky and I'll see if can type "gottle of geer".
in reply to SuspciousCarrot78

Yep, definitely talking to myself again

Jokes aside, i haven’t seen you mention anything for media streaming.

I highly passionately recommend Navidrome for music. It is my absolute favorite and most used self hosted service.

For acquiring media like film and music depending where you live ripping those from your local library is in some places arguably a protected fair use. (Comes from the time mp3
players became common and runners used to take rented cds in their walkman outside before). In my experience, 480p dvd is much higher quality then internet 480p streams and the total size is much smaller then what you find in downloads.

ARM can help you automatically rip these as long as you have a drive in your pc. I got ARM running in a proxmox lxc with drive passtrough but that honestly was a pain to setup so not sure you should go that exact route, either way the moment arm is functional its smooth sailing and your only concern becomes storage space.

This entry was edited (1 week ago)
in reply to webghost0101

This entry was edited (1 week ago)
in reply to SuspciousCarrot78

in reply to ropatrick

Here. Begin your walk down the path to a power some would call...unnatural

get.kiwix.org/en/solutions/app…

get.kiwix.org/en/solutions/hot…

get.kiwix.org/en/solutions/hot…

This entry was edited (1 week ago)
in reply to SuspciousCarrot78

in reply to maegul (he/they)

This entry was edited (2 weeks ago)
in reply to SuspciousCarrot78

Oh I hear you (and appreciate the response).

For me, I can’t help but think of another alternative, which I’m surprised I haven’t heard of yet …

stripping down one’s personal technological cognitive load to a stack of systems that can fit into one’s brain (like the Python mantra), focusing on learning that stack well building sustainable and stable systems, and then just detoxing from the increasingly polluted digital information stream (protected commons, traditional formats such as books and in person engagement … dunno).

Depends on what the end goal is, but AIs seem to be about using tech more or just opting out of sovereignty. Something like the above seems to me to be about using tech less (in the end) and pushing toward being a secondary tool rather than an end of its own.

in reply to maegul (he/they)

I agree.

God help me, I'm actually reading books again.

Books.

It's...harder than it use to be. A lot harder, actually.

But there's something to be said about marginalia etc.

This entry was edited (2 weeks ago)
in reply to SuspciousCarrot78

Ha yes … on the other hand, it was easy to forget how good damn expansive non-internet information is: the whole world ran on that shit for millennia.
in reply to SuspciousCarrot78

Being responsible with powerful technology starts with knowing who is using it.


The fuck it does. Claude is already censored - you can't get a recipe of a poison, schematics for a bomb, an advice on how to hide a body. If you can, then it's Anthropic engineers didn't do their job.

Knowing who is using it helps either with conditional censorship, or helping governments to track people based on their prompts, or just plainly lying and using data for analytics and training. All these easons are shit.

And don't tell me this is to protect the kids again. Let the parents do their job.