I’m a big proponent of self-hosting, right to repair, and rolling your own whatever when you can. That probably started as teenage rebellion that got baked in - I was lucky enough to read both Walden and The Hobbit during a week-long cyclone lockdown several decades ago - but I suspect there’s a non-trivial overlap between that space and privacy-minded people in general.

My endgame is a self-sufficient intranet for myself and family: if the net goes down tomorrow, we’d barely notice.

I also use LLMs as a tool. True self-hosted equivalence to state-of-the-art models is still an expensive proposition, so like many, I use cloud-based tools like Claude or Codex for domain-specific heavy lifting - mostly coding. Not apologising for it; I think it’s a reasonable trade-off while local hardware catches up.

That context is just to establish where I’m coming from when I say this caught my attention today:

https://support.claude.com/en/articles/14328960-identity-verification-on-claude

To be accurate about what it actually says: this isn’t a blanket “show us your passport to use Claude.” Not yet.

The policy as written is narrower than it might first appear.

My concern isn’t what it says - it’s that the precedent now exists. OAI will do doubt follow suite.

Scope creep is a documented pattern with this kind of thing, and “we only use it for X” describes current intent, not a structural constraint.

Given the nature of this community, figured it was worth flagging.

  • steel_for_humans@piefed.social
    link
    fedilink
    English
    arrow-up
    2
    ·
    11 hours ago

    Say I have a GPU with 32GB VRAM and I am on Linux, what local LLM would be good for coding?

    Currently I just have an iGPU ;) but that’s always an option, albeit a very expensive one.

    • andrew0@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      2
      ·
      10 hours ago

      Get llama.cpp and try Qwen3.6-35B-A3B. Just came out and looks good. You’ll have to look into optimal settings, as it’s a Mixture of Experts (MoE) model with only 3B parameters active. That means that the rest can stay in RAM for quick inference.

      You could also try the dense model (Qwen3.5-27B), but that will be significantly slower. Put these in a coding harness like Oh-My-Pi, OpenCode, etc. and see how it fares for your tasks. Should be ok for small tasks, but don’t expect Opus / Sonnet 4.6 quality, more like better than Haiku.

    • SuspciousCarrot78@lemmy.worldOP
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      10 hours ago

      Sadly…none. Well, I mean…it depends what you mean by “coding”. If you mean “replace Claude with local?”. Then…none. Sorry.

      If you mean “actually, if I use ECA to call a cloud model from OpenRouter for planning, then have it direct a local LLM to do the scutt work”, then the Qwen series of models (like Qwen 3 Next) are pretty awesome.

      The iGPU will make you want to kill yourself though. Get a GPU :) Even a 4-16GB one can make a difference.

      PS: You said GPU and iGPU, so I’m not sure which one has the 32GB or what rig your running. I have suspicion though you’re running on a i5 or i7 with something like a intel 630 igpu inbuilt? In which case, the iGPU is pretty slow and depending on the exact chip, you likely won’t be able to use CUDA or Vulkan acceleration.

      So, the “get a GPU” thing still holds :)

      • steel_for_humans@piefed.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 hours ago

        I meant that I can buy one of those Radeons dedicated to AI work, like the ASRock Radeon AI PRO R9700 Creator 32GB GDDR6. If I need to.

        Currently my Ryzen iGPU is all I need, because all I need is to see the graphical desktop environment on my screen ;) It does the job well.

        I use Claude Code as well and I am slightly concerned with that ID verification news, even more so because of the technology partner that they chose.