I’m a big proponent of self-hosting, right to repair, and rolling your own whatever when you can. That probably started as teenage rebellion that got baked in - I was lucky enough to read both Walden and The Hobbit during a week-long cyclone lockdown several decades ago - but I suspect there’s a non-trivial overlap between that space and privacy-minded people in general.

My endgame is a self-sufficient intranet for myself and family: if the net goes down tomorrow, we’d barely notice.

I also use LLMs as a tool. True self-hosted equivalence to state-of-the-art models is still an expensive proposition, so like many, I use cloud-based tools like Claude or Codex for domain-specific heavy lifting - mostly coding. Not apologising for it; I think it’s a reasonable trade-off while local hardware catches up.

That context is just to establish where I’m coming from when I say this caught my attention today:

https://support.claude.com/en/articles/14328960-identity-verification-on-claude

To be accurate about what it actually says: this isn’t a blanket “show us your passport to use Claude.” Not yet.

The policy as written is narrower than it might first appear.

My concern isn’t what it says - it’s that the precedent now exists. OAI will do doubt follow suite.

Scope creep is a documented pattern with this kind of thing, and “we only use it for X” describes current intent, not a structural constraint.

Given the nature of this community, figured it was worth flagging.

  • pound_heap@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    4
    ·
    14 hours ago

    Being responsible with powerful technology starts with knowing who is using it.

    The fuck it does. Claude is already censored - you can’t get a recipe of a poison, schematics for a bomb, an advice on how to hide a body. If you can, then it’s Anthropic engineers didn’t do their job.

    Knowing who is using it helps either with conditional censorship, or helping governments to track people based on their prompts, or just plainly lying and using data for analytics and training. All these easons are shit.

    And don’t tell me this is to protect the kids again. Let the parents do their job.

  • maegul (he/they)@lemmy.ml
    link
    fedilink
    English
    arrow-up
    14
    ·
    24 hours ago

    Probably a shallow response …

    But I always figured AI/LLMs are basically apocalyptic for all sorts of individualistic values in computing (including privacy but also independence and diversity).

    Whether they’re good or useful etc, I just struggle to see how they will ever be justifiable against these sorts of values.

    Sure, local models and our hardware will get better … but better than the state of the art from the big labs and providers? Given that data and training are the big bottlenecks on quality … I struggle to see how AI isn’t a complete feudal capturing of information computing and processing. Not to mention what happens to the pipeline that produces information content if everyone is only consuming it through the models that train on it.

    So for me the big question is, what’s our call on a possible (likely even?) future where we are forever stuck using cloud provided AI along with all of its negatives, in the same way that basically all of us has been and still is stuck using MS windows, Google and the big-social-media hellscape?

    For me, I baulk at this.

    • ☂️-@lemmy.ml
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      12 hours ago

      these people are losing a lot of money on every prompt with the state of the art models.

      on this specific front it’s only a matter of time.

    • SuspciousCarrot78@lemmy.worldOP
      link
      fedilink
      arrow-up
      7
      ·
      23 hours ago

      I’m right there with you…but may I offer an alternative narrative in two parts and then address the pipeline issue you raise.

      The first part:

      There’s a small (but real) subset of people turning their back on big corpo. Retro-tech, dumb-phones, self hosting, linux, right-to-repair advocates, OSS and FOSS, privacy groups … everyone can smell the enshittification and are (in their own ways) pushing back. That’s not nothing.

      I think the way forward is not to play the game. Big corpo will do what big corpo always does. But we can use the tools we have to make the things we want.

      Will it compete with SOTA? No. But…does it need to? At an individual level, I’d argue “probably not”. It just needs to work for the individual.

      More to the point, there’s something to be said about doing more with less. Constraints can bring about real innovation. If the answer cannot be “Throw more X at it” (where X is $$$, compute, whatever)…then how can you leverage the tools and intelligence you have to build what you want? I think that’s the real question.

      Now for the second part:

      So for me the big question is, what’s our call on a possible (likely even?) future where we are forever stuck using cloud provided AI along with all of its negatives, in the same way > that basically all of us has been and still is stuck using MS windows, Google and the big-social-media hellscape?

      I’m more sanguine about it because I think this is down to the individual. Look at where you are now - it’s not Reddit or Facebook :). You and I choose to be here because…reasons. We can choose to run Linux, LibreOffice, Mullivad, llama.cpp, SearXNG, Syncthing, Immich etc for the same reasons.

      I think the trick will be figuring out how to navigate from your home ecosystem into the wider world, without getting f’d in the a.

      The one thing I don’t have a clean answer for is your pipeline point. If the content web collapses into AI slop - and it’s already going that way - then the human-generated signal that makes these models worth using starts to degrade. You may need to hold onto your “Good Old LLMs” for a while yet (or start training your own from scratch. There are ways and means but that’s beyond the scope of this conversation I think).

      In any case, individual sovereignty doesn’t fix that. You can opt out personally and still live in a world where the epistemic commons has been strip-mined.

      That…probably what WILL happen, come to think of it. Ok, fine. But partial answers already exist - cryptographic provenance of human content, federated communities being structurally harder to slop-flood (maybe).

      Honestly? Nobody has solved that problem just yet. The people building the biggest models know it’s a problem and don’t have a clean answer either. Anyone who says they do is selling something.

      All I can say is the only way to win is not to play the game. Which WORP would no doubt meep-morp at.

      • maegul (he/they)@lemmy.ml
        link
        fedilink
        English
        arrow-up
        4
        ·
        22 hours ago

        Oh I hear you (and appreciate the response).

        For me, I can’t help but think of another alternative, which I’m surprised I haven’t heard of yet …

        stripping down one’s personal technological cognitive load to a stack of systems that can fit into one’s brain (like the Python mantra), focusing on learning that stack well building sustainable and stable systems, and then just detoxing from the increasingly polluted digital information stream (protected commons, traditional formats such as books and in person engagement … dunno).

        Depends on what the end goal is, but AIs seem to be about using tech more or just opting out of sovereignty. Something like the above seems to me to be about using tech less (in the end) and pushing toward being a secondary tool rather than an end of its own.

        • SuspciousCarrot78@lemmy.worldOP
          link
          fedilink
          arrow-up
          6
          ·
          22 hours ago

          I agree.

          God help me, I’m actually reading books again.

          Books.

          It’s…harder than it use to be. A lot harder, actually.

          But there’s something to be said about marginalia etc.

          • maegul (he/they)@lemmy.ml
            link
            fedilink
            English
            arrow-up
            3
            ·
            22 hours ago

            Ha yes … on the other hand, it was easy to forget how good damn expansive non-internet information is: the whole world ran on that shit for millennia.

  • Eager Eagle@lemmy.world
    link
    fedilink
    English
    arrow-up
    21
    arrow-down
    1
    ·
    1 day ago

    I hope people in the us wake the fuck up and these companies start losing users to foreign ones by the millions

    • unitedwithme@lemmy.today
      link
      fedilink
      arrow-up
      6
      ·
      1 day ago

      We spread word to migrate to federated and non US tech platforms to start taking their financial leg, then ad revenue drops and slowly they weaken. The weaker they are, the less political influence (aka purchasing power) they have!

    • SuspciousCarrot78@lemmy.worldOP
      link
      fedilink
      arrow-up
      5
      ·
      1 day ago

      I’m not sure that solves the issue or just changes the actors. Still, I’m all for “fight the power”.

      I’m just a silly man with a box of scraps. But I hope enough silly men with boxes can come together to form some sort of co-op. Maybe. I don’t know. But…I hope, people smarter and better resourced than I can find a way forward.

      The writing is on the wall here.

  • utopiah@lemmy.ml
    link
    fedilink
    arrow-up
    9
    ·
    1 day ago

    IMHO LLM usage isn’t coherent with independence. That being said I wrote quite a bit on self-hosting LLMs. There are quite a few tools available, like ollama itself relying on llama.cpp that can both work locally and provide an API compatible replacement to cloud services. As you suggested though typically at home one doesn’t have the hardware, GPUs with 100+GB of VRAM, to run the state of the art. There is a middle ground though between full cloud, API key, closed source vs open source at home on low-end hardware : running STOA open models on cloud. It can be done on any cloud but it’s much easier to start with dedicated hardware and tooling, for that HuggingFace is great but there are multiples.

    TL;DR: closed cloud -> models on clouds -> self-hosted provide a better path to independence, including training.

    • SuspciousCarrot78@lemmy.worldOP
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      1 day ago

      Yeah, me too :)

      https://bobbyllm.github.io/llama-conductor/

      https://codeberg.org/BobbyLLM/llama-conductor

      I’m thinking about coding a >>cloud side car at the moment, with the exact feature you mentioned…but…that’s scope creep for what I have in mind.

      Irrespective of all that, I agree: an open cloud co-op could be a good way to have SOTA (or near SOTA - GLM 5.1 is about as close as we have right now) access for when needed.

      (Not teaching you to suck eggs, so this comment is for the lay-reader):

      For coding, you can do some interesting stuff where the cloud model is the “general” and the locally hosted LLM is the “soldier” that does the grunt work. We have some pretty decent, consumer-level-hardware runnable “soldiers” now (I still like Qwen 3 coder)…they just don’t quite have the brains to see the full/big picture for coding.

  • ropatrick@lemmy.world
    link
    fedilink
    arrow-up
    3
    ·
    1 day ago

    I love the sound of this but can I ask, if the net goes down and you hardly notice, where do you get your ‘net’ from? Or is it that your intranet doesn’t need internet as such and everything is just local?

    I might have answered my own question there but I’m interested to understand it a bit more.

    Thanks!

    • SuspciousCarrot78@lemmy.worldOP
      link
      fedilink
      arrow-up
      6
      ·
      edit-2
      12 hours ago

      The intranet becomes the internet :) Everything is local, accessible from multiple devices within my wLAN. The main box plugs into the router and serves everything over Wifi to trusted devices - my documents, media, books, games etc.

      I wrote (flippantly) about the bones of the system here, 3 or 4 months ago. It’s more complex now, but the endgame has always been “what if cloud, but you are your own cloud?”

      https://lemmy.world/post/41315607/21438607

      It may not be fresh (if the net goes down), but it would be local. The only real question I have to grok for myself is if I want to mirror curated section of Wikipedia, books etc.

      https://en.wikipedia.org/wiki/Kiwix

      Probably I should. May as well go full data-horder. Good excuse to get a few more TB of storage. What I’ve done so far is all within 4TB, using clever tricks and black magic but there’s a limit to 4TB. Fortunately, hard drives are still not too $$$. +4TB is about $200 here locally. So the entire set up is still around $600-700 AUD (around $350-400 USD)

      All the other stuff I have more or less tee-ed up (barring the UPS + solar kit I am building later in the year).

      https://www.youtube.com/watch?v=1q4dUt1yK0g

      Anyhow, 8TB local should just about cover what I have in mind (he said, fully aware of dragon horder sickness). Then I’ll grab something for offsite storage for critical docs - I have an old raspberry pi with a 256GB NVMe ssd I can use for that.

      I’m semi tempted (because fuck it, why not have fun) to look into LoRA after that.

      https://en.wikipedia.org/wiki/LoRa

      What I really NEED to do is finish the LLM stack (I’m on it and nearly done) and then do a curated Youtube replacement with yl-dlp feeding into Nova-player or Jellyfin, once/if SmartTube etc gets shit-canned. The youtube thing I’m kinda excited about because I’ve figured out how to squeeze ~1500 videos in around 250ish GB of storage, with TTL (time to live) mechanics, download replacement schedule etc. The kids watch too much random shit on YT, so daddy will make YT at home (ha!).

      I have some other wild ideas too…it’s a whole other thing…don’t get me started :)

      Once I’m finished, I will open-source the entire thing, post about it here, and let others replicate / improve on it. And so it goes. Once you begin walking down the dark path, you are forever doomed. Be careful :)

      • ropatrick@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        23 hours ago

        OK I have you. You dont need the internet because you have the internet in your terabyte farm. Pretty cool.

        Thanks for the detailed reply.

        One final question, I’m sure its dark at the bottom of the deep rabbit hole you are in, what do you do for batteries for your head torch?! 😀

        • SuspciousCarrot78@lemmy.worldOP
          link
          fedilink
          arrow-up
          5
          ·
          edit-2
          20 hours ago

          Exactly so. Mom - can we get the internet? Mom: we have the internet at home.

          Batteries? I don’t need batteries. I have the never-ending warm glow of weaponized autism. And that’s not even a joke.

          I tend to hyper-fixate on something until either it breaks or I do. It’s usually 70/30 in my favour :)

          • webghost0101@sopuli.xyz
            link
            fedilink
            arrow-up
            2
            ·
            15 hours ago

            Thats weird, i don’t remember having an alt account called SuspiciousCarrot78 but surely you must be me, same project, same neurology… same fixation pattern.

              • webghost0101@sopuli.xyz
                link
                fedilink
                arrow-up
                3
                ·
                14 hours ago

                Yep, definitely talking to myself again

                Jokes aside, i haven’t seen you mention anything for media streaming.

                I highly passionately recommend Navidrome for music. It is my absolute favorite and most used self hosted service.

                For acquiring media like film and music depending where you live ripping those from your local library is in some places arguably a protected fair use. (Comes from the time mp3 players became common and runners used to take rented cds in their walkman outside before). In my experience, 480p dvd is much higher quality then internet 480p streams and the total size is much smaller then what you find in downloads.

                ARM can help you automatically rip these as long as you have a drive in your pc. I got ARM running in a proxmox lxc with drive passtrough but that honestly was a pain to setup so not sure you should go that exact route, either way the moment arm is functional its smooth sailing and your only concern becomes storage space.

                • SuspciousCarrot78@lemmy.worldOP
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  edit-2
                  12 hours ago

                  Well, this is going to freak you out, because I am (literally, right now) explicitly scoping out offline YouTube integration into Jellyfin, as a sort of rolling library. Jellyfin has been good to me, but I’ve been using Nova Player for a while now, since my Pi borked itself (Nova player is plug hard drive into router, install app on TVs, done). The limit is that yt-dlp doesn’t integrate very well with it. I mean, I could build something, or fork the repo myself…or I could just use what already exists.

                  So it might be time to restore the entire *arr stack.

                  The TL;DR: I want one front end for ALL my media - YouTube, instructionals, movies, TV shows. That immediately speaks to Jellyfin, which I’m very familiar with. The issue is YouTube. There’s too much slop on there, I want a curated experience for the kids, SmartTube won’t work forever, and the eldest is starting to go black-hat and screw around with settings. That’s accelerating the timeline.

                  The stack I’m scoping:

                  • Jellyfin - front end for everything
                  • Tube Archivist - YouTube archive, metadata, download manager
                  • Tube Archivist Jellyfin plugin - maps channels as Shows, videos as Episodes, bidirectional playback sync
                  • *The usual arr stack (Sabnzbd, Sonarr, Radarr, etc.) - for maximum yarr me hearties. I’ve been downloading from 1337 like a pleb.
                  • Handbrake (+ usual media ripping stuff from DVD as needed)

                  The YT stack: rolling library logic:

                  • Core “keepers” - permanent, protected, not touched by auto-delete
                  • TA rescans subscribed channels twice a week
                  • Auto-delete watched videos after N days, per channel, marks them ignored so they don’t re-download
                  • Whole thing surfaces in Jellyfin as a YouTube-style shelf

                  Scoping the maths at 200GB, 30-min average per vid, using compressed modern codecs:

                  Planning numbers per video: assume average video is 30mins. At 360p, that’s ~100MB per video. 480p ~160MB, 540p ~220MB, 720p ~320MB.

                  If I have a selection of “core keepers” at 720p H.265 (~300 videos), taking up ~80GB, that leaves ~120GB for the rolling pool:

                  Rotating quality Rotating count Total library
                  360p ~1,200 ~1,500 (garbage; ok for kids cartoons)
                  480p ~750 ~1,050 (surprising ok)
                  540p ~545 ~845 (good to my eyes)
                  720p ~375 ~675 (very nice.)

                  I don’t need 4K…hell, 1080p is wasted on me. So I’m thinking… 300 core vids at 720p + rolling library at 540p = 845 videos, give or take. More than enough to keep the fam off my back once SmartTube goes tits up (they can’t play whack-a-mole for ever).

                  I would prefer a clean migration to other, live sources (I have those scoped out as well) but not all the Minecraft / gaming / pretend play / blah blah stuff the family watches is on Peertube/Odysee/Curiosity Stream.

                  PS: I see your 480p and raise you 60, because 540p is the forbidden resolution :)

                  PPS: I was planning on using JF for music too…but maybe I should look at Navidrome like you said.

                  The crazy idea that I had was to use AI to create an infinite playlist of sorts. Seed it with your own music, get it to generate tracks in THAT style as filler, intermingle them (so there’s always something new).

                  Finish off with AI DJ’s that pulls in “local news” from your curated RSS feeds.

                  Think: Three Dog from Fallout 3.

                  Basically what I spoke about here -

                  https://lemmy.world/post/43936980/22784324

                  I have a pretty clear idea of how to get that done. It could be amusing.

                  https://huggingface.co/ACE-Step/acestep-5Hz-lm-0.6B

  • Sims@lemmy.ml
    link
    fedilink
    arrow-up
    3
    ·
    1 day ago

    Yucky… The identity partner looks nasty. I bet all these corps are linked to the Epstein class and their think tanks. The 1% are taking control of their cattle…

  • steel_for_humans@piefed.social
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 day ago

    Say I have a GPU with 32GB VRAM and I am on Linux, what local LLM would be good for coding?

    Currently I just have an iGPU ;) but that’s always an option, albeit a very expensive one.

    • andrew0@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      2
      ·
      1 day ago

      Get llama.cpp and try Qwen3.6-35B-A3B. Just came out and looks good. You’ll have to look into optimal settings, as it’s a Mixture of Experts (MoE) model with only 3B parameters active. That means that the rest can stay in RAM for quick inference.

      You could also try the dense model (Qwen3.5-27B), but that will be significantly slower. Put these in a coding harness like Oh-My-Pi, OpenCode, etc. and see how it fares for your tasks. Should be ok for small tasks, but don’t expect Opus / Sonnet 4.6 quality, more like better than Haiku.

    • SuspciousCarrot78@lemmy.worldOP
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      1 day ago

      Sadly…none. Well, I mean…it depends what you mean by “coding”. If you mean “replace Claude with local?”. Then…none. Sorry.

      If you mean “actually, if I use ECA to call a cloud model from OpenRouter for planning, then have it direct a local LLM to do the scutt work”, then the Qwen series of models (like Qwen 3 Next) are pretty awesome.

      The iGPU will make you want to kill yourself though. Get a GPU :) Even a 4-16GB one can make a difference.

      PS: You said GPU and iGPU, so I’m not sure which one has the 32GB or what rig your running. I have suspicion though you’re running on a i5 or i7 with something like a intel 630 igpu inbuilt? In which case, the iGPU is pretty slow and depending on the exact chip, you likely won’t be able to use CUDA or Vulkan acceleration.

      So, the “get a GPU” thing still holds :)

      • steel_for_humans@piefed.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 day ago

        I meant that I can buy one of those Radeons dedicated to AI work, like the ASRock Radeon AI PRO R9700 Creator 32GB GDDR6. If I need to.

        Currently my Ryzen iGPU is all I need, because all I need is to see the graphical desktop environment on my screen ;) It does the job well.

        I use Claude Code as well and I am slightly concerned with that ID verification news, even more so because of the technology partner that they chose.

  • lsjw96kxs@sh.itjust.works
    link
    fedilink
    Français
    arrow-up
    1
    ·
    1 day ago

    Personally, I would like to use AI, but I don’t due to it being non local. I know there are local AI that could do things, but I don’t know which models are the good one for each task. If someone can give me pointers for it, I’d be grateful, for exemple a good model for local coding :)

    • lime!@feddit.nu
      link
      fedilink
      arrow-up
      2
      ·
      1 day ago

      depends on your hardware and your preferred language. i think wizardcoder is a pretty common choice but the smallest useful version is around 14GB so you need the vram to accommodate it.

      • lsjw96kxs@sh.itjust.works
        link
        fedilink
        Français
        arrow-up
        1
        ·
        56 minutes ago

        Thanks, I’ll dig into this. BTW, I have a 9070 XT, with 16 Gb VRAM, so it should do the job I guess.

    • SuspciousCarrot78@lemmy.worldOP
      link
      fedilink
      arrow-up
      1
      ·
      1 day ago
      1. How much VRAM do you have?
      2. Which GPU?
      3. What sort of coding do you want to do?

      No point in telling you “yo, dude, just grab MinMax 2.7 or GLM5.1”…unless you happen to have several GPUs running concurrently with a total combined VRAM pool of 500GB or more.

      There are strong local contenders… (Like Qwen3-Coder-Next but as you can see, the table ante is probably in the 45GB vram range just to load them up. Actually running them with a decent context length is likely to mean you need to be in the 80-100GB range.

      Do-able…but maybe pay $10 on OpenRouter first to test drive them before committing to $2000+ worth of hardware upgrades.

      There are other, more reasonable, less hardware dependent uses for local LLMs, but if you want fully local coders, it’s the same old story: pay to play (and that’s even if you don’t mind slow speed / overnight batch jobs).

      Right now, cloud-based providers are hemorrhaging money because they know it will lead to lock-in (ie: people will get use to what can be achieved with SOTA models, forgetting the multi-million dollar infrastructure required to run them). Then, when they realize you can’t quite do the same with local gear (at least, without spending $$$), they can ratchet the prices up.

      Codex pro-plan just went to $300/month.

      We’ve seen this playbook before, right?

      • lsjw96kxs@sh.itjust.works
        link
        fedilink
        Français
        arrow-up
        1
        ·
        50 minutes ago

        Thanks for the pointers. For the hardware, I have a 9070 XT with 16 Gb of VRAM. It’s sure that it can be very expensive. As I only do this as a hobby, I don’t want to pay that amount of money. I’m okay with having a slow llm as it wouldn’t be a tool I’d use often. I prefer to try doing things on my own and use the ai to help for little tasks first, such as checking why this one line of code didn’t want to work correctly or things like that.