• Yerbouti@sh.itjust.works
    link
    fedilink
    arrow-up
    28
    arrow-down
    1
    ·
    15 hours ago

    The future of AI has to be local and self-hosted. Soon enough you’ll have super powerful models that can run on your phone. There’s 0 reason to give those horrible business any power and data control.

    • yucandu@lemmy.world
      link
      fedilink
      arrow-up
      12
      arrow-down
      2
      ·
      14 hours ago

      Not to mention the one that I run locally on my GPU is trained on ethically-sourced data without breaking any copyright or data licensing laws, and yet it somehow works BETTER at ChatGPT for coding.

      • Bob Robertson IX @discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        2
        ·
        10 hours ago

        I agree with you that it needs to be local and self-hosted… I currently have an incredible AI assistant running locally using Qwen3-Coder-Next. It is fast, smart and very capable. However, I could not have gotten it setup as well as I have without the help of Claude Code… and even now, as great as my local model is, it still isn’t to the point that it can handle modifying its own code as well as Claude. The future is local, but to help us get there a powerful cloud-based AI adds a lot of value.

        • SuspciousCarrot78@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          13 minutes ago

          Thank you for honestly stating that. I am in similar position myself.

          How do you like Qwen 3 next? With only 8GB vram I’m limited in what I can self host (maybe the Easter bunny will bring me a Strix lol).

      • TheFinn@discuss.tchncs.de
        link
        fedilink
        arrow-up
        4
        ·
        13 hours ago

        I’m having difficulty with getting off the ground with these. Primarily I don’t trust the companies or individuals involved. I’m hoping for open source, local, with a GUI for desktop use and an API for automation.

        What model do you use? And in what kind of framework?

        • wonderingwanderer@sopuli.xyz
          link
          fedilink
          arrow-up
          2
          ·
          5 hours ago

          Huggingface lists thousands of open source models. Each one has a page telling you what base model it’s based on, what other models are merged into it, what data its fine-tuned on, etc.

          You can search by number of parameters, you can find quantized versions, you can find datasets to fine-tune your own model on.

          I don’t know about GUI, but I’m sure there are some out there. Definitely options for API too

        • Alloi@lemmy.world
          link
          fedilink
          arrow-up
          3
          ·
          12 hours ago

          R1 last i checked seems to be decent enough for a local model. customizable. but that was a while ago. its release temporarily crashed Nvidia stock because they showed how smart software design trumps mass spending on cutting edge hardware.

          at the end of the day its all of our data. we should own the means, especially if we built it by simply existing on the internet. without consent.

          if we wish to do this, its crucial that we do everything in our power to dismantle the “profit” structure and investment hype. sooner or later someone will leak the data, and we will have access to locally run versions we can train ourselves. as long as we dont allow them to monopolize hardware, we can have the brain, and the body of it run local.

          thats the only time it will be remotely ethical to use, unless its the persuit of attaining these goals.

    • brucethemoose@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      13 hours ago

      RAM constraints make phone running difficult. As do the more restricted quantization schemes NPUs require. 1B-8B LLMs are shockingly good backed with RAG, but still kind of limited.

      It seemed like Bitnet would solve all that, but the big model trainers have ignored it, unfortunately. Or at least not told anyone about their experiments with it.

      • SuspciousCarrot78@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        10 minutes ago

        M$ are dragging its feet with BITNET for sure and no one else seems to be cooking. They were meant to have released 8b and 70b models buy now (according to source files in repo). Here’s hoping.