Obviously there’s not a lot of love for OpenAI and other corporate API generative AI here, but how does the community feel about self hosted models? Especially stuff like the Linux Foundation’s Open Model Initiative?

I feel like a lot of people just don’t know there are Apache/CC-BY-NC licensed “AI” they can run on sane desktops, right now, that are incredible. I’m thinking of the most recent Command-R, specifically. I can run it on one GPU, and it blows expensive API models away, and it’s mine to use.

And there are efforts to kill the power cost of inference and training with stuff like matrix-multiplication free models, open source and legally licensed datasets, cheap training… and OpenAI and such want to shut down all of this because it breaks their monopoly, where they can just outspend everyone scaling , stealiing data and destroying the planet. And it’s actually a threat to them.

Again, I feel like corporate social media vs fediverse is a good anology, where one is kinda destroying the planet and the other, while still niche, problematic and a WIP, kills a lot of the downsides.

  • @[email protected]
    link
    fedilink
    English
    43 months ago

    I read localllama through redlib but I don’t contribute. I am not technical enough to contribute and I don’t understand the math.

    I have been looking at YouTube for some videos to try to explain it, but I haven’t found anything that is in the sweet spot between “video for non-technical people” and “video for people with PhD and quantum physics”

    • @brucethemooseOP
      link
      3
      edit-2
      3 months ago

      It’s a giant mess. Even the technical vidoes tend to be theoretical, and are either obsolete or do nothing to help you actually run them.

      I would know nothing if I hadn’t been following the community since the Pygmalion/ESRGAN days

      • Bob Robertson IX
        link
        23 months ago

        I’ve spent the past 2 years looking for the open source AI community, but haven’t really found it. I’ve tinkered with Stable Diffusion and Ollama and I want to learn more, but haven’t found the right places online yet.

        • @brucethemooseOP
          link
          83 months ago

          I’ll give you one hint, a lot of the community is locked away in various Discords.

          This is one of the many reasons I hate Discord.

        • @brucethemooseOP
          link
          33 months ago

          And just to be more helpful, I can point you in the right direction depending on your hardware.

          • Bob Robertson IX
            link
            13 months ago

            Yeah, I hate Discord too but that has been the best place I’ve found the best information, but even then it doesn’t really feel like a community.

            I’m running on an Apple M1 at the moment, likely to upgrade to an M4 when it is released.

            • @brucethemooseOP
              link
              1
              edit-2
              3 months ago

              What RAM capacity?

              Honestly, if LLMs are your focus, you should just upgrade to a used M2 Max (or Ultra) when the M4 comes out, lol. Basically the only thing that matters is RAM capacity and bandwidth, and the M2 is just going to be faster and better than a similarly priced M4.

              Or better yet, upgrade to and AMD Strix Halo. This will buy you into linux and the cuda ecosystem (through AMD rocm), which is going to open a lot of doors and save headaches (while admittedly creating other headaches).

              • Bob Robertson IX
                link
                13 months ago

                Honestly I’ve mostly been playing around with image generators and learning how to write a good prompt. But LLMs are where I see real value and would love to learn more about self-hosting one, and custom training it.

                I find it interesting that the M2 is better for LLM than the M4 will be… what’s the reason for this?

                • @brucethemooseOP
                  link
                  1
                  edit-2
                  3 months ago

                  RAM capacity and bandwidth.

                  That basically the only two things that matter for local LLM performance, as it has to read the entire model from memory for every token (aka half word). And for the same money, a “higher end” M2 (like an M2 Max or Ultra) will just have more of it than the equivalent cost M3 or (probably) M4.

        • sunzu2
          link
          fedilink
          13 months ago

          Hate to suggested it but have you checked reddit localllama?