Obviously there’s not a lot of love for OpenAI and other corporate API generative AI here, but how does the community feel about self hosted models? Especially stuff like the Linux Foundation’s Open Model Initiative?

I feel like a lot of people just don’t know there are Apache/CC-BY-NC licensed “AI” they can run on sane desktops, right now, that are incredible. I’m thinking of the most recent Command-R, specifically. I can run it on one GPU, and it blows expensive API models away, and it’s mine to use.

And there are efforts to kill the power cost of inference and training with stuff like matrix-multiplication free models, open source and legally licensed datasets, cheap training… and OpenAI and such want to shut down all of this because it breaks their monopoly, where they can just outspend everyone scaling , stealiing data and destroying the planet. And it’s actually a threat to them.

Again, I feel like corporate social media vs fediverse is a good anology, where one is kinda destroying the planet and the other, while still niche, problematic and a WIP, kills a lot of the downsides.

  • @brucethemooseOP
    link
    63 months ago

    If you have a 24GB card, just go straight to the most recent Command R, a 3.75bpw-4bpw quantization. It’s incredible, and you can do the full 131K context on a 24GB GPU easy.

    Gemma 27B Is actually quite good, but “narrow.” Its super low context and seems to be hyper optimized for short chatbot-arena style questions.

    • @[email protected]
      link
      fedilink
      English
      4
      edit-2
      3 months ago

      Gemma 27B Is actually quite good, but “narrow.” Its super low context and seems to be hyper optimized for short chatbot-arena style questions.

      This is the stuff I love to know so thanks for sharing. I will be pulling Command R tomorrow.

      • @brucethemooseOP
        link
        33 months ago

        Good! So Command-R excels at “RAG” style tasks like asking questions about a huge document, continuing a long story or so on. You should also read up on its super intricate system prompt format, which can steer it quite well.

        I dunno about code, I tend to use Mistral Code 22B (or deepseek v2 API) for that.

        I am happy to ramble on about this stuff, just ask.