• @[email protected]
    link
    fedilink
    143 months ago

    Ironically thanks in no small part to Facebook releasing Llama and kind of salting the earth for similar companies trying to create proprietary equivalents.

    Nowadays you either have gigantic LLMs with hundreds of billions of parameters like Claude and ChatGPT or you have open Models that are sub-200B.

    • Possibly linux
      link
      fedilink
      English
      83 months ago

      I personally think the really large models are useless. What is very impressive is the small ones that somehow manage to be good. It blows my mind that so much information can fit in 8b.

      • bruhduh
        link
        13 months ago

        True that, llms could be the future of lossy compression

      • bruhduh
        link
        13 months ago

        Waiting till mixtral gonna optimise it enough to run on home computer, and then till dolphin uncensor it