• LoafyLemon
      link
      fedilink
      7
      edit-2
      9 months ago

      Old news pal! 😄

      [4/27] Thanks to the community effort, LLaVA-13B with 4-bit quantization allows you to run on a GPU with as few as 12GB VRAM! Try it out here.

      • @[email protected]
        link
        fedilink
        English
        29 months ago

        12GB of VRAM is still an upgrade away for most people and a 4bit quantized 13B model is barely going to be a tech demo. When open source ai is proclaimed to be near/on par/better then gpt4 they are talking about nothing else than their biggest models in a prime environment.

        • @just_another_person
          link
          English
          19 months ago

          Sure, but not for standard cloud instances that are very affordable for companies wanting to get away from OpenAI.

          • @[email protected]
            link
            fedilink
            English
            19 months ago

            I usually don’t think much about companies and cloud instances when it comes to Fossai but fair enough.

            For me its all about locally run consumer models. If we cannot archive that it means we will always need to rely on the wims and decisions of others to acces the most transforming technology ever invented.

    • @sudo22
      link
      English
      29 months ago

      Holy shit a terabyte?