• @foggy
    link
    English
    1
    edit-2
    6 months ago

    Saying you wouldn’t use ‘ollama or gpt3.5’ is such a… I want to say uneducated statement? These are not two like terms

    You’re aware that ollama isn’t an LLM? You’re aware there are LLMs available via ollama that exceed gpt4s capabilities? I mean, you’re right that you need an array of expensive gpus to run them effectively, but… Just comparing ollama to gpt-3.5 is like comparing an NCAA basketball star to the Harlem globe trotters. It’s ridiculous at its face. A player compared to a team, for starters.

    • @[email protected]
      link
      fedilink
      English
      2
      edit-2
      6 months ago

      Correct, i kept it simple on purpose and could probably have worded it better.

      It was a meant as a broader statement including “both publicly available free to download models like those based on the ollama architectures as well as free to acces proprietary llm’s like gpt3.5”

      I personally tried variations of the vicuna, wizardLM and a few other models (mostly 30B, bigger was to slow) which are all based on ollama’s architecture but i consider those individual names to be less known.

      Neither of these impressed me all that much. But of course this is a really fast changing industry. Looking at the hf leaderboard i don’t see any of the models i tried. Last time i checked was January.

      I may also have an experience bias as i have become much more effective using gpt4 as a tool compared to when i just started to use it. This influences what I expect and how i write prompts for other models.

      I’d be happy to try some new models that have since archived new levels. I am huge supporter for self-hosting digital tools and frankly i cant wait to stop funding ClosedAi

      • @foggy
        link
        English
        16 months ago

        Llama3-70b is probably the most general purpose capable open source

        There are a bunch of contenders for specific purposes, like coding and stuff, though. I wanna say Mistral has a brand new enormous one that you’d need like 4 4090s to run smoothly.