• Eager Eagle
    link
    English
    184 hours ago

    I bet he just wants a card to self host models and not give companies his data, but the amount of vram is indeed ridiculous.

    • Jeena
      link
      fedilink
      English
      83 hours ago

      Exactly, I’m in the same situation now and the 8GB in those cheaper cards don’t even let you run a 13B model. I’m trying to research if I can run a 13B one on a 3060 with 12 GB.

      • The Hobbyist
        link
        fedilink
        English
        32 hours ago

        You can. I’m running a 14B deepseek model on mine. It achieves 28 t/s.

        • @[email protected]
          link
          fedilink
          English
          141 seconds ago

          I also have a 3060, can you detail which framework (sglang, ollama, etc) you are using and how you got that speed? i’m having trouble reaching that level of performance. Thx

        • Jeena
          link
          fedilink
          English
          31 hour ago

          Oh nice, that’s faster than I imagined.