Couldn’t make this shit up if I tried.

  • @brucethemoose
    link
    English
    7
    edit-2
    6 hours ago

    Dual 3060s are an option. LLMs can be split across GPUs reasonably well.

    3090s used to be like $700 used, but ironically they’ve gone up in price. I got mine for around $800 awhile ago, and stuffed it into 10L PC.

    Some people buy used P40s. There are rumors of a 24GB Arc B580. Also, AMD Strix Halo APU laptops/mini PCs can host it quite well, with the right software setup… I might buy an ITX board if anyone ever makes one.

    Also, there are 12GB/6GB VRAM distillations too, but 24GB is a huge intelligence step-up.

      • @brucethemoose
        link
        English
        26 hours ago

        For sure.

        The 14B distillation is still quite good, and usable on like 10GB GPUs. Maybe 8 with the right settings.