Couldn’t make this shit up if I tried.

  • @brucethemoose
    link
    English
    49
    edit-2
    7 hours ago

    In case anyone missed it, Deepseek just released models that make OpenAI’s best nearly irrelevent… in the open, for anyone to host. For a tiny fraction of the hosting cost.

    Even the small distillation that fits on a 24GB VRAM desktop is incredible. And you can host it for others to use for free, with room for batching, like I’m doing right now. And there is so much that’s awesome about it, like the SFT training pipeline/code being published and the smaller models being built on top of models from another company (Qwen 2.5).

    I don’t even care what he’s saying now, but don’t believe a word that comes out of Altman’s mouth. He’s a just as much of a greedy con man as Musk, trying to gaslight everyone into thinking OpenAI will be relevant in a year, not a hollow, closed shell that sold out its research directive for cheap short term profit.

      • @brucethemoose
        link
        English
        7
        edit-2
        7 hours ago

        Dual 3060s are an option. LLMs can be split across GPUs reasonably well.

        3090s used to be like $700 used, but ironically they’ve gone up in price. I got mine for around $800 awhile ago, and stuffed it into 10L PC.

        Some people buy used P40s. There are rumors of a 24GB Arc B580. Also, AMD Strix Halo APU laptops/mini PCs can host it quite well, with the right software setup… I might buy an ITX board if anyone ever makes one.

        Also, there are 12GB/6GB VRAM distillations too, but 24GB is a huge intelligence step-up.

          • @brucethemoose
            link
            English
            27 hours ago

            For sure.

            The 14B distillation is still quite good, and usable on like 10GB GPUs. Maybe 8 with the right settings.