• @[email protected]
    link
    fedilink
    93 days ago

    They’re going to keep making more powerful hardware either way, since parallel processing capability supports graphics and AI just fine.

    It’s not quite as simple as that. AI needs less precision than regular graphics, so chips developed with AI in mind do not necessarily translate into higher performance for other things.

    In science/engineering, people want more—not less—precision. So we look for GPUs with capable 64-bit processing, while AI is driving the industry in the other direction, from 32 down to 16.

      • @[email protected]
        link
        fedilink
        13 days ago

        That’s true, but I would like to see improvements driven along the consumer segment also. AI rendering is a nice software addition but I could easily see it becoming a distraction from hardware improvements.

        Consumers generally can’t just throw more money at a problem in the way that professional and business can.