• copygirl
    link
    fedilink
    English
    3411 days ago

    Great, so it’s still wrong 1 out of 20 times, and just got even more energy intensive to run.

    • @kippinitreal
      link
      English
      811 days ago

      Genuine question: how energy intensive is it to run a model compared to training it? I always thought once a model is trained it’s (comparatively) trivial to query?

        • @[email protected]
          link
          fedilink
          English
          010 days ago

          How much energy does it take for the PC to be on and the user to type out that email manually?

          I assume we will get to a point where energy required starts to reduce as the computing power increases with moores law. However, it’s awful for the environment in the mean time.

          I don’t doub that rather than reducing energy, instead they will use more complex models requiring more power for these tasks for the foreseeable future. However eventually it will be diminishing returns on power and efficiency will be more profitable.

      • @[email protected]
        link
        fedilink
        English
        611 days ago

        For the small ones, with GPUs a couple hundred watts when generating. For the large ones, somewhere between 10 to 100 times that.

        With specialty hardware maybe 10x less.

        • Pennomi
          link
          English
          311 days ago

          A lot of the smaller LLMs don’t require GPU at all - they run just fine on a normal consumer CPU.

          • copygirl
            link
            fedilink
            English
            311 days ago

            Wouldn’t running on a CPU (while possible) make it less energy efficient, though?

            • Pennomi
              link
              English
              311 days ago

              It depends. A lot of LLMs are memory-constrained. If you’re constantly thrashing the GPU memory it can be both slower and less efficient.

          • @[email protected]
            link
            fedilink
            English
            110 days ago

            yeah but 10x slower, at speeds that just don’t work for many use cases. When you compare energy consumption per token, there isn’t much difference.

      • @[email protected]
        link
        fedilink
        English
        211 days ago

        Still requires thirsty datacenters that use megawatts of power to keep them online and fast for thousands of concurrent users

      • @dustyData
        link
        English
        811 days ago

        Not a very good, or easy comparison to make. Against the average, sure, the AI is above the average. But a domain expert like a doctor or an accountant is way much more accurate than that. In the 99+% range. Sure, everyone makes mistakes. But when we are good at something, we are really good.

        Anyways this is just a ridiculous amount of effort and energy wasted just to reduce hallucinations to 4.4%.

        • LughOPM
          link
          fedilink
          English
          311 days ago

          But a domain expert like a doctor or an accountant is way much more accurate

          Actually, not so.

          If the AI is trained on narrow data sets, then it beats humans. There’s quite a few examples of this recently with different types of medical expertise.

        • Ogmios
          link
          fedilink
          English
          0
          edit-2
          11 days ago

          It’s also notable that human error tends to occur in predictable ways which can be prepared for and noticed much more easily, while machine errors tend to be entirely random and unpredictable. For example: When a human makes a judgment on a medical issue which poses a very significant risk to the patient, they will generally put more effort into ensuring an accurate result/pay more attention to what they’re doing.

      • copygirl
        link
        fedilink
        English
        711 days ago

        I would not accept a calculator being wrong even 1% of the time.

        AI should be held to a higher standard than “it’s on average correct more often than a human”.