As if EA didn’t already make bland, derivative games…

  • @phoneymouse
    link
    398 months ago

    Because players are just dying for generative AI!!!

    • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 ℹ️
      link
      fedilink
      English
      18
      edit-2
      8 months ago

      I unironically am; but I don’t think EA is gonna use it the way I would actually like to see it used in a video game. They hear “I want AI dialogue” and will use AI to write (and probably code) the game, instead of having it generated real time so you can communicate with the NPCs as if they were actually there.

      • @moriquende
        link
        248 months ago

        The problem is that you can’t really control what AI spits out. May fit perfectly into the story, or it may be immersion breaking nonsense that doesn’t even fit into the narrative. What if a character suddenly makes a promise or tells you a key plotline point that it has just made up? I, for one, prefer games to be handcrafted to deliver a quality reliable experience instead of being a coinflip.

        • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 ℹ️
          link
          fedilink
          English
          -8
          edit-2
          8 months ago

          You can control what it spits out, though. They already do somewhat.

          Edit: Gonna go out on a limb and assume most of you haven’t actually played any of the projects currently doing this. Or mess with chatbots at all.

          • @moriquende
            link
            198 months ago

            Somewhat is key. You can try to guide it in a direction, but that’s it. Also, as a player, you can never be sure if the dialogue is meaningful or not. Does it reveal something about the plot? Is it a key information about the character? Is it just hallucinated gibberish to fill the space?

            • @Jesus_666
              link
              58 months ago

              Besides, LLMs struggle with retaining contextual information for long and they’re pretty dang resource hungry. Expect a game with LLM-driven dialogue to reserve several gigs of VRAM and a fair chunk of GPU processing power solely for that.

              And then you still get characters who hallucinate plot points or suddenly speak gibberish.

          • MentalEdge
            link
            fedilink
            3
            edit-2
            8 months ago

            You really can’t.

            You can run checks and fence it in with traditional software, you can train it more narrowly…

            I haven’t seen anything that suggests AI hallucinations are actually a solvable problem, because they stem from the fact that these models don’t actually think, or know anything.

            They’re only useful when their output is vetted before use, because training a model that gets things 100% right 100% of the time, is like capturing lightning in a bottle.

            It’s the 90/90 problem. Except with AI it’s looking more and more like a 90/99.99999999 problem.