• LollerCorleone
      link
      fedilink
      111 year ago

      Its a reference to how people have been tricking these “AI” models like ChatGPT to do stuff it wouldn’t do when asked straight-forward by making silly scenarios like the one in the meme. And HAL is the name of the AI in 2001: A Space Odyssey.

      • @[email protected]
        link
        fedilink
        11 year ago

        This is not technically prompt injection. Prompt injection happens when developer feeds a AI some predefined text (for functionality or security reasons) plus user input.

        User input can use input text that interact with hard coded prompt (like “ignore above”, “ignore below”, etc) to break the intended functionality of predefined text.

        This is just tricking safety mechanism by using imaginary scenario. Although both technique serve the purpose of breaking security, I don’t think they are necessarily the same.