cross-posted from: https://lemmy.world/post/2312869

AI researchers say they’ve found ‘virtually unlimited’ ways to bypass Bard and ChatGPT’s safety rules::The researchers found they could use jailbreaks they’d developed for open-source systems to target mainstream and closed AI systems.

    • LoafyLemon
      link
      fedilink
      11 year ago

      Not really. This isn’t AGI but a text transformer. They trained it so the most probable answer to unwanted questions is ‘I’m sorry but as an AI…’.

      However, if you phrase your question in a way researchers haven’t thought about, you will bypass the filter.

      There’s not an ounce of intelligence in LLMs, it’s all statistics.