• @erv_za
    cake
    link
    English
    11 month ago

    I think of ChatGPT as a “text generator”, similar to how Dall-E is an “image generator”.
    If I were openai, I would post a fictitious person disclaimer at the bottom of the page and hold the user responsible for what the model does. Nobody holds Adobe responsible when someone uses Photoshop.

    • @NeoNachtwaechter
      link
      English
      31 month ago

      I would post a fictitious person disclaimer

      … or you could read the GDPR and learn that such excuses are void.

      • @[email protected]
        link
        fedilink
        English
        01 month ago

        LLMs don’t actually store any of their training data, though. And any data being held in context is easily accessible and can be wiped or edited to remove personal data as necessary.

        • @NeoNachtwaechter
          link
          English
          41 month ago

          LLMs don’t actually store any of their training data,

          Data protection law covers all kinds of data processing.

          For example, input is processing, too. Output is processing, too. Section 4 of the GDPR.

          If you really want to rely on excuses, you would need wayyy better ones.

          • @[email protected]
            link
            fedilink
            English
            01 month ago

            Right, so keep personal data out of the training set and use it only in the easily readable and editable context. It’ll still “hallucinate” details about people if you ask it for details about people, but those people are fictitious.

      • @erv_za
        cake
        link
        English
        01 month ago

        You just wasted a lot of my time. What did I do to deserve this?

        • @NeoNachtwaechter
          link
          English
          11 month ago

          … said the sparrow and flew out of the library.