• @CrayonRosary
    link
    English
    21 year ago

    Lemmy users: Copyright law is broken and stupid.

    Also Lemmy users: A.I. violates copyright law!

    • User
      link
      fedilink
      English
      161 year ago

      Where’s the contradiction though

    • Alien Nathan Edward
      link
      fedilink
      English
      51 year ago

      yes. there are myriad ways that copyright law is broken and stupid, but protecting the creations of independent artists isn’t one of them

      take this bullshit back to reddit

    • Liz
      link
      fedilink
      English
      01 year ago

      I mean, consistency is better than inconsistency, even if we don’t agree with the rules.

    • @UnderpantsWeevil
      link
      English
      -2
      edit-2
      1 year ago

      A.I. doesn’t violate copywrite laws. It is the data-mining done to train A.I. and the regurgitation of said data in the responses that ultimately violate these laws. A model trained on privately owned, properly licensed, or exclusively public works wouldn’t be a problem.

      Even then, I would argue that lack of attribution is a bigger problem than merely violating copywrite. A big part of the LLM mystique is in how it can spit out a few lines of Shakespeare without accreditation and convince its users that its some kind of master poet.

      Copywrite law is stupid and broken. But plagarism is a problem in its own right, as it seeks to effectively sell people their own creative commons at an absurd markup.

      • @[email protected]
        link
        fedilink
        English
        6
        edit-2
        1 year ago

        A model trained on privately owned, properly licensed, or exclusively public works wouldn’t be a problem.

        This is how we end up with only corpo owned AIs being allowed to exist imo, places like stock photo sites are the only ones with large enough repositories of images to train AI that they have all the legal rights to

        The way I see it, either generative AI is legal, free for everyone to run locally, and the created works are public domain, OR, everyone pays $20/mo to massive faceless corpos for the rest of their lives to have the privilege of access to it because they’re the only ones who own all (or have enough money to license) the IP needed to train them

        • @UnderpantsWeevil
          link
          English
          01 year ago

          This is how we end up with only corpo owned AIs being allowed to exist imo

          Its how you end up with sixteen different streaming services that only vend a sliver of the total available content, sure. But the underlying technology of AI grows independent of what its trained on.

          The way I see it, either generative AI is legal, free for everyone to run locally, and the created works are public domain, OR, everyone pays $20/mo to massive faceless corpos for the rest of their lives to have the privilege of access to it

          There are other alternatives. These sites can be restricted to data within the public domain. And we can increase our investment in public media. The problem of NYT articles being digested and regurgitated as ChatGPT info-vomit isn’t a problem if the NYT is a publicly owned and operated enterprise. Then its not struggling to profit off journalism, but treating this information as a loss-leading public service open to all, with ChatGPT simply operating as a tool to store, process, and present the data.

          Similarly, if you limit generative AI to the old Mickey Mouse and Winnie-the-Pooh films from the 1930s, you leave plenty of room for original artists to create new works without fear that their livelihoods get chews up and fed back into the system. If you invest in public art exhibitions then these artists can get paid to pursue their craft, the art becomes public domain immediately, and digital tools that want to riff on the original are free to do so without undermining the artists themselves.