• @[email protected]
    link
    fedilink
    41 year ago

    *"The first technique is called an encoder attack. PhotoGuard adds imperceptible signals to the image so that the AI model interprets it as something else. For example, these signals could cause the AI to categorize an image of, say, Trevor Noah as a block of pure gray. As a result, any attempt to use Stable Diffusion to edit Noah into other situations would look unconvincing.

    The second, more effective technique is called a diffusion attack. It disrupts the way the AI models generate images, essentially by encoding them with secret signals that alter how they’re processed by the model. By adding these signals to an image of Trevor Noah, the team managed to manipulate the diffusion model to ignore its prompt and generate the image the researchers wanted. As a result, any AI-edited images of Noah would just look gray." *

    What about copyright? I don’t see why as a copyright holder of an image we can’t just forbid them from using the image and if they do, then we have grounds to sue. Why are we ok with just letting companies steal and “use” our work? You wouldn’t let them just create a new ad with your image. And I would argue that taking someone’s face and superimposing it onto porn is not fair use, but hey, I am definitely not a lawyer, just someone dumping thoughts.

    • @[email protected]
      link
      fedilink
      11 year ago

      Because they right now AI companies have deep pockets and some amount of goodwill from the tech community just because of the massive potential they hold, so they’re given a free pass. It will take a while for the more privacy minded countries to come around and reign these companies in, and establish precedents to sue these companies if they use copyrighted material.