Surprised pikachu face

    • @UnderpantsWeevil
      link
      English
      12 months ago

      The entire training set isn’t used in each permutation. Your keywords are building the samples based on metadata tags tied back to the original images.

      If you ask for “Iron Man in a cowboy hat”, the toolset will reach for some catalog of Iron Man images and some catalog of cowboy hat images and some catalog of person-in-cowboy-hat images, when looking for a basis of comparison as it renders the image.

      These would be the images attributed to the output.

      • @[email protected]
        link
        fedilink
        English
        22 months ago

        Do you have a source for this? This sounds like fine-tuning a model, which doesn’t prevent data from the original training set from influencing the output. The method you described would only work if the AI is trained from scratch on only images of iron man and cowboy hats. And I don’t think that’s how any of these models work.