• TimeSquirrel
    link
    fedilink
    58
    edit-2
    2 months ago

    This has been obvious for a while to those of us using GitHub Copilot for programming. Start a function, and then just keep hitting tab to let it autotype based on what it already wrote. It quickly devolves into strange and random bullshit. You gotta babysit it.

    • @0laura
      link
      English
      132 months ago

      very unlikely to stem from model collapse. why would they use a worse model? it’s probably because they neutered it or gave it less resources.

      • TimeSquirrel
        link
        fedilink
        11
        edit-2
        2 months ago

        It learns from your own code as you type so it can offer more relevant suggestions unlike the web-based LLMs. So you can make it feed back on itself.

    • NekuSoul
      link
      fedilink
      English
      92 months ago

      Same thing with Stable Diffusion if you’ve ever used a generated image as an input and repeated the same prompt. You basically get a deep-fried copy.

      • FaceDeer
        link
        fedilink
        32 months ago

        img2img is not “training” the model. Completely different process.

        • NekuSoul
          link
          fedilink
          English
          12 months ago

          Oh yeah, you’re right. It’s both degradation in some way, but through entirely different causes.