• Hoxton
    link
    English
    216 months ago

    Well said! I’m still wondering what happens when the enviable ouroboros of AI content referencing AI content referencing AI content makes the whole internet a self perpetuating mess of unreadable content and makes anything of value these companies once gained basically useless.

    Would that eventually result in fresh, actual human created content only coming from social media? I guess clauses about using your likeness will be popping up in TikTok at some point (if they aren’t already)

    • maegul (he/they)
      link
      fedilink
      English
      10
      edit-2
      6 months ago

      I dunno, my feeling is that even if the hype dies down we’re not going back. Like a real transition has happened just like when Facebook took off.

      Humans will still be in the loop through their prompts and various other bits and pieces and platforms (Reddit is still huge) … while we may just adjust to the new standard in the same way that many reported an inability to do deep reading after becoming regular internet users.

      • @[email protected]
        link
        fedilink
        English
        86 months ago

        I think it’ll end up like Facebook (the social media platform, not the company). Eventually you’ll hit model collapse for new models trained off uncurated internet data once a critical portion of all online posts are made by AI, and it’ll become Much more expensive to create quality, up-to-date datasets for new models. Older/less tech literate people will stay on the big, AI-dominated platforms getting their brains melted by increasingly compelling, individually-tailored AI propaganda and everyone else will move to newer, less enshittified platforms until the cycle repeats.

        Maybe we’ll see an increase in discord/matrix style chatroom type social media, since it’s easier to curate those and be relatively confident everyone in a particular server is human. I also think most current fediverse platforms are also marginally more resistant to AI bots because individual servers can have an application process that verifies your humanity, and then defederate from instances that don’t do that.

        Basically anything that can segment the Unceasing Firehose of traffic on the big social media platforms into smaller chunks that can be more effectively moderated, ideally by volunteers because a large tech company would probably just automate moderation and then you’re back at square 1.

        • Hoxton
          link
          English
          36 months ago

          Honestly, that sounds like the most realistic outcome. If the history of the internet is anything to go by, the bubble will reach critical mass and not so much pop, as slowly deflate when something else begins to grow and take its place of hype.

        • maegul (he/they)
          link
          fedilink
          English
          16 months ago

          Great take.

          Older/less tech literate people will stay on the big, AI-dominated platforms getting their brains melted by increasingly compelling, individually-tailored AI propaganda

          Ooof … great way of putting it … “brain melting AI propaganda” … I can almost see a sci-fi short film premised on this image … with the main scene being when a normal-ish person tries to have a conversation with a brain-melted person and we slowly see from their behaviour and language just how melted they’ve become.

          Maybe we’ll see an increase in discord/matrix style chatroom type social media, since it’s easier to curate those and be relatively confident everyone in a particular server is human.

          Yep. This is a pretty vital project in the social media space right now that, IMO, isn’t getting enough attention, in part I suspect because a lot of the current movements in alternative social media are driven by millennials and X-gen nostalgic for the internet of 2014 without wanting to make something new. And so the idea of an AI-protected space doesn’t really register in their minds. The problems they’re solving are platform dominance, moderation and lock-in.

          Worthwhile, but in all serious about 10 years too late and after the damage has been done (surely our society would be different if social media didn’t go down the path it did from 2010 onward). Now what’s likely at stake is the enshitification or en-slop-ification (slop = unwanted AI generated garbage) of internet content and the obscuring of quality human-made content, especially those from niche interests. Algorithms started this, which alt-social are combating, which is great.

          But good community building platforms with strong privacy or “enclosing” and AI/Bot protecting mechanisms are needed now. Unfortunately, all of these clones of big-social platforms (lemmy included) are not optimised for community building and fostering. In fact, I’m not sure I see community hosting as a quality in any social media platforms at the moment apart from discord, which says a lot I think. Lemmy’s private and local only communities (on the roadmap apparently) is a start, but still only a modification of the reddit model.

          • @afraid_of_zombies
            link
            English
            26 months ago

            person tries to have a conversation with a brain-melted person and we slowly see from their behaviour and language just how melted they’ve become.

            I see you have met my Fox News watching parents.

            • maegul (he/they)
              link
              fedilink
              English
              06 months ago

              LOL (I haven’t actually met someone like that, in part because I’m not a USian and generally not subject to that sort of type ATM … but I am morbidly curious TBH.

      • Hoxton
        link
        English
        26 months ago

        You’re absolutely right about not going back. Web 3.0 I guess. I want to be optimistic that a distinction between all the garbage and actual useful or real information will be visible to people, but like you said, general tech and media literacy isn’t encouraging, hey?

        Slightly related, but I’ve actually noticed a government awareness campaign where I live about identifying digital scams. Be nice if that could be extended to incorrect or misleading AI content too.

    • @assassin_aragorn
      link
      English
      66 months ago

      It should end up self regulating once AI is using AI material. That’s the downfall of the companies not bothering to put very clear identification of AI produced material. It’ll spiral into a hilarious mess.

      • Hoxton
        link
        English
        46 months ago

        I’m legit looking forward to when Google returns completely garbled and unreadable search results, because someone is running an automated Ads campaign that sources another automated campaign and so on, with the only reason it rises to the top is that they put the highest bid amount.

        I doubt Google will do shit about it, but at least the memes will be good!

      • @afraid_of_zombies
        link
        English
        16 months ago

        Hasn’t it already happened? All culture is derivative, yes all of it. And look at how much of it is awful, yet we navigate fine. I keep hearing stats like every one second YouTube gets 4 hours more content and yet I use YouTube daily. Despite being very very confident that all but a fraction of a percent of what it has is of any value to me.

        Same for books, magazines, news, podcasts, radio programs, music, art, comics, recipes, articles…

        We already live in the post information explosion. Where the same stuff gets churned over and over again. All I am seeing AI doing is speeding this up. Now instead of a million YouTube vids I won’t watch getting added next week it will be ten million.

    • @afraid_of_zombies
      link
      English
      36 months ago

      Tik Tok was banned so it ain’t coming from there. Can’t get universal healthcare but we can make sure to protect kids from the latest dance craze.

    • Gnome Kat
      link
      fedilink
      English
      26 months ago

      Thats a technical issue that likely can be solved. I doubt some feedback loop of training data will be the downfall of AI… The way to stop it is to refuse to use it( lets be real the regulators arnt gana do shit)