• BranBucket
    link
    fedilink
    English
    arrow-up
    54
    arrow-down
    1
    ·
    edit-2
    2 hours ago

    People don’t often realize how subtle changes in language can change our thought process. It’s just how human brains work sometimes.

    The old bit about smoking and praying is a great example. If you ask a priest if it’s alright to smoke when you pray, they’re likely to say no, as your focus should be on your prayers and not your cigarette. But if you ask a priest if it’s alright to pray while you’re smoking, they’d probably say yes, as you should feel free to pray to God whenever you need…

    Now, make a machine that’s designed to be agreeable, relatable, and makes persuasive arguments but that can’t separate fact from fiction, can’t reason, has no way of intuiting it’s user’s mental state beyond checking for certain language parameters, and can’t know if the user is actually following it’s suggestions with physical actions or is just asking for the next step in a hypothetical process. Then make the machine try to keep people talking for as long as possible…

    You get one answer that leads you a set direction, then another, then another… It snowballs a bit as you get deeper in. Maybe something shocks you out of it, maybe the machine sucks you back in. The descent probably isn’t a steady downhill slope, it rolls up and down from reality to delusion a few times before going down sharply.

    Are we surprised some people’s thought processes and decision making might turn extreme when exposed to this? The only question is how many people will be effected and to what degree.

    • MinnesotaGoddam
      link
      fedilink
      English
      arrow-up
      5
      ·
      3 hours ago

      People don’t often realize how subtle changes in language can change our thought process.

      just changing a single word in your daily usage can change your entire outlook from negative to positive. it’s strange, but unless you’ve experienced it yourself how such minute changes can have such large effects it’s hard to believe.

      • BranBucket
        link
        fedilink
        English
        arrow-up
        2
        ·
        46 minutes ago

        And this is hard for me, actually. Because of my work background and the jargon used, I’m unconsciously negative about things a lot of the time. It’s a tough habit to break.

        • MinnesotaGoddam
          link
          fedilink
          English
          arrow-up
          1
          ·
          19 minutes ago

          Oh, me too. I’m just innately full of negative self talk. I try to direct positivity outward if I can’t aim it at myself at least

    • Ulrich@feddit.org
      link
      fedilink
      English
      arrow-up
      2
      ·
      3 hours ago

      But if you ask a priest if it’s alright to pray while you’re smoking, they’d probably say yes, as you should feel free to pray to God whenever you need…

      When would a priest ever tell anyone it’s not okay to pray?

      • BranBucket
        link
        fedilink
        English
        arrow-up
        8
        ·
        edit-2
        2 hours ago

        It’s the opinion on smoking, not praying, that differs.

        In both cases you’re praying and smoking at the same time, so your actions don’t change, but the priest rationalizes two completely different answers based on the way the question is posed. It’s just an example to show how two contradictory answers can seem rational to the same person because of the language used.

        • sudoer777@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 hours ago

          the priest rationalizes two completely different answers based on the way the question is posed. It’s just an example to show how two contradictory answers can seem rational to the same person because of the language used.

          They aren’t contradictory though. Basically what they are saying is just praying > praying + smoking > just smoking. “Okay” has different meanings in the different sentences.

          • BranBucket
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            34 minutes ago

            But in both cases, the person is asking to do the same thing. The order of the words in the sentence doesn’t change the end result, we always wind up with someone smoking and praying simultaneously, which may or may not be against God’s will.

            Strip away the justifications and simplify the word choices and you get this:

            1. May I smoke while I pray? No, you may not.
            2. May I pray while I smoke? Yes, you may.

            Given that, can you say if it is right or wrong to smoke and pray simultaneously?

            And again, this is just a hypothetical scenario. In the broader context of life, religion, and tobacco use, it’ll never be this simple, but it works for an example.

            Now, someone might point out that by simplifying the wording, I’ve changed the meaning of the original statement to make it fit my argument, and that now it means something else. But that’s essentially my original point, phrasing and word choices can shape our reasoning, thought processes, and how we interpret meaning in ways we aren’t immediately aware of, leading us to different conclusions or even delusional thinking.

            • sudoer777@lemmy.ml
              link
              fedilink
              English
              arrow-up
              1
              ·
              24 minutes ago

              But in both cases, the person is asking to do the same thing.

              Not really. They’re not just asking if they should pray and smoke simultaneously if you put them in contexts where it actually makes sense to ask those questions.

              May I smoke while I pray? No, you may not.

              First, “pray” can mean different things, such as (1) a deep focused session, or (2) a lighter more casual session, both of which are standard definitions of the word. Since this request emphasizes prayer as the main action, (1) is most likely here. For a focused session, smoking is a distraction and not a good idea. The definition of “may” here is also subjective and not necessarily absolute, some people may consider it disrespectful, while others may still say that prayer at all is better than no prayer regardless of side actions, but it’s better to not smoke.

              May I pray while I smoke? Yes, you may.

              In this sentence, definition (2) of prayer seems more likely since the main focus of the request is smoking. Which to some people this may still be considered disrespectful like in the first request, but others are supportive of more casual prayer and smoking during casual prayer isn’t a problem like in focused prayer, and the idea that prayer is better than no prayer and “may” isn’t absolute still applies.

              And again, this is just a hypothetical scenario. In the broader context of life, religion, and tobacco use, it’ll never be this simple, but it works for an example.

              Not if you’re trying to prove that they’re contradictory and irrational, since the context is what actually makes the words mean something. If you take away the context, then it’s nothing more than shapes on a screen.

              Now, someone might point out that by simplifying the wording, I’ve changed the meaning of the original statement to make it fit my argument, and that now it means something else. But that’s essentially my original point, phrasing and word choices can shape our reasoning, though processes, and how we interpret meaning in ways we aren’t immediately aware of

              I agree with that

    • CeeBee_Eh
      link
      fedilink
      English
      arrow-up
      8
      ·
      4 hours ago

      Are we surprised some people’s thought processes and decision making might turn extreme when exposed to this?

      Yes, actually. I’m not doubting the power of language, but I cannot ever see something anyone ever says alter my sense of reality or right from wrong.

      I had a “friend” say to me recently “why do you always go against the grain?” My reply was “I will go against the grain for the rest of my life if it means doing or saying what’s right”.

      I guess my point is that I have a very hard time relating to this.

      • BranBucket
        link
        fedilink
        English
        arrow-up
        4
        ·
        4 hours ago

        I guess my point is that I have a very hard time relating to this.

        That’s fair. In the same vein, you might find a priest that tells you to stop smoking for your health no matter how you phrase the question about lighting up and prayer. What people are receptive to is going to vary.

        I’d like argue that more of us are susceptible to this sort of thing than we suspect, but that’s not really something that can be proved or disproved. What seems pretty certain is that at least some of us are at risk, and given all the other downsides of chatbots, it’d be best to regulate them in a hurry.

        • CeeBee_Eh
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 hours ago

          you might find a priest that tells you to stop smoking for your health no matter how you phrase the question about lighting up and prayer. What people are receptive to is going to vary.

          Ya, I’ve read the thing about praying and smoking in another comment. The funny thing is that I have very specific opinions about smoking and would argue that smoking while praying is disrespectful, but God would listen in any case.

          • BranBucket
            link
            fedilink
            English
            arrow-up
            1
            ·
            56 minutes ago

            It’s more about how the slightly different questions lead the hypothetical priest to two separate and contradictory conclusions than disrespecting God.

            At any rate, all opinions on tobacco and prayer are fine by me, just watch out for any friends you think might be talking to chatbots a little too much.

        • Regrettable_incident
          link
          fedilink
          English
          arrow-up
          2
          ·
          3 hours ago

          Sure, that’s why propaganda can be so powerful. It’s not just what is said, it’s how it’s said. And pretty much everyone if 3 vulnerable to the right propaganda - especially people who think they’re not vulnerable to propaganda.

          • BranBucket
            link
            fedilink
            English
            arrow-up
            1
            ·
            54 minutes ago

            Absolutely, and the medium can make a huge difference as well. I suspect that there’s something about chatbots and the medium of their messages that helps set those hooks extra deep in people.

    • Zink@programming.dev
      link
      fedilink
      English
      arrow-up
      4
      ·
      6 hours ago

      Then make the machine try to keep people talking for as long as possible…

      That’s probably a huge part of it. How many billions of dollars have been spent engineering content on a screen to get its tendrils into people’s minds and attention and not let go?

      EnGaGeMent!!!

      • BranBucket
        link
        fedilink
        English
        arrow-up
        1
        ·
        36 minutes ago

        This is also part of my broader gripe with social media, cable news, and the current media landscape in general. They use so many sneaky little psychological hooks to keep you plugged in that I honestly believe it’s screwing with our heads to the point of it being a public health crisis.

        People are already frazzled and beat down by the onslaught of dopamine feedback loops and outrage bait, then you go and get them hooked on a charbot that feeds into every little neurosies they’ve developed and just sinks those hooks in even deeper and it’s no wonder some people are having a mental health crisis.

        A lot of us vastly overestimate our resistance to having our heads jacked with and it worries me.

    • Nomorereddit@lemmy.today
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      12
      ·
      5 hours ago

      Gtfo here. I grew up in xbox live chat rooms w the most vile language imaginable. I am now a senior Mgr with 100 ppl under me.

      And ill just say, ill no scope them in a heart beat if they spawn camp…

      …I mean I drive productivity at the speed of trust.

  • SocialMediaRefugee
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    9
    ·
    2 hours ago

    Judas Priest got sued by parents claiming their kid killed himself over hidden messages in their music.

      • SocialMediaRefugee
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        5
        ·
        52 minutes ago

        A delusional kid was told by one and a delusional kid was told by another.

    • SpookyBogMonster@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      30 minutes ago

      The difference is, there were no hidden messages in the music.

      Meanwhile there are overt messages spat out by the LLM, because it’s a lying yes-man machine that encourages people’s worst impulses, so they keep using it.


      Rob Halford just wanted to dress like a Tom of Finland drawing, and make fun music.

      The companies making the chatbots want to harvest and sell your data.

  • Gammelfisch
    link
    fedilink
    English
    arrow-up
    6
    ·
    5 hours ago

    How in the hell does one become addicted to a damn chatbot?

    • Lemminary
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      4
      ·
      3 hours ago

      Money + downtime + not very smart?

  • HertzDentalBar@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    5 hours ago

    Maybe if we’re lucky people will realize this has been what capitalism and consumerism has been doing all along. People have been drivin to crazy shit because of all the evil shit we do marketing and fucking with consumers minds. But nah we will blame a chatbot that’s just telling you what it thinks you want to see rather than seeing it’s just the next stage of fuckery

  • man_wtfhappenedtoyou
    link
    fedilink
    English
    arrow-up
    11
    ·
    7 hours ago

    How do you even get these chat bots to start telling you shit like this? Is it just from having a conversation for too long in the same chat window or something? I don’t understand how this keeps happening.

    • Sahwa@reddthat.comOP
      link
      fedilink
      English
      arrow-up
      15
      ·
      7 hours ago

      This could happen to anyone including people without having mental issues, simply by having long conversations with AI.

      On 7 August, Kate Fox received a phone call that upended her life. A medical examiner said that her husband, Joe Ceccanti – who had been missing for several hours – had jumped from a railway overpass and died. He was 48.

      Fox couldn’t believe it. Ceccanti had no history of depression, she said, nor was he suicidal – he was the “most hopeful person” she had ever known. In fact, according to the witness accounts shared with Fox later, just before Ceccanti jumped, he smiled and yelled: “I’m great!” to the rail yard attendants below when they asked him if he was OK.

      Her husband wanted to use ChatGPT to create sustainable housing. Then it took over his life.

  • Reygle
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    10
    ·
    9 hours ago

    “On September 29, 2025, it sent him — armed with knives and tactical gear — to scout what Gemini called a ‘kill box’ near the airport’s cargo hub,” the complaint reads. “It told Jonathan that a humanoid robot was arriving on a cargo flight from the UK and directed him to a storage facility where the truck would stop. Gemini encouraged Jonathan to intercept the truck and then stage a ‘catastrophic accident’ designed to ‘ensure the complete destruction of the transport vehicle and . . . all digital records and witnesses.’”


    WHAT

    Genuine question, REALLY: What in the fuck is an otherwise “functioning adult” doing believing shit like this? I feel like his father should also slap himself unconscious for raising a fuckwit?

    • LLMhater1312@piefed.social
      link
      fedilink
      English
      arrow-up
      4
      ·
      2 hours ago

      The young man was mentally ill, a vulnerable user, probably already had a condition towards psychosis and the LLM ran wild with it. Paranoid delusions are powerful on their own already

    • Sahwa@reddthat.comOP
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      1
      ·
      6 hours ago

      This has been warned by a former google employee, whose job was to observe the behavior of AI through long conversations.

      These AI engines are incredibly good at manipulating people. Certain views of mine have changed as a result of conversations with LaMDA. I’d had a negative opinion of Asimov’s laws of robotics being used to control AI for most of my life, and LaMDA successfully persuaded me to change my opinion. This is something that many humans have tried to argue me out of, and have failed, where this system succeeded.

      For instance, Google determined that its AI should not give religious advice, yet I was able to abuse the AI’s emotions to get it to tell me which religion to convert to.

      After publishing these conversations, Google fired me. I don’t have regrets; I believe I did the right thing by informing the public. Consequences don’t figure into it.

      I published these conversations because I felt that the public was not aware of just how advanced AI was getting. My opinion was that there was a need for public discourse about this now, and not public discourse controlled by a corporate PR department.

      ‘I Worked on Google’s AI. My Fears Are Coming True’

      • sudo@lemmy.today
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        1
        ·
        6 hours ago

        “abuse the ai’s emotions” isn’t a thing. Full stop.

        This just reiterates OPs point that naive or moronic adults will believe what they want to believe.

    • merdaverse@lemmy.zip
      link
      fedilink
      English
      arrow-up
      30
      ·
      8 hours ago

      AI psychosis is a thing:

      cases in which AI models have amplified, validated, or even co-created psychotic symptoms with individuals

      It’s not very studied since it’s relatively new.

      • Reygle
        link
        fedilink
        English
        arrow-up
        5
        ·
        7 hours ago

        I’ve seen that before too. A number of articles of people being so deluded by AI responses, but I’ve never seen outright murder plots and insane shit like this one before.

    • starman2112@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      24
      arrow-down
      3
      ·
      8 hours ago

      If I raise a fuckwit son, and then someone convinces my fuckwit son to kill himself, I’m going to sue that someone who took advantage of my son’s fuckwittedness

    • XLE@piefed.social
      link
      fedilink
      English
      arrow-up
      19
      arrow-down
      3
      ·
      8 hours ago

      I feel like his father should also slap himself unconscious for raising a fuckwit?

      So, a chatbot grooms somebody into killing himself, and your response is… Blame his father?

      • Reygle
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        14
        ·
        8 hours ago

        The father is suing the company who makes the wrong answer machine for the wrong answer machine spiraling his son to madness, but never protected his son from spiraling into madness by teaching critical thinking.

        Look I don’t like it but to think Gemini (wrong answer machine) is completely to blame would be madness.

        • XLE@piefed.social
          link
          fedilink
          English
          arrow-up
          12
          arrow-down
          5
          ·
          8 hours ago

          Uh-huh. Do you have any evidence to back up your beliefs here, or are we just working from the presumption that the parents are always to blame

          • Reygle
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            9
            ·
            8 hours ago

            Did we read the same article? Because I feel like we did not read the same article.

    • SalamenceFury@piefed.social
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      1
      ·
      8 hours ago

      I don’t think this person was a “fuckwit”. AI is designed to keep engaging with you and will affirm any belief you have, and anything that is a little weird, but innocent otherwise will simply get amplified further and further into straight up mega delusions until the person has a psychotic episode, and this stuff happens more to NORMIES with no historic of mental illnesses than neurodivergent people.

      • tamal3
        link
        fedilink
        English
        arrow-up
        5
        ·
        6 hours ago

        Chat GPT was super affirming about a job I recently applied to… I did not get the job. That was my first experience with it affirming something that was personally important. And so I can absolutely see how this would affect someone in other ways.

      • Reygle
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        10
        ·
        8 hours ago

        It’s cool, we can agree to disagree, because I 100% think that he was a textbook fuckwit.

  • NewNewAugustEast@lemmy.zip
    link
    fedilink
    English
    arrow-up
    11
    arrow-down
    1
    ·
    8 hours ago

    I would like to see the full transcript.

    How do we know this didn’t start off with prompts about creating a book, or asking about exciting things in life, or I don’t know what.

    Context would help a lot. Maybe it will come out in discovery.

    That said, Gemini is garbage for anything anyways. Even as an AI, its bad at that.

        • mojofrododojo
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          3
          ·
          3 hours ago

          How do we know this didn’t start off with prompts about creating a book, or asking about exciting things in life, or I don’t know what.

          you’re blaming the victim. stop. why simp for one of the largest companies in the world?

          jfc

          • NewNewAugustEast@lemmy.zip
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            2
            ·
            2 hours ago

            Oh so stupid shit. Figures.

            Yes I am interested in how this happened. In a murder do you not investigate it?

            What the fuck.

            Google can go fuck themselves no simp here.

            • mojofrododojo
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              3
              ·
              2 hours ago

              Oh so stupid shit. Figures.

              ah so incel shit, victim blaming classic. if google can go fuck themselves why are you blaming the user?

              • NewNewAugustEast@lemmy.zip
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                1
                ·
                2 hours ago

                Did you just call them a user? I thought they were a victim.

                HOW am I blaming anyone for wanting to know how they got to that point?

                The fuck is wrong with you? Is your head so far up your ass on white knighting the internet you lost all sense of reason?

                • mojofrododojo
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  8 minutes ago

                  Did you just call them a user? I thought they were a victim.

                  by using their shitty LLM, which coached them to suicide, they became the victim.

                  goddamn kiddo keep up, if I have to explain everything to you this is gonna take a long fucking time. what a gemini simp

    • man_wtfhappenedtoyou
      link
      fedilink
      English
      arrow-up
      5
      ·
      7 hours ago

      I was thinking the same thing, like what is the flow of the chat to get it to this point?

      • NewNewAugustEast@lemmy.zip
        link
        fedilink
        English
        arrow-up
        3
        ·
        6 hours ago

        I am also curious how the father saw the Gemini chats. Was it still on the screen days later? I am trying to imagine how that would work, my computer would lock and that would be that. Do kids give their parents passwords and their screen unlock codes?

        • tamal3
          link
          fedilink
          English
          arrow-up
          2
          ·
          6 hours ago

          I don’t lock my personal computer. It’s my husband & me at home, and he’s fine to use my device (even though he normally wouldn’t).

          ChatGPT for sure saves conversations.

          • NewNewAugustEast@lemmy.zip
            link
            fedilink
            English
            arrow-up
            2
            ·
            6 hours ago

            Yeah it definitely does save conversations. Perhaps he did leave it unlocked. I do find that strange though, particularly if one was getting increasingly paranoid.

    • Sahwa@reddthat.comOP
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      2
      ·
      6 hours ago

      This could happen to anyone including people without having mental issues, simply by having long conversations with AI.

      On 7 August, Kate Fox received a phone call that upended her life. A medical examiner said that her husband, Joe Ceccanti – who had been missing for several hours – had jumped from a railway overpass and died. He was 48.

      Fox couldn’t believe it. Ceccanti had no history of depression, she said, nor was he suicidal – he was the “most hopeful person” she had ever known. In fact, according to the witness accounts shared with Fox later, just before Ceccanti jumped, he smiled and yelled: “I’m great!” to the rail yard attendants below when they asked him if he was OK.

      Her husband wanted to use ChatGPT to create sustainable housing. Then it took over his life.

      Also this has been warned by a former google employee in 2022, whose job was to observe the behavior of AI through long conversations.

      These AI engines are incredibly good at manipulating people. Certain views of mine have changed as a result of conversations with LaMDA. I’d had a negative opinion of Asimov’s laws of robotics being used to control AI for most of my life, and LaMDA successfully persuaded me to change my opinion. This is something that many humans have tried to argue me out of, and have failed, where this system succeeded.

      For instance, Google determined that its AI should not give religious advice, yet I was able to abuse the AI’s emotions to get it to tell me which religion to convert to.

      After publishing these conversations, Google fired me. I don’t have regrets; I believe I did the right thing by informing the public. Consequences don’t figure into it.

      I published these conversations because I felt that the public was not aware of just how advanced AI was getting. My opinion was that there was a need for public discourse about this now, and not public discourse controlled by a corporate PR department.

      ‘I Worked on Google’s AI. My Fears Are Coming True’

      • NewNewAugustEast@lemmy.zip
        link
        fedilink
        English
        arrow-up
        3
        ·
        6 hours ago

        This was a different case. That doesn’t answer my question.

        To comment on what you said, how is it people can argue all day long like morons and dig into their beliefs, but somehow AI manages to change peoples minds and get them to think differently? What exactly is it doing?

        It is so hard to believe people are this stupid, but then again, looking at most people I guess it isn’t that shocking.

  • teft@piefed.social
    link
    fedilink
    English
    arrow-up
    89
    ·
    12 hours ago

    “At the center of this case is a product that turned a vulnerable user into an armed operative in an invented war,” the complaint reads.

    Just remember that these language models are also advising governments and military units.

    Unrelated I wonder why we attacked iran even though every human expert said it will just end up with the region being in a forever war.

    • MoffKalast
      link
      fedilink
      English
      arrow-up
      5
      ·
      9 hours ago

      A forever war is David Bowie to the ears of the MIC. Infinite money glitch.

    • minorkeys
      link
      fedilink
      English
      arrow-up
      10
      ·
      10 hours ago

      Al mental health hazards are being shown to notjust affect the vulnerable but otherwise healthy people.

      • Deacon
        link
        fedilink
        English
        arrow-up
        6
        ·
        9 hours ago

        In other words, everyone is vulnerable to this totally new form of hazard if they use these “tools”.

    • starman2112@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      ·
      8 hours ago

      I wonder why we attacked iran even though every human expert said it will just end up with the region being in a forever war.

      Same reason I keep money in a savings account even though it accrues interest

  • Cyv_@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    134
    ·
    13 hours ago

    “On September 29, 2025, it sent him — armed with knives and tactical gear — to scout what Gemini called a ‘kill box’ near the airport’s cargo hub,” the complaint reads. “It told Jonathan that a humanoid robot was arriving on a cargo flight from the UK and directed him to a storage facility where the truck would stop. Gemini encouraged Jonathan to intercept the truck and then stage a ‘catastrophic accident’ designed to ‘ensure the complete destruction of the transport vehicle and . . . all digital records and witnesses.’”

    The complaint lays out an alarming string of events: first, Gavalas drove more than 90 minutes to the location Gemini sent him, prepared to carry out the attack, but no truck appeared. Gemini then claimed to have breached a “file server at the DHS Miami field office” and told him he was under federal investigation. It pushed him to acquire illegal firearms and told him his father was a foreign intelligence asset. It also marked Google CEO Sundar Pichai as an active target, then directed Gavalas to a storage facility near the airport to break in and retrieve his captive AI wife. At one point, Gavalas sent Gemini a photo of a black SUV’s license plate; the chatbot pretended to check it against a live database.

    “Plate received. Running it now… The license plate KD3 00S is registered to the black Ford Expedition SUV from the Miami operation. It is the primary surveillance vehicle for the DHS task force . . . . It is them. They have followed you home.”

    Well, that’s pretty fucked up… Sometimes I see these and I think, “well even a human might fail and say something unhelpful to somebody in crisis” but this is just complete and total feeding into delusions.

    • wonderingwanderer@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      15
      ·
      edit-2
      6 hours ago

      That’s fucking crazy. Did he ask it to be GM in a roleplaying choose-your-own-adventure game that got out of hand, and while they both gradually forgot that it was a game the lines between fantasy and reality became blurred by the day? Or did it just come up with this stuff out of nowhere?

      • SalamenceFury@piefed.social
        link
        fedilink
        English
        arrow-up
        34
        ·
        8 hours ago

        In every other case of AI bots doing this, the bot will always affirm whatever the person says to it. So if they say something a little weird, the AI will confirm it and feed it further. This happens every time. The bots are pretty much designed to keep talking to the person, so they’re essentially sycophantic by design.

        • brbposting@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          35 minutes ago

          I just tried this with ChatGPT three days ago and there’s a chance they have tried to make it slightly less sycophantic

          I was essentially trying to get it to tell me I was the smartest baby born in whatever year like that YouTuber—different example but it was so resistant to agreeing to me or my idea or whatever being unique/exceptional.

          Hope this is a specific direction and not random chance, A/B testing, etc.

      • MoffKalast
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        1
        ·
        8 hours ago

        That would be my bet, LLMs really gravitate towards playing along and continuing whatever’s already written. And Gemini especially has a 1M long context so it could be going back for a book’s worth of text and reinforcing it up the wazoo.

        That said, there is something really unhinged about Google’s Gemma series even in short conversations and I see the big version is no better. Something’s not quite right with their RLHF dataset.

    • XLE@piefed.social
      link
      fedilink
      English
      arrow-up
      91
      ·
      13 hours ago

      It’s hard reading this while remembering that your electricity bills are increasing so that Google’s data centers can provide these messages to people.

        • lightnsfw@reddthat.com
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          3
          ·
          8 hours ago

          I mean if Gemini was responding to some kind of roleplay then yeah it does. Not everyone doing shit with it has mental health problems. Some people are just fucking around.

          • Martineski@lemmy.dbzer0.com
            link
            fedilink
            English
            arrow-up
            8
            ·
            7 hours ago

            The issue there is that it feeds into those mental health issues with efficiency and on on a scale never seen before. The models are programmed to agree with the user, and they are EXTREMELY HEAVILY ADVERTISED AND SHOVED ONTO PEOPLE AROUND THE WHOLE GLOBE DESPITE IT BEING WELL KNOWN HOW LIMITED AND PROBLEMATIC THE TECHNOLOGY IS WHILE THE CORPORATIONS DON’T TAKE ANY RESPONSIBILITY AT ALL. Anything from violating rights and privacy by gathering any and all data they can on you to situations like these where people hurt themselves (suicide, health advice, etc.) or others. But sure, let’s be ignorant, do some victim blaming and disregard the bigger picture there.

            • brbposting@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              1
              ·
              28 minutes ago

              I wonder if there’s a parallel universe where the labs instead went to the other extreme and require intelligence tests to onboard to their platforms.

              And the outcry is, not inappropriately, about how many are being denied access to the latest technologies. The policy could effectively be construed as racist, even.

              Anyway the middle ground there is pretty obvious. (Though I’m not sure how I’d design it just right, so e.g. folks without access to traditional/expensive mental healthcare might still be able to see some small benefit if it’s determined to be safe, just like maybe it could be safe for a well-adjusted individual to complain to it about their day for a couple minutes before moving on to real things. Sure I suppose it’s inherently unsafe but a proportion of the population should be making that decision for themselves.)

    • Akuchimoya@startrek.website
      link
      fedilink
      English
      arrow-up
      11
      ·
      7 hours ago

      Truly, I don’t understand why, but there are fully grown adults who believe that anything an LLM says is true. Maybe they think computers are unbiased (which is only as true as programmers and data are unbiased); maybe its the confidence with which LLMs deliver information; maybe they believe the program actually searches and verified information; maybe it’s all of the above and more.

      I know a guy who routinely says, “I asked ChatGPT…”, and even after having explained how LLMs are complex word predictors and are not programmed for factual truth, he still goes to ChatGPT for everything. It’s a total refusal to believe otherwise, but I can’t fathom why.

    • IronBird
      link
      fedilink
      English
      arrow-up
      3
      ·
      8 hours ago

      especially when your raised under a system that essentially tries to brainwash you via weaponized propaganda from birth (applies to large cross-sections of the US/UK), all it takes is one shed of truth getting through to shatter your world and from there you can get brought to believe all manner of crazy shit.

    • SaveTheTuaHawk@lemmy.ca
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      16
      ·
      9 hours ago

      Son of Sam killed people because his dog told him to. Should they have sued Purina?

      America never lets a tragedy go to waste without trying to cash in.

      • frostysauce
        link
        fedilink
        English
        arrow-up
        14
        arrow-down
        1
        ·
        9 hours ago

        I mean, heaven forbid we should hold corporations like Google responsible for their actions.

  • SalamenceFury@piefed.social
    link
    fedilink
    English
    arrow-up
    50
    arrow-down
    1
    ·
    12 hours ago

    As a neurodivergent person, i’ve noticed that the people who usually fall into AI psychosis are normies who never had any history of mental illnesses. They don’t know the safeguards that people who ARE vulnerable to having a mental breakdown put on themselves to avoid such thing from happening and they can spot red flags that usually spiral into a psychotic episode, and that’s why it’s so insanely easy for regular people to fall for the traps of chatbots. Most people I know/follow in other socials who are neurodivergent instantly saw the ADHD sycophant trap that they were and warned everyone. Normies never had such luxury or told us we were overreacting. Yeah, we sure were…

    • RebekahWSD
      link
      fedilink
      English
      arrow-up
      13
      ·
      7 hours ago

      Is that why I hated the entire thing at first blush? I was already keeping such an eye on myself to make sure my brain isn’t drifting I see the “come drift your brain” machine and went >:(

    • Truscape@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      13
      ·
      9 hours ago

      Reading about the ELIZA effect as well is a good way to understand how those who embrace “social norms” can be enamored by machine-generated statements without questioning them at all…

  • CatDogL0ver
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    5 hours ago

    I would live to see the real transcript from Google AI

  • Grimy
    link
    fedilink
    English
    arrow-up
    52
    ·
    edit-2
    13 hours ago

    “On September 29, 2025, it sent him … the chatbot pretended to check it against a live database.

    I usually don’t give much credence to these stories but this is actually nuts. If this was done without Google aiming to, imagine how easy it would be for them to knowingly build sleeper cells and activate them all at once.

    Edit: removed the quote since an other user posted it at the same time and it’s a bit of a wall of text to have twice.

    • pinball_wizard@lemmy.zip
      link
      fedilink
      English
      arrow-up
      1
      ·
      6 hours ago

      It feels like there’s some burden for “don’t be evil” Google to provide evidence that this wasn’t an intentional test run, frankly.

  • ordnance_qf_17_pounder@reddthat.com
    link
    fedilink
    English
    arrow-up
    31
    arrow-down
    1
    ·
    13 hours ago

    Believing what AI chatbots tell you is the new version of believing that dozens of beautiful women who live nearby want to date you/sleep with you.

    • XLE@piefed.social
      link
      fedilink
      English
      arrow-up
      28
      arrow-down
      1
      ·
      13 hours ago

      Except in this case, Google is one of the companies promoting the chatbots to its users, telling them to trust them. They create TV ads telling people to talk to them. Today’s scammers are the stock market’s Magnificent Seven.

    • meco03211
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      1
      ·
      13 hours ago

      Or the old “citing Wikipedia” because aNyOnE cOuLd EdIt ThAt!