• @[email protected]
    link
    fedilink
    English
    214 days ago

    But you wouldn’t call it lying if a person tells you something they think is true but turns out to be false. Lying means intentionally giving out false information. LLMs don’t have intentions.

    • ℍ𝕂-𝟞𝟝
      link
      fedilink
      English
      514 days ago

      Yeah I think it’s more fitting to use the term bullshitting.

      LLMs actually know that some of their answers have low probability to be the right ones, they give them out regardless, and don’t mention the low confidence of it.

    • @[email protected]
      link
      fedilink
      English
      314 days ago

      You would call what they said bullshit though.

      Intent is irrelevant. Bullshit is bullshit.

      • @[email protected]
        link
        fedilink
        English
        014 days ago

        Depends which definition of bullshit you use, I guess.

        Frankfurt determines that bullshit is speech intended to persuade without regard for truth. The liar cares about the truth and attempts to hide it; the bullshitter doesn’t care whether what they say is true or false.

        Wiki

        • @[email protected]
          link
          fedilink
          English
          214 days ago

          the bullshitter doesn’t care whether what they say is true or false.

          That’s another way to say “intent is irrelevant”.

          It’s also effectively the perfect definition of LLM output. Content for the sole purpose of looking the part with absolutely no consideration for reality.

          • @[email protected]
            link
            fedilink
            English
            -113 days ago

            …bullshit is speech intended to persuade…

            Quoting out of context is not going to score you any points

            • @[email protected]
              link
              fedilink
              English
              113 days ago

              It is intended (by the designer) to persuade. It’s intended to persuade you that it’s something a human would say.

              Ignoring that you’re trying to claim one dude’s definition of bullshit as the law, that one dude’s definition is an exact flawless match for what LLMs are.

              • @[email protected]
                link
                fedilink
                English
                013 days ago

                It is intended (by the designer) to persuade.

                According to you, I presume? Or can you back that up somehow?

                LLMs were developed to simulate human-like understanding and generation of language. They’re called large language models for a reason.

                • @[email protected]
                  link
                  fedilink
                  English
                  113 days ago

                  No, they weren’t. There was never at any point any theoretical possibility that an LLM would resemble understanding in any way.

                  • @[email protected]
                    link
                    fedilink
                    English
                    0
                    edit-2
                    13 days ago

                    That’s why they simulate it. Just like I said.

                    Look, there’s no point going any further with this. You just keep making baseless claims without any explanation or even attempt to try and convince me otherwise. When called out, you ignore it and move on. I’m not interested in discussions where people are just talking past each other while disregarding everything said in the previous messages. Take care now.

    • @[email protected]
      link
      fedilink
      English
      114 days ago

      …but if they don’t know I expect them to say so. An LLM isn’t trustworthy until it says “I don’t know”.