• @BitSound
    link
    01 year ago

    Your concept of a chair is an abstract thought representation of a chair. An LLM has vectors that combine or decompose in some way to turn into the word “chair,” but are not a concept of a chair or an abstract representation of a chair. It is simply vectors and weights, unrelated to anything that actually exists.

    Just so incredibly wrong. Fortunately, I’ll have save myself time arguing with such a misunderstanding. GPT-4 is here to help:

    This reads like a misunderstanding of how LLMs (like GPT) work. Saying an LLM’s understanding is “simply vectors and weights” is like saying our brain’s understanding is just “neurons and synapses”. Both systems are trying to capture patterns in data. The LLM does have a representation of a chair, but it’s in its own encoded form, much like our neurons have encoded representations of concepts. Oversimplifying and saying it’s unrelated to anything that actually exists misses the point of how pattern recognition and information encoding works in both machines and humans.

    • Veraticus
      link
      fedilink
      English
      01 year ago

      Are you kidding me? I sourced GPT4 itself disagreeing with you that it is intelligent and you told me it’s lying. And here you are, using it to try to reinforce your point? Are you for real or is this some kind of complicated game?

        • Veraticus
          link
          fedilink
          English
          -1
          edit-2
          1 year ago

          Here, let’s ask GPT4 itself since you’ve decided it’s suddenly an okay source:

          Your statement is correct in asserting that the vector representation in a language model is not an abstract representation. It’s purely a mathematical construct. However, saying it’s “unrelated to anything that actually exists” might be an overstatement. These vectors do capture statistical patterns in human language, which are reflections of human thought and culture. They’re just not capable of the deep, nuanced understanding that comes from human experience.

          I accept it’s an overstatement. But it is neither “incredibly wrong,” nor is it thought. (Or intelligence.)

          • @[email protected]
            link
            fedilink
            11 year ago

            I’d just like to step in here and mention that asking an LLM is probably not a good proof (and this is directed at both of you). Its understanding of AI is from before it was trained, so it is wildly out of date at this point given how much has happened in the space since.

            • Veraticus
              link
              fedilink
              English
              11 year ago

              GPT4 has knowledge of its own training since it was trained in 2022.

              • @[email protected]
                link
                fedilink
                11 year ago

                Care to provide some proof of that? They did update their system prompt to include a few things like it is now GPT4 (it used to always say GPT3). Other than that, I don’t think it knows anything. But in general, I was more talking about developments in AI since it was trained which it certainly does not know.

          • @BitSound
            link
            11 year ago

            So you admit that you were wrong?

            • Veraticus
              link
              fedilink
              English
              -11 year ago

              I was in this case – but the overall point I made is still correct. If winning this minor battle is what you were seeking, congratulations. You are no closer to understanding the truth of this or what we were actually talking about. Not that that was either your point or within your capabilities.