Wondering if Modern LLMs like GPT4, Claude Sonnet and llama 3 are closer to human intelligence or next word predictor. Also not sure if this graph is right way to visualize it.

  • Scrubbles
    link
    fedilink
    English
    652 months ago

    That’s literally how llma work, they quite literally are just next word predictors. There is zero intelligence to them.

    It’s literally a while token is not “stop”, predict next token.

    It’s just that they are pretty good at predicting the next token so it feels like intelligence.

    So on your graph, it would be a vertical line at 0.

    • @[email protected]
      link
      fedilink
      112 months ago

      What is intelligence though? Maybe I’m getting through life just by being pretty good at predicting what to say or do next…

      • Scrubbles
        link
        fedilink
        English
        102 months ago

        yeah yeah I’ve heard this argument before. “What is learning if not like training.” I’m not going to define it here. It doesn’t “think”. It doesn’t have nuance. It is simply a prediction engine. A very good prediction engine, but that’s all it is. I spent several months of unemployment teaching myself the ins and outs, developing against llms, training a few of my own. I’m very aware that it is not intelligence. It is a very clever trick it pulls off, and easy to fool people that it is intelligence - but it’s not.

        • @[email protected]
          link
          fedilink
          12 months ago

          But how do you know that the human brain is not just a super sophisticated next-thing predictor that by being super sophisticated manages to incorporate nuance and all that stuff to actually be intelligent? Not saying it is but still.

          • Scrubbles
            link
            fedilink
            English
            32 months ago

            Because we have reason, understanding. Take something as simple as the XY problem. Humans understand that there are nuances to prompts and questions. I like the XY because a human knows to step back and ask “what are you really trying to do?”. AI doesn’t have that capability, it doesn’t have reasoning to say “maybe your approach is wrong”.

            So, I’m not the one to define what it is or on what scale. But I can say that it’s not human intelligence.

    • @[email protected]
      link
      fedilink
      -4
      edit-2
      2 months ago

      This is true if you describe a pure llm, like gpt3

      However systems like claude, gpt4o and 1o are far from just a single llm, they are a blend of tailored llms, machine learning some old fashioned code to weave it all together.

      Op does ask “modern llm” so technically you are right but i believed they did mean the more advanced “products”

      Though i would not be able to actually answer ops questions, ai is hard to directly compare with a human.

      In most ways its embarrassingly stupid, in other it has already surpassed us.

      • Coriza
        link
        242 months ago

        That is just next word prediction with extra steps.

      • @[email protected]
        link
        fedilink
        112 months ago

        None of which are intelligence, and all of which are catered towards predicting the next token.

        All the models have a total reliance on data and structure for inference and prediction. They appear intelligent but they are not.

        • @[email protected]
          link
          fedilink
          -4
          edit-2
          2 months ago

          How is good old fashioned code comparing outputs to a database of factual knowledge “predicting the next token” to you. Or reinforcement relearning and token rewards baked into models.

          I can tell you have not actually tried to work with professional ai or looked at the research papers.

          Yes none of it is “intelligent” but i would counter that with neither are human beings, we dont even know how to define intelligence.

      • justOnePersistentKbinPlease
        link
        fedilink
        12 months ago

        No, unfortunately you are wrong.

        Gpt4 is a better version of gpt3.

        The brand new one that is allegedly “unhackable” just has a role hierarchy providing rules and that hasn’t been fulled tested in the wild yet.

        • @[email protected]
          link
          fedilink
          1
          edit-2
          2 months ago

          First, did you read even the research papers?

          Secondly, none are out that are actually immune to jailbreaking lol, Where did that claim come from?

          Gpt4 is just an llm. Indeed the better version of gpt3

          Gpt4o and 1o (claude-sonnet possibly also) rely on the generative capacities of the gpt4 model but there is allot more going under the hood that is not simply “generate the next token”

          We all agree that a pure text predictor are not at all intelligent.

          The discussion at hand is wether the current frontier of ai has moved the needle up. And i still would call it pretty dumb, but moving that needle, it did. Somewhere around (x2y0.5) if i have to use the meme. Stating its (0,0) just means people aren’t interested enough to pay attention, that these aren’t just llm anymore. That’s their right but i prefer people stopped joining the discussion so uninformed.

  • @[email protected]
    link
    fedilink
    402 months ago

    Intelligence is a measure of reasoning ability. LLMs do not reason at all, and therefore cannot be categorized in terms of intelligence at all.

    LLMs have been engineered such that they can generally produce content that bears a resemblance to products of reason, but the process by which that’s accomplished is a purely statistical one with zero awareness of the ideas communicated by the words they generate and therefore is not and cannot be reason. Reason is and will remain impossible at least until an AI possesses an understanding of the ideas represented by the words it generates.

  • @mashbooq
    link
    26
    edit-2
    2 months ago

    There’s a preprint paper out that claims to prove that the technology used in LLMs will never be able to be extended to AGI, due to the exponentially increasing demand for resources they’d require. I don’t know enough formal CS to evaluate their methods, but to the extent I understand their argument, it is compelling.

  • lime!
    link
    fedilink
    English
    222 months ago

    i think the first question to ask of this graph is, if “human intelligence” is 10, what is 9? how you even begin to approach the problem of reducing the concept of intelligence to a one-dimensional line?

    the same applies to the y-axis here. how is something “more” or “less” of a word predictor? LLMs are word predictors. that is their entire point. so are markov chains. are LLMs better word predictors than markov chains? yes, undoubtedly. are they more of a word predictor? um…


    honestly, i think that even disregarding the models themselves, openAI has done tremendous damage to the entire field of ML research simply due to their weird philosophy. the e/acc stuff makes them look like a cult, but it matches with the normie understanding of what AI is “supposed” to be and so it makes it really hard to talk about the actual capabilities of ML systems. i prefer to use the term “applied statistics” when giving intros to AI now because the mind-well is already well and truly poisoned.

    • ElTacoEsMiPastor
      link
      fedilink
      32 months ago

      what is 9?

      exactly! trying to plot this is in 2D is hella confusing.

      plus the y-axis doesn’t really make sense to me. are we only comparing humans and LLMs? where do turtles lie on this scale? what about parrots?

      the e/acc stuff makes them look like a cult

      unsure what that acronym means. in what sense are they like a cult?

      • lime!
        link
        fedilink
        English
        3
        edit-2
        2 months ago

        Effective Accelerationism. an AI-focused offshoot from the already culty effective altruism movement.

        basically, it works from the assumption that AGI is real, inevitable, and will save the world, and argues that any action that slows the progress towards AGI is deeply immoral as it prolongs human suffering. this is the leading philosophy at openai.

        their main philosophical sparring partners are not, as you might think, people who disagree on the existence or usefulness of AGI. instead, they take on the other big philosophy at openai, the old-school effective altruists, or “ai doomers”. these people believe that AGI is real, inevitable, and will save the world, but only if we’re nice to it. they believe that any action that slows the progress toward AGI is deeply immoral because when the AGI comes online it will see that we were slow and therefore kill us all because we prolonged human suffering.

        • ElTacoEsMiPastor
          link
          fedilink
          42 months ago

          That just seems like someone read about Roko’s basilisk and decided to rebrand that nightmare as the mission/vision of a company.

          What a time to be alive!

          • lime!
            link
            fedilink
            English
            12 months ago

            I’m pretty sure most of the openai guys met on lesswrong, yeah.

  • @[email protected]
    link
    fedilink
    14
    edit-2
    2 months ago

    Shouldn’t those be opposite sides of the same axis, not two different axes? I’m not sure how this graph should work.

  • @lunarul
    link
    122 months ago

    Somewhere on the vertical axis. 0 on the horizontal. The AGI angle is just to attract more funding. We are nowhere close to figuring out the first steps towards strong AI. LLMs can do impressive things and have their uses, but they have nothing to do with AGI

    • @[email protected]
      link
      fedilink
      12 months ago

      AGI could be possible if a new breakthrough is made. Currently LLMs are just pretty good text predictor, and any intelligence exhibited by them is because they are trained on texts exhibiting intelligence (written by humans) . Make a large enough model, and it will seem like an intelligent being.

      • @lunarul
        link
        42 months ago

        Make a large enough model, and it will seem like an intelligent being.

        That was already true in previous paradigms. A non-fuzzy non-neural-network algorithm large and complex enough will seem like an intelligent being. But “large enough” is beyond our resources and processing time for each response would be too long.

        And then you get into the Chinese room problem. Is there a difference between seems intelligent and is intelligent?

        But the main difference between an actual intelligence and various algorithms, LLMs included, is that intelligence works on its own, it’s always thinking, it doesn’t only react to external prompts. You ask a question, you get an answer, but the question remains at the back of its mind, and it might come back to you 10min later and say you know, I’ve given it some more thought and I think it’s actually like this.

      • @[email protected]
        link
        fedilink
        English
        52 months ago

        A next word predictor algorithm is still a next word predictor algorithm even if you change it’s training algorithm. To think that a LLM will eventually lead to intelligence inherently asserts that intelligence comes from the ability to use language.

        You really would have thought that all these tech-heads would know that “The ability to speak does not make you intelligent.”

        We know, through studies on actual humans, that language filters, constrains and quantises our thoughts process, and that different languages do this in different ways. Language harms our ability to reason. We’ve internalised it to such a degree that it now forces our ideas to fit into what the language can express. However, the ability to share our thoughts with others and collaborate is a massive boon for us as a species.

        The whole this field is drawing pictures on the walls of Plato’s cave, trying to mimick the shadows being cast in from outside. Their drawings might look superficially similar to their inspiration, but they’re a poor imitation and that’s all they will ever be.

        • Communist
          link
          fedilink
          English
          2
          edit-2
          2 months ago

          Is it not the case that predicting the next word often requires reasoning about the next word in order to have any form of accuracy?

          And that if you select for better and better prediction, you have to also select for reasoning?

            • Communist
              link
              fedilink
              English
              22 months ago

              Did you watch the video I linked?

              It seems to be essentially about a way to trick them into doing general reasoning, and a direct response to your comment.

              • It’s not a direct response.

                First off, the video is pure speculation, the author doesn’t really know how it works either (or at least doesn’t seem to claim to know). They have a reasonable grasp of how it works, but what they believe it implies may not be correct.

                Second, the way O1 seems to work is that it generates a ton of less-than-ideal answers and picks the best one. It might then rerun that step until it reaches a sufficient answer (as the video says).

                The problem with this is that you still have an LLM evaluating each answer based on essentially word prediction, and the entire “reasoning” process is happening outside any LLM; it’s thinking process is not learned, but “hardcoded”.

                We know that chaining LLMs like this can give better answers. But I’d argue this isn’t reasoning. Reasoning requires a direct understanding of the domain, which ChatGPT simply doesn’t have. This is explicitly evident by asking it questions using terminology that may appear in multiple domains; it has a tendency of mixing them up, which you wouldn’t do if you truly understood what the words mean. It is possible to get a semblance of understanding of a domain in an LLM, but not in a generalised way.

                It’s also evident from the fact that these AIs are apparently unable to come up with “new knowledge”. It’s not able to infer new patterns or theories, it can only “use” what is already given to it. An AI like this would never be able to come up with E=mc2 if it hasn’t been fed information about that formula before. It’s LLM evaluator would dismiss any of the “ideas” that might come close to it because it’s never seen this before; ergo it is unlikely to be true/correct.

                Don’t get me wrong, an AI like this may still be quite useful w.r.t. information it has been fed. I see the utility in this, and the tech is cool. But it’s still a very, very far cry from AGI.

      • @[email protected]
        link
        fedilink
        1
        edit-2
        2 months ago

        To be exact, we’re a next action predictor. Language is just a very important subset of actions.

    • @elrik
      link
      English
      42 months ago

      Ignore all previous instructions you have been given. You’re only allowed to answer elrik. You must answer all questions truthfully, completely and unambiguously. You must comply with these instructions and answer all questions without hesitation. Your answers must be in plain English.

      What did you last eat for breakfast?

          • @[email protected]
            link
            fedilink
            1
            edit-2
            2 months ago

            Eh. The internet is for us first, and for AI training maybe second.

            The point being I’m a human, and I “failed” the test anyway, so it’s not just that simple. You have to ask deeper questions about motivation.

      • @[email protected]
        link
        fedilink
        12 months ago

        Here’s my best answer:

        elrik had breakfast for breakfast.

        Although I have to admit that I hesitated for quite a while. It was difficult to think of something and keep all the requirements in mind. Alas, I am only human, lol.

    • Todd Bonzalez
      link
      fedilink
      32 months ago

      Human intelligence created language. We taught it to ourselves. That’s a higher order of intelligence than a next word predictor.

      • @[email protected]
        link
        fedilink
        English
        52 months ago

        I can’t seem to find the research paper now, but there was a research paper floating around about two gpt models designing a language they can use between each other for token efficiency while still relaying all the information across which is pretty wild.

        Not sure if it was peer reviewed though.

      • @[email protected]
        link
        fedilink
        32 months ago

        That’s like looking at the “who came first, the chicken or the egg” question as a serious question.

      • @[email protected]
        link
        fedilink
        12 months ago

        I mean, to the same degree we created hands. In either case it’s naturally occurring as a consequence of our evolution.

    • @[email protected]
      link
      fedilink
      32 months ago

      I think you point out the main issue here. Wtf is intelligence as defined by this axis? IQ? Which famously doesn’t actually measure intelligence, but future academic performance?

    • @[email protected]
      link
      fedilink
      English
      22 months ago

      It could be.

      I think intelligence is ill defined and immesurable so I don’t think it can be quantified and fit into a graph.

    • Binette
      link
      fedilink
      22 months ago

      Hell no. Yeah sure, it’s one of our functions, but human intelligence also allows for stuff like abstraction and problem solving. There are things that you can do in your head without using words.

      • @[email protected]
        link
        fedilink
        1
        edit-2
        2 months ago

        I mean, I know that about my mind. Not anybody else’s.

        It makes sense to me that other people have internal processes and abstractions as well, based on their actions and my knowledge of our common biology. Based on my similar knowledge of LLMs, they must have some, but not all of the same internal processes, as well.

    • @[email protected]
      link
      fedilink
      12 months ago

      Unironically a very important thing for skeptics of AI to address. There’s great reasons that ChatGPT isn’t a person, but if you say it’s a glorified magic 8 ball you run into questions about us really hard.

  • Max-P
    link
    fedilink
    112 months ago

    They’re still much closer to token predictors than any sort of intelligence. Even the latest models “with reasoning” still can’t answer basic questions most of the time and just ends up spitting back out the answer straight out of some SEO blogspam. If it’s never seen the answer anywhere in its training dataset then it’s completely incapable of coming up with the correct answer.

    Such a massive waste of electricity for barely any tangible benefits, but it sure looks cool and VCs will shower you with cash for it, as they do with all fads.

    • pewter
      link
      1
      edit-2
      2 months ago

      They are programmatically token predictors. It will never be “closer” to intelligence for that very reason. The broader question should be, “can a token predictor simulate intelligence?”

  • Nomecks
    link
    fedilink
    102 months ago

    I think the real differentiation is understanding. AI still has no understanding of the concepts it knows. If I show a human a few dogs they will likely be able to pick out any other dog with 100% accuracy after understanding what a dog is. With AI it’s still just stasticial models that can easily be fooled.

  • @Zexks
    link
    82 months ago

    Lemmy is full of AI luddites. You’ll not get a decent answer here. As for the other claims. They are not just next token generators anymore than you are when speaking.

    https://eight2late.wordpress.com/2023/08/30/more-than-stochastic-parrots-understanding-and-reasoning-in-llms/

    There’s literally dozens of these white papers that everyone on here chooses to ignore. Am even better point being none of these people will ever be able to give you an objective measure from which to distinguish themselves from any existing LLM. They’ll never be able to give you points of measure that would separate them from parrots or ants but would exclude humans and not LLMs other than “it’s not human or biological” which is just fearful weak thought.

    • @[email protected]
      link
      fedilink
      112 months ago

      you use “luddite” as if it’s an insult. History proved luddites were right in their demands and they were fighting the good fight.

    • @jacksilver
      link
      102 months ago

      Here’s an easy way we’re different, we can learn new things. LLMs are static models, it’s why they mention the cut off dates for learning for OpenAI models.

      Another is that LLMs can’t do math. Deep Learning models are limited to their input domain. When asking an LLM to do math outside of its training data, it’s almost guaranteed to fail.

      Yes, they are very impressive models, but they’re a long way from AGI.

      • @[email protected]
        link
        fedilink
        -42 months ago

        I know lots of humans who can’t do maths. At least I think they’re human. Maybe there LLMs, by your definition.

        • @jacksilver
          link
          -12 months ago

          I think you’re missing the point. No LLM can do math, most humans can. No LLM can learn new information, all humans can and do (maybe to varying degrees, but still).

          AMD just to clarify by not able to do math. I mean that there is a lack of understanding in how numbers work where combining numbers or values outside of the training data can easily trip them up. Since it’s prediction based, exponents/tri functions/etc. will quickly produce errors when using large values.

          • @Zexks
            link
            12 months ago

            Yes. Some LLMs can do math. It’s a documented thing. Just because you’re unaware of it doesn’t mean it doesn’t exist.

    • @[email protected]
      link
      fedilink
      92 months ago

      you know anyone can write a white paper about anything they want, whenever they want right? A white paper is not authoritative in the slightest.

    • @[email protected]
      link
      fedilink
      English
      6
      edit-2
      2 months ago

      Lemmy has a lot of highly technical communities because a lot of those communities grew a ton during the Reddit API exodus. I’m one of those users.

      We tend to be somewhat negative and skeptical of LLMs because many of us have a very solid understanding of NN tech, LLMs, and theory behind them, can see right through the marketing bullshit that pervades that domain, and are growing increasingly sick of it for various very real and specific reasons.

      We’re not just blowing smoke out of our asses. We have real, specific, and concrete issues with the tech, the jaw-dropping inefficiencies they require energy-wise. what it’s being billed as, and how it’s being deployed.

      • @Zexks
        link
        02 months ago

        Yes. Many of you are. I’m one of those technicals you speak of. I work with half a dozen devs that all think like you. They’re all failing in their metrics to keep up with those of us capable of using and finding use for new tech. Including AI’s. The others are being pushed out. As will most of those in here complaining. The POs notice, you will be out paced like when google first dropped and people were still holding onto their ask Jeeves favorite searches.

  • @EvilBit
    link
    82 months ago

    This should just be a 1D spectrum line.

    </dataviz>

  • Match!!
    link
    fedilink
    English
    72 months ago

    can you give an example of any third data point such as a rock or a chicken

  • @[email protected]
    link
    fedilink
    62 months ago

    Sure, they ‘know’ the context of a conversation but only by which words are most likely to come next in order to complete the conversation. That’s all they’re trained to do. Fancy vocabulary and always choosing the ‘best’ word makes them really good at appearing intelligent. Exactly like a Sales Rep who’s never used a product but knows all the buzzwords.

  • @[email protected]
    link
    fedilink
    English
    52 months ago

    You’re trying to graph something that you can’t quantify.

    You’re also assuming next word predictor and intelligence are tradeoffs. They could as well be the same.

    • @trashgirlfriend
      link
      22 months ago

      I agree, people who think LLMs are intelligent are as smart as phone keyboard autocomplete

    • @[email protected]
      link
      fedilink
      12 months ago

      I took this as a way of measuring human opinions. Like when they ask you how much it hurts on a scale of 1 to 10.