• GreenKnight23
    link
    fedilink
    arrow-up
    3
    arrow-down
    1
    ·
    2 days ago

    nice astroturfing there schmuck.

    because although LLMs are not good at many things, what they absolutely are good at is taking large data sets of writing and finding a kind of “average” of that data.

    who knew that Large LANGUAGE Models do math (they don’t)

    gtfo of here with your bullshit.

    • BluesF
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      1 day ago

      I’m not talking about numerical data, the way LLMs work is to find a “most likely response” based on the input text. There is absolutely maths happening inside the model, how else do you think they work? I’m not saying they take numbers and find an average.

      • GreenKnight23
        link
        fedilink
        arrow-up
        1
        arrow-down
        1
        ·
        21 hours ago

        LLMs are trained on language based content. it doesn’t know how to extract answers from mathematical based problems. it only gives approximations based on model input. it also can be trained wrong based on user input of data.

        to a purely mathematical logical operator 2+2=4.

        to a LLM if told 2+2=9 it will then always respond with 2+2=9.

        1000003363

        LLMs don’t count because they can’t count. without the ability to count it can never understand the proof behind mathematical formulas.

        • BluesF
          link
          fedilink
          arrow-up
          1
          ·
          2 hours ago

          Yes, I understand that, you are not understanding what I’m describing. I am not talking about taking an average of numerical data. LLMs take something that can be thought of as an “average” of text. It says “given all the text I have seen, and this new text input, what’s the most likely output?” In some numerical contexts the expected value is also an average, LLMs find a similar result, and that is what I am drawing a parallel between here.