Junk websites filled with AI-generated text are pulling in money from programmatic ads::More than 140 brands are advertising on low-quality content farm sites —and the problem is growing fast.

  • @Cabrio
    link
    English
    11
    edit-2
    1 year ago

    ChatGPT is a predictive text engine than can already generate more coherent and accurate information than 90% of Internet Contributers and it doesn’t even have the capacity to add 5+5 together. I welcome our new overlords.

    • @Aceticon
      link
      English
      5
      edit-2
      1 year ago

      LLMs like ChatGTP are accurate only by chance, which is why you can’t really trust the info contained in what they output: if your question ended up in or near a cluster in the “language token N-space” were a good answer is, then you’ll get a good answer, otherwise you’ll get whatever is closest in the language token N-space, which might very well be complete bollocks whilst delivered in the language of absolute certainty.

      It is however likely more coherent that “90% of Internet Contributers” for just generated texts (not if you get to do question and answer though: just ask something from it and if you get a correct answer say that “it’s not correct” and see how it goes).

      This is actually part of the problem: in the stuff outputted by LLMs you can’t really intuit the likely accuracy of a response from the gramatical coherence and word choice of the response itself: it’s like being faced with the greatest politician in the World who is an idiot savant - perfect at memorizing what he/she heard and creating great speeches based on it whilst being a complete total moron at everything else including understanding the meaning of what he or she heard and just reshuffles and repeats to others.

      • @Cabrio
        link
        English
        1
        edit-2
        1 year ago

        Oh I know. 54% of American adults read below a 6th grade comprehension level.

        Most of the answers you get out of them are accurate only by chance, which is why you can’t really trust the info contained in what they output: if your question ended up in or near a cluster in the “educated N-space” where a good answer is, then you get a good answer, otherwise you’ll get whatever is the closest “response N-space”, which is practically guaranteed to be complete bollocks whilst delivered in the language of absolute certainty.

        I’d go on but I’m sure the point is made.