• BigFig
    link
    English
    526 months ago

    What do you do when ChatGPT just makes shit up or answers incorrectly to yes or no questions, you’d have no way of knowing it was wrong

    • @gridleaf
      link
      English
      156 months ago

      ChatGPT is most useful when you may not know the right answer, but you know a wrong answer when you see one. It’s very useful for technical issues. Much quicker for troubleshooting than searching page after page for a solution.

      • @TheDarkKnight
        link
        English
        26 months ago

        It’s actually great at troubleshooting linux stuff weirdly enough lol

          • Avid Amoeba
            link
            fedilink
            English
            86 months ago

            Yeah that makes sense. The success rate might fall off a cliff in more complex software projects. E.g. applications that require designs beyond 10 UML boxes with hundreds of thousands of lines, especially not written in JS/Python.

    • Otter
      link
      fedilink
      English
      16 months ago

      Not the other commenter:

      I usually have an idea about the thing I’m asking, and if not then I’ll look up the topics mentioned after some guided brainstorming

      I’ve also found that asking the same question again, after resetting the chat, can give you an idea of what is happening

    • @[email protected]
      link
      fedilink
      English
      16 months ago

      While this is an important thing to understand about AI, it’s an overstated issue once understood. For most questions I ask AI, it doesn’t matter if it’s correct as long as it pulls some half useful info to get me on track (i.e programming). For other questions, I only ask it if I need to figure out where to look next, which it will usually do just fine.

      The first page of my search results is all AI generated garbage articles anyway, at least I know what I am getting with GPT and can take it as such.

      • @Womble
        link
        English
        26 months ago

        Yup, as long as you are aware that it could be wrong and look at it critically LLMs at GPT scale are very useful tools. The best way I’ve heard it described is having a lightning fast intern who often gets things wrong but will always give it a go.

        So long as you’re calibrated to “how might this be wrong” when looking at the results it is exceptionally useful.

    • @Evotech
      link
      English
      1
      edit-2
      6 months ago

      Bing AI provides reference in the “more precise” version