Anyone who has been surfing the web for a while is probably used to clicking through a CAPTCHA grid of street images, identifying everyday objects to prove that they’re a human and not an automated bot. Now, though, new research claims that locally run bots using specially trained image-recognition models can match human-level performance in this style of CAPTCHA, achieving a 100 percent success rate despite being decidedly not human.

ETH Zurich PhD student Andreas Plesner and his colleagues’ new research, available as a pre-print paper, focuses on Google’s ReCAPTCHA v2, which challenges users to identify which street images in a grid contain items like bicycles, crosswalks, mountains, stairs, or traffic lights. Google began phasing that system out years ago in favor of an “invisible” reCAPTCHA v3 that analyzes user interactions rather than offering an explicit challenge.

Despite this, the older reCAPTCHA v2 is still used by millions of websites. And even sites that use the updated reCAPTCHA v3 will sometimes use reCAPTCHA v2 as a fallback when the updated system gives a user a low “human” confidence rating.

  • @[email protected]
    link
    fedilink
    English
    90
    edit-2
    2 months ago

    This is actually a good sign for self driving. Google was using this data as a training set for Waymo. If AI is accurately identifying vehicles and traffic markings, it should be able to process interactions with them easier.

    • @[email protected]
      link
      fedilink
      English
      712 months ago

      As I understand it, the point of those captchas was never really “bots can’t identify these things” (though you’re right on that it was used to train). They use cursor movement, clicks, and other behaviours while you’re solving it to detect if you are a bot or not.

      • @Grimy
        link
        English
        432 months ago

        The image choosing was always just to train their own bots

      • @Takumidesh
        link
        English
        11
        edit-2
        2 months ago

        It’s a combination.

        Most captchas goals generally aren’t 100% prevention, it’s to put a workload in front, this makes spamming the site cost money, a bankrolled attempt could just as easily outsource the captchas to real humans.

        • @[email protected]
          link
          fedilink
          English
          12 months ago

          a bankrolled attempt could just as easily outsource the captchas to real humans.

          Exactly. I’ve been using 2captcha for that for over a decade now

      • @[email protected]
        link
        fedilink
        English
        92 months ago

        Since I started getting good at yosu and that fishing mini game in farmrpg I’ve been failing more captchas. I wonder if they’re related knowing this

      • @nieceandtows
        link
        English
        52 months ago

        Is that why I’m asked to do this over and over for 14 million times when I’m on a VPN?

        • @[email protected]
          link
          fedilink
          English
          42 months ago

          It is probably part of it, yeah. But to be clear I’m not a captcha expert or anything, just a layman.

    • @grue
      link
      English
      322 months ago

      The annoying thing is that they held us hostage for our free labor, but the results are proprietary for Google’s benefit only.

      That training data ought to be forced to be made freely available to the public, since we’re the ones who actually created it.

    • @[email protected]
      link
      fedilink
      English
      32 months ago

      Afaik this is precisely what the captcha data was intended for - training AI models. Originally leveraged machine learning. LLMs are a slightly different paradigm but same purpose and results here.

      • @[email protected]
        link
        fedilink
        English
        13
        edit-2
        2 months ago

        Its never been confirmed by Google, so I may be wrong. It still tracks that the data harvesting company with a AI self driving car project would use free human labor to identify road hazards.

        • Arthur Besse
          link
          fedilink
          English
          -9
          edit-2
          2 months ago

          I was referring to the “This is actually a good sign for self driving” part of their comment.

          The captcha circumvention arms race has been going on for over two decades, and every new type of captcha has and will continue to be broken as soon as it’s widely deployed enough that someone is motivated to spend the time to.

          So, the notion that an academic paper about breaking the current generation of traffic-related captchas (something which the captcha solving industry has been doing for years with a pretty high success rate already) is “good news” for the autonomous vehicle industry (who has also been able to identify such objects well enough to continue existing and getting more regulatory approval for years now) is…

          fry not sure meme template, no text

          • @[email protected]
            link
            fedilink
            English
            16
            edit-2
            2 months ago

            Not really. I’m not even sure what you’re disagreeing with based on the above comment.

            My point is that if bog standard AI can accurately identify all of the road information from pictures, that is good news for self driving.

            What was once a nearly impossible task for computers is now mundane, and can be used to improve safety/utility for self driving, especially for FOSS projects like comma.ai