A Tesla was in its self-driving mode when it crashed into a parked patrol vehicle responding to a fatal crash in Orange County Thursday morning, police said.

The officer was on traffic control duty blocking Orangethorpe Avenue in Fullerton for an investigation into a suspected DUI crash that left a motorcyclist dead around 9 p.m. Wednesday when his vehicle was struck.

A Fullerton Police Department spokesperson said the officer was standing outside his vehicle around midnight when he saw a Tesla driving in his direction and not slowing down.

  • @[email protected]
    link
    fedilink
    187 months ago

    A new interesting tech failing isn’t news on a site mostly made of tech nerds. Could you explain that further?

    • @IsThisAnAI
      link
      -17
      edit-2
      7 months ago

      Yes. It’s a single crash, no details, and happens every day with non assisted driving, LKA, ACC, etc. But since it’s a Tesla anti Elon zealots have to post every single rage bait article they can get their hands on.

      It’s an obsession and has nothing to do with technology.

      • @[email protected]
        link
        fedilink
        12
        edit-2
        7 months ago

        In 59 crashes examined by NHTSA, the agency found that Tesla drivers had enough time, “five or more seconds,” prior to crashing into another object in which to react. In 19 of those crashes, the hazard was visible for 10 or more seconds before the collision. Reviewing crash logs and data provided by Tesla, NHTSA found that drivers failed to brake or steer to avoid the hazard in a majority of the crashes analyzed.

        NHTSA also compared Tesla’s Level 2 (L2) automation features to products available in other companies’ vehicles. Unlike other systems, Autopilot would disengage rather than allow drivers to adjust their steering. This “discourages” drivers from staying involved in the task of driving, NHTSA said.
        “A comparison of Tesla’s design choices to those of L2 peers identified Tesla as an industry outlier in its approach to L2 technology by mismatching a weak driver engagement system with Autopilot’s permissive operating capabilities,” the agency said.
        Even the brand name “Autopilot” is misleading, NHTSA said, conjuring up the idea that drivers are not in control. While other companies use some version of “assist,” “sense,” or “team,” Tesla’s products lure drivers into thinking they are more capable than they are.
        https://www.theverge.com/2024/4/26/24141361/tesla-autopilot-fsd-nhtsa-investigation-report-crash-death

        It is not a single crash. There are assisted driving system out there using pupil tracking, to make sure drivers are still paying attention.
        Tesla’s solution is something along you need to be resting at least one hand on the steering wheel. And don’t get me started on how they are diluting the concept of “full self driving”…

        But yeah, you’re right, the only reason I’m sceptical of Tesla’s semi-self-driving tech; is because I think Elon is an egomaniac little bitch, who is incapable of ever admitting he was wrong in even the smallest way.

        • @IsThisAnAI
          link
          -4
          edit-2
          7 months ago

          It doesn’t increase the total volume of crashes per mile driven. Humans are shitty drivers, the bar is low. We’ve heard ad nauseum about the name of FSD. It’s a truth in advertising issue snd idiotic driver issue. Not safety.