- cross-posted to:
- news
- cross-posted to:
- news
Tesla recalls nearly all vehicles sold in US to fix system that monitors drivers using Autopilot::Tesla is recalling nearly all vehicles sold in the U.S., more than 2 million, to update software and fix a defective system when using Autopilot.
This is the important part, it’s not just Tesla not giving a shit about their customers safety, it’s dangerous to literally anyone in the same road as a Tesla
This case went to trial already and the jury sided with Tesla because the driver was holding his foot on the accelerator to override cruise control, ignored the vehicles warnings, drove through numerous flashing lights, crashed through a stop sign, and then hit the couple, later stating “I expect to be the driver and be responsible for this… I was highly aware that was still my responsibility to operate the vehicle safely." Any vehicle is capable of doing this with a reckless driver behind the wheel.
You may be confused, there’s a lot of Tesla court cases to keep track of tho, I should have been specific.
https://arstechnica.com/tech-policy/2023/12/tesla-fights-autopilot-false-advertising-claim-with-free-speech-argument/
You may be confused as I’m referring to the case from Angulo who was being quoted in the above comment.
I don’t see how a dispute between the California DMV and Tesla over the terms “Autopilot” and “Full Self Driving” make Autopilot any more dangerous than calling it something else and other companies use similar names like “ProPilot” or “PilotAssist” for their LKAS systems.
Idiot drivers do idiot things, hardly unique to Tesla drivers. The whole autopilot feature-set is pretty clearly marked as beta on the screen with a clear warning, that you have to acknowledge to enable, that it is considered a beta-feature and that you should be attentive when using it.
I agree that the FSD feature-set is advertised with capabilities it in no way possesses. But everyone seems to forget that autopilot and FSD are two separate things. Autopilot is only TACC+Lane-Assist and isn’t advertised as a fully self driving technology.
Because Tesla made it as confusing as possible on purpose and misleads consumers…
Which is why they’re getting sued about it as I type this…
I’m not sure why people have issues differentiating FSD and Autopilot honestly, they have separate names that are not similar sounding.
I 100% agree that FSD has been falsely advertised for almost a decade now though.
It’s because they think they’re experts after superficially reading through some news headlines and don’t actually do any research nor use either of these systems to gain the appropriate knowledge. They’re just playing a game of telephone with what they think they know.
The system should nevertheless be designed to handle all types of road and weather conditions. It’s a safety-related system. To not do so, regardless of the reason (probably cost savings) is negligence on the part of Tesla.
Autopilot isn’t any more dangerous than any other vehicle sold with cruise control over the past 30 years. I don’t understand why people are so desperate to give reckless drivers a pass rather than making them face consequences for their actions. Is it Honda’s fault if I hold my foot on the gas and drive a 1995 Civic through a red light and T-bone someone?
It handles the same roads and conditions as pretty much all other manufacturers do with their TACC+Lane-Assist solutions. It maintains distance to vehicles ahead, and keeps the car centered in the lane on straight road and gentle curves…nothing more.
The issue is people using this simple ADAS way outside it’s constraints, and wanting it to be more than it is, and is advertised as. Autopilot is not a self driving solution, and isn’t advertised as it either. It has the same limitations as other ADAS solutions in other cars, but apparently because Tesla calls their solution “Autopilot” people completely disregard warnings and limitations of the system and insist on using it as a self driving solution.
Sounds like Tesla should market it as adaptive cruise and lane assist, since clearly their clientel think autopilot means autonomous driving.
We’re getting to the point where these features might need to be locked behind a specific driving license designation. Drivers need to demonstrate to a proctor that they understand where the systems do and do not work.
We already have license classifications for commercial equipment and motorcycles, having one for automated features seems fairly justifiable at this point.
You would think a road with few drivers would be easier for the autopilot? But maybe the road lacked lines and marking? But wouldn’t you want the car to default to human control not just keep going? Any car I’ve had with lane keep assist turns off of it can’t find the lines. It’s a pretty simple failsafe. Rather have the driver a little annoyed than injured.
You can’t even enable it if it can’t detect lines on the road
Exactly. Not sure if there’s any regulation for this but from a controls standpoint you always want to make sure you fail to a “safe” state. I’d the system can’t find any of the input it needs the outcome will be unpredictable.
And the driver in the above case was holding his foot on the accelerator to override AP and stated “I expect to be the driver and be responsible for this… I was highly aware that was still my responsibility to operate the vehicle safely."
TIL