• 1 Post
  • 29 Comments
Joined 9 months ago
cake
Cake day: February 10th, 2024

help-circle
















  • In 59 crashes examined by NHTSA, the agency found that Tesla drivers had enough time, “five or more seconds,” prior to crashing into another object in which to react. In 19 of those crashes, the hazard was visible for 10 or more seconds before the collision. Reviewing crash logs and data provided by Tesla, NHTSA found that drivers failed to brake or steer to avoid the hazard in a majority of the crashes analyzed.

    NHTSA also compared Tesla’s Level 2 (L2) automation features to products available in other companies’ vehicles. Unlike other systems, Autopilot would disengage rather than allow drivers to adjust their steering. This “discourages” drivers from staying involved in the task of driving, NHTSA said.
    “A comparison of Tesla’s design choices to those of L2 peers identified Tesla as an industry outlier in its approach to L2 technology by mismatching a weak driver engagement system with Autopilot’s permissive operating capabilities,” the agency said.
    Even the brand name “Autopilot” is misleading, NHTSA said, conjuring up the idea that drivers are not in control. While other companies use some version of “assist,” “sense,” or “team,” Tesla’s products lure drivers into thinking they are more capable than they are.
    https://www.theverge.com/2024/4/26/24141361/tesla-autopilot-fsd-nhtsa-investigation-report-crash-death

    It is not a single crash. There are assisted driving system out there using pupil tracking, to make sure drivers are still paying attention.
    Tesla’s solution is something along you need to be resting at least one hand on the steering wheel. And don’t get me started on how they are diluting the concept of “full self driving”…

    But yeah, you’re right, the only reason I’m sceptical of Tesla’s semi-self-driving tech; is because I think Elon is an egomaniac little bitch, who is incapable of ever admitting he was wrong in even the smallest way.