Wrong. If you come onto a stretch of super icy road, or it starts raining cats and dogs, most drivers will disengage AP, and the accident could very well happen after 5 seconds.
Why would that be AP's fault then, if you disengaged? How did AP cause the crash?
You're completely missing the point. I'm not saying it's AP's fault. I'm saying that if people disengage AP every time they are in precarious situations, then that majorly taints the data, because you have way fewer accidents being recorded as due to AP. So when Elon/Tesla compares AP data to NHTSA data as a whole, it looks like AP is the most amazing safe auto-driver in the world, but in reality it may only be slightly safer than a human driver considering how little it can do to maneuver complex situations, like let's say two trucks merging into the same lane on either side of you.
I'm not trying to be an a-hole when I say this, but it honesty shocks me at how little awareness the general public seems to have about statistics. I'm truly having a hard time processing it.
Let's see if I can clarify your point:
- Driver A is running Autopilot for 1000 miles. Every time he sees a risky situation (that has nothing to do with Autopilot's behavior) he disengages Autopilot and handles the situation himself (total 10 miles). He never crashes
- Driver B is running Autopilot for 1000 miles. Every time he sees a risky situation (that has nothing to do with Autopilot's behavior) he disengages Autopilot and handles the situation himself (total 10 miles). He crashes once
- Driver C is running manually for 1000 miles. He never crashes
- Driver D is running manually for 1000 miles. He crashes once
- Driver E is running Autopilot for 1000 miles. AP never crashes
- Driver F is running Autopilot for 1000 miles. Driver does nothing in crash-imminent situation. AP crashes once
- Driver G is running Autopilot for 1000 miles. One crash-imminent situation occurs due to AP. He disengages 2 seconds before the crash. Crash occurs
- Driver H is running Autopilot for 1000 miles. One crash-imminent situation occurs due to AP. He disengages 6 seconds before the crash. Crash occurs, enough criteria are met that Tesla doesn't count it as AP caused
Result:
Driver A: 1000 miles on Autopilot. 10 miles non-Autopilot. Zero crashes AP. Zero crashes non-AP
Driver B: 1000 miles on Autopilot. 10 miles non-Autopilot. Zero crashes AP. One crash non-AP
Driver C: 0 miles on Autopilot. 1000 miles non-Autopilot. N/A crashes AP. Zero crashes non-AP
Driver D: 0 miles on Autopilot. 1000 miles non-Autopilot. N/A crashes AP. One crash non-AP
Driver E: 1000 miles on Autopilot. 0 miles non-Autopilot. Zero crashes AP. N/A crashes non-AP
Driver F: 1000 miles on Autopilot. 0 miles non-Autopilot. One crash AP. N/A crashes non-AP
Driver G: 999.98 miles on Autopilot. 0.02 miles non-Autopilot. One crash AP. Zero crashes non-AP
Driver H: 999.95 miles on Autopilot. 0.05 miles non-Autopilot. Zero crashes AP. One crash non-AP
Analysis:
Driver A and B are skewing the data. They are letting AP run during safe times but taking over during risky times. They are not letting AP take responsibility for handling the risk
Driver C, D, E, and F are either taking full responsibility or letting AP take full responsibility. Any crashes are assigned appropriately
Driver G has an AP crash that is assigned appropriately to AP
Driver H is Tesla skewing the data. It has an AP crash that is not assigned to AP. Tesla assumes 6 seconds is enough time that the driver should have avoided the crash
If a lot of drivers act like A and B then AP gets the benefit of all the good miles and none of the crashes. Drivers A and B have skewed the data themselves and made AP look better.
Drivers C, D, E, F, and G are correctly recorded statistics.
If there are many situations like with Driver H then Tesla has skewed the data and AP gets no penalty for crashes that it caused.
ProtonSF is that what you meant?