I don't follow the autopilot issue closely but others here are more competent and up-to-date. Please correct any misunderstandings.
By some measures, including comparison with autopilot 1.0, the current autopilot falls behind or is more betaish than we would like. I think there are at least two reasons for this. 1) Tesla is doing this the hard way by letting a real driver's experience be sent to the AI main computer for it to learn from zero prior information. No software human magician used to develop the algos. I have also heard those in the know say only two cameras are being "audited" at present. No mention of radar inputs, gps, or microsound (? I've lost the right word) sensing of soft targets. Implementation should be slower by this method but eventually the finished product should be spectacular once all sensor data has been uploaded and analyzed. 2) I've read somewhere the current state of either the hardware, software, or both is such the damn machines take a very long time digesting what is inputed, not so much because the computing is so difficult, but the process of ingesting it is slow, probably just because of the volume of data. (Apologies, use of alimentary words needed because I'm not a computer nerd by any means.) If I remember correctly, there is an article by Harry Jerison who says our eyes and immediate neural linkages preprogram images into two dimensional frames before they are sent to the brain which is why others note babies seem to smile at faces way earlier than expected and, of course, this capacity aids in locating a source of milk.
Sorry to bother you smart guys, but can you clean up these understandings/misunderstandings?
Edit: Maybe I've used the wrong term. By AI I mean machine learning, if that makes a difference.