Hmmmm. Maybe I'm wrong on this, but I thought the "no teaching" thing was limited to rules, ie, the car will stop at red lights because it has seen thousands of clips of other cars doing so. But in order to do this, the car will still need to be able to identify objects; it will need to be able to identify a human in the road so it can then refer to the clips that show other cars stopping for this object type (that may be a bit of an extreme example, as I'm sure there's still going to be some hard code for life or death situations).One person's rewrite is another's increment?
I am curious how visualization works with E2E. If the car is taking in photons and outputting vehicle control, with no teaching for identifying specific things like stop signs, other vehicles, etc, how will the system create the visualization? Perhaps some of the legacy perception NNs are kept to run alongside the E2E stuff?
So object recognition will still be at the heart of the system, but how it responds to those objects is what's changing.