Incremental progress in basic image recognition is fairly well understood at this point...
It would be very interesting to gain some insight into Tesla's development approach for _how_ the car drives. Driving policy - not just for other road users, but in general... this is an area that Tesla haven't spoken about (all of Karpathy's talks focus on image recognition and data collection). Teaching AP how to read, react and understand the world around it (rather than just recognise it) is the really interesting area we need to see a lot of advancement in before it becomes a viable approach.
Would be super interested in any thoughts from
@verygreen on any news in that area...
Yes. The latest builds seem to do lane keeping like a neural network that has been trained on driving. Ie. It doesn’t appear to be using programmatic rules. It does appear to use rules for deciding actions on things like stop lights, stop signs, etc. So, a hybrid approach. Some neural net, some programmatic rules. I suspect this is the best way forward for Tesla.
I think human brains also use a similar hybrid approach. When you are a beginner driver, your cortex is consciously giving inputs into the car and furiously trying to keep the car centered in the lane. Over time, the autonomous parts of the brain start to take over. An experienced driver can be thinking quite complex thoughts (daydreaming) while driving because they aren’t consciously doing lane keeping anymore, that is being done unconsciously.
In a way, Tesla’s off line learning and neural net builds are similar to what the human brain does in this case.
But this approach breaks down for things that have hard and fast rules like stop sign intersections (4 way, Not 4 way, left turn, right turn, etc.). Here, you see that even human brains struggle. How many times have you been at a 4 way (they are common where I live), and four cars have approach their stops, but they don’t enter the intersection in the correct order? What you’ll often see is a person entering after they have given a single driver the right of way, instead of giving possibly three other drivers the right of way. These drivers are letting their brains drive autonomously without thinking about the rules, and human brains will often shortcut and treat a 4 way like a two way stop sign intersection, especially when another part of their brain is stressed about getting somewhere fast.
Given that human neural nets can’t properly handle 4 way intersections, I doubt Tesla will figure it out (but I could be wrong). So it’ll probably be a combo of neural net and some interesting programmatic coding.