Still digesting the info from the Autonomy Day. Tesla explained how they train their cars. Well, I think unless Tesla changes how they teach their neural nets, with any future computer power, IMHO that level of L5 autonomy is impossible.
Why? Well, think about it. The way Tesla teach cars to drive is no split it into features, and train the NN to implement each feature. Some examples:
1. Traffic lights. Ok, let’s teach car to figure it out. One more NN.
2. Need to be able to interpret human signs? Ok, one more NN, feed lots of humans waving hands to train it.
3. Cars with bicycles? No need for bew NN, but people have to manually label data to mark ‘cars with bicycles’ as just cars.
4. Junk on the road? One more NN, feed lots of junk images to it.
So you get it 99%. Maybe 99.99%. And then you have to start to implement very weird, rare features, which maybe only 1 in a 1,000,000 person will ever encounter. Cost to teach the cars will become prohibitive.
So to go beyond that, someone has to figure out how to let the cars teach at a more abstract level on its own, basically feed traffic code + raw video feeds, without any labeling, triggers that feed corner cases back to Tesla for Humans to fugure out.
Now, there is no solution for that. At high level, car is constrained by a set of ‘features’, hard-coded by Tesla. This will stop the exponential improvement somewhere at 99.99%. This is not a Tesla-specific problem, of course. But this is why we are talking about 10+ years, until a more fundamental way for cars to learn by itsels how to drive is invented.
Why? Well, think about it. The way Tesla teach cars to drive is no split it into features, and train the NN to implement each feature. Some examples:
1. Traffic lights. Ok, let’s teach car to figure it out. One more NN.
2. Need to be able to interpret human signs? Ok, one more NN, feed lots of humans waving hands to train it.
3. Cars with bicycles? No need for bew NN, but people have to manually label data to mark ‘cars with bicycles’ as just cars.
4. Junk on the road? One more NN, feed lots of junk images to it.
So you get it 99%. Maybe 99.99%. And then you have to start to implement very weird, rare features, which maybe only 1 in a 1,000,000 person will ever encounter. Cost to teach the cars will become prohibitive.
So to go beyond that, someone has to figure out how to let the cars teach at a more abstract level on its own, basically feed traffic code + raw video feeds, without any labeling, triggers that feed corner cases back to Tesla for Humans to fugure out.
Now, there is no solution for that. At high level, car is constrained by a set of ‘features’, hard-coded by Tesla. This will stop the exponential improvement somewhere at 99.99%. This is not a Tesla-specific problem, of course. But this is why we are talking about 10+ years, until a more fundamental way for cars to learn by itsels how to drive is invented.