powertoold
Active Member
Pardon? The 300+K lines of control code have no perception capabilities whatsoever. It just reacts to inputs provided from the planner.
Here's my logic on this, someone please break it down:
1) the 300k lines of code pertain to both planning and control, essentially all heuristics
2) the heuristics are dependent on perceptual objects labeled in a human-understandable way. For example, the vision perception NN will predict 100s of objects in human-heuristics ("stop sign", "traffic light", "pedestrian", etc.) and convert them to a BEV vector space.
3) V12 gets rid of all these human-understandable heuristics, so there's no way to code the planner with human-understandable objects. Per Ashok, V12 has internal representations of human heuristics, and there's no good way to ask V12 to respond to a "stop sign" or "traffic light" for example. All of V12's normal driving functions are only based on video training of good drivers.
4) so it begs the question, if perceptual objects are no longer represented in human heuristics for all the driving, then what do we make of all the labeled objects from V11?
Last edited: