Okay I tried not to get involved but I feel like people are speaking on different wavelengths. I, as a mediocre machine learning scientist, mediocre TSLA investor, and mediocre meme generator have a well balanced analysis of how Tesla is positioned in the autonomy space.
Where your friend is right:
Your friend is right that camera-based depth perception takes a lot more processing power. Yes the voxel depth resolution output currently is likely worse than what Waymo / Cruise are getting from lidar. But these are nothing new and no one should be surprised by this. I've wrote about this years ago on TMC, like here
Tesla won’t have to wait for FSD to solve all weather conditions in everywhere in the world before rolling out TN and generate significant value. With current FSD computer they should be able to drive better than human under most conditions, that would enable TN in enough markets to consume...
teslamotorsclub.com
Maybe Tesla has already used up all of HW 3.0 processing & memory, this is hard to tell because of course they are going to use as much as possible before optimization (induced demand)
He is right that sensor fusion gives higher accuracy than individual sensors (more on this later).
I would agree that its quite unlikely that Tesla will be able get to to say 10x safer than humans with
current hardware.
Where you friend is wrong:
Everything else, and every conclusion he makes from those observations above. He basically discounts any of the data Tesla collects because he doesn't think the camera resolution is sufficient to be worth anything. This is inane and myopic.
The Big Picture
Tesla is focused on developing a profitable approach to autonomy
Tesla is starting with a L2 assist system that will probably soon make the entire Tesla AI team cash flow positive. Oh and they'll be able to sell excellent upgraded passive safety features (all Tesla's produced from 2019 onward will be able to detect pedestrians / cyclists / running a red light/ into a curb etc... soon). Let's assume the path to L4/L5 robotaxi autonomy takes 5 years, well in the meantime Tesla is still making money off their development with the best in class driver assist features.
Meanwhile, Waymo / Cruise were essentially forced from the beginning to rely on sensor fusion w/ LIDAR / RADAR. Sure it's more accurate, but they had to chose it because they rely on VC funding. And you aren't getting the next round of funding if you can't demo clear progress. That means focusing on the easy way to get something working (lots of expensive sensors and focusing on limited geographical areas). And still no clear path to profitability.
If Tesla need to add sensors, improve cameras, or improve processing power, they can do that later. What, excactly have they lost in the meantime?
Whose path do you think is most stable in the current moment?
Tesla is focused on developing a scalable approach to autonomy
By starting out with trying to make FSD work almost everywhere, Tesla is forcing itself to make a scalable solution. There won't need to be as much "going back to the drawing board" as there might be at Waymo / Cruise etc who barely have gotten things to work outside of one city. Of course a downside of that is going to be performance of the general algorithm will be much worse at first. The data shows most of the disengagements are currently due to mapping issues. Tesla clearly hasn't finished whatever their generalized mapping solution is, but I'm confident they have the diverse data to figure it out. It's hilarious to think people believe Tesla has flatlined progress when even that one issue clearly has a path to improve.
Meanwhile Waymo has "solved" perception, but nary a peep about global deployment? If it was scalable, they would be touting their scaling prowless, demo in 50 cities and IPO for a trillion dollar valuation. Oh, not happening?
Tesla is doing the best compromise between what a data scientist and an accountant would do.
If you told the data scientist you had unlimited funds to solve FSD, he/she would load 100,000 cars with cameras / radars, lidars and have them driving around every area of the U.S. or world, collecting the data and working to develop a robust, scalable solution. But that's not cost effective, so no one is doing that. Tesla is chosing to collect more diverse data while competitors are focused on collecting less diverse, but better resolution data.
By developing a "cheap" perception engine (cameras + AI), Tesla is allowing their AI team to move on to focus on how to solve all the other problems in full-scale autonomy. Planning, prediction, whatever else. There is no need to wait for centimeter level depth & perception precision in order to work on all the other areas of the tech stack. This is where your friend is totally wrong. And any good data scientist knows you need a wealth of diverse data in order to make a complicated algorithm generalize well. That's something that Tesla has and competitors absolutely do not.
So Tesla is going what a good data scientist would do given a constraint on funding. And guess what? Say that the camera only stack isn't good enough in a few years for robotaxi level precision? Is Tesla farked? Um no, they can simply add high-res radar or lidar in a few years and fuse the data then.
TLDR Tesla is making a cheap, scalable, and profitable autonomy software that they can upgrade with better sensors and hardware in a few years if they need to.