Welcome to Tesla Motors Club
Discuss Tesla's Model S, Model 3, Model X, Model Y, Cybertruck, Roadster and More.
Register

FSD Beta Videos (and questions for FSD Beta drivers)

This site may earn commission on affiliate links.
At a very high level-

Previous system- Each camera ran its own single-frame NN to act as visual perception for the car. Radar was a primary sensor for forward speed and distance of objects. Nothing persisted over time.

Current beta FSD system- All camera inputs are fed into a series of cascading NNs that essentially process a BEV 4D perception of the world (think of it as 360 surround video that persists over time). Radar is not used at all. Speed and distance are determined from the video inputs creating a point cloud.

There were steps in between, where for example some BEV views existed by still frame by frame, and still using some radar inputs as well. But that's the general "How it started and how's it going" of the design philosophy.

This also required rewriting of a lot of the training code too--- the upside being if it's rewritten to understand 360 view and time, you only have to manually label something in frame 1, and the new code can self-label that object for the rest of the video as long as it remains in view of any camera (and even to make predictions about it reappearing if it moves behind something briefly- they gave examples of this on AI day).

Here's Karpathy discussing the overall transition with examples in practice from mid 2020:
Do you happen to know if the Tesla approach is fundamentally different from say the method Waymo and Mobileye are using? I presume they all use neural net processing. I would guess that the Mobileye Lidar/Radar version works from a whole-scan sensor fusion and works down to individual elements, whereas Tesla's (and maybe Mobileye's) vision systems work on camera inputs and create the whole-world later. Possibly Waymo combines all sensor inputs at the beginning?

What's the simplest way of explaining the different approaches? (Without needing to qualify them as better/worse, just how are they doing it in a few sentences, as I'd just like the overview).
 
A little late but it looks like it will happen. Tweets from various people.

1631348085360.png


1631348431909.png


1631348970581.png
 
Last edited:
  • Like
Reactions: eggzrule
I wonder if Brandon is going to do a quick drive after the download? I seem to remember Tesla not allowing people to do live streams of drives, but I don’t know if that’s still an issue or not.
He confirmed that live streams of driving are not allowed but he is going to go through all of the release notes and UI while he is still live.
 
Last edited:
I may have missed it, but has anyone confirmed whether v10 is single stack for city streets and highway? This is the biggest change we were anticipating for v10, so I’m surprised I haven’t seen any confirmation one way or the other.

My bet is no single stack yet, but would love to be wrong.
 
I may have missed it, but has anyone confirmed whether v10 is single stack for city streets and highway? This is the biggest change we were anticipating for v10, so I’m surprised I haven’t seen any confirmation one way or the other.

My bet is no single stack yet, but would love to be wrong.

According to Whole Mars who has V10, it does not have the single stack:

E_AAZBrXoAEkbiw
 
Haven't seen any highway videos yet. I'm sure we'll get them within next few hours.
Ok cool. If I was a beta tester I would have ma
According to Whole Mars who has V10, it does not have the single stack:

E_AAZBrXoAEkbiw
Thanks for finding this.

Not surprising but does make me wonder whether they’ll expand the beta without it (ie the button in a couple of weeks). I suppose it’s not a gating factor to widening the beta, as NoA is already out there as-is.
 
  • Like
Reactions: diplomat33