No idea. May as well shake a magic 8 ball. We are also past 80% install from last nights drop.It's been 24 hours since the last wave... what are the odds of another drop tonight?
You can install our site as a web app on your iOS device by utilizing the Add to Home Screen feature in Safari. Please see this thread for more details on this.
Note: This feature may not be available in some browsers.
No idea. May as well shake a magic 8 ball. We are also past 80% install from last nights drop.It's been 24 hours since the last wave... what are the odds of another drop tonight?
with all the good feedback, I’m hoping to see a large deployment anytime this evening. Fingers crossed!It's been 24 hours since the last wave... what are the odds of another drop tonight?
- Even once you have E2E, imo any realistic self-driving car deployment will demand a vector space stack because sometimes you just want explicit control. E.g. if a regulator in some country comes to you with demands around time/distance for various maneuvers, you just want to be able to implement that.The amazing thing is that V11 was/is required to get to V12...
In order to curate / test / simulate / etc. the data, you need V11 heuristics
Yes, and there are real reasons of that can be added to clarify the word "need":In order to curate / test / simulate / etc. the data, you need V11 heuristics
Do you have a sense of whether Rohan Patel's posts are more of what Tesla is focused on or perhaps he's also excited to talk more about FSD?Rohan has increased his postings recently showing Elon has a lot of faith in him
The pure E2E examples I see start with something very simple. like keeping in the lane, then figuring when to turn left or right (and how to get through the curve to do so), not having to deal with more parameters. Then slowly build from there.Yes, and there are real reasons of that can be added to clarify the word "need":
Starting from v11 isn't theoretically needed, but to start from scratch without heuristic code, you'd have to constrain the problem to a small toy-like driving problem that can converge, then constrain (protect) that network from fundamental disruption while you add next-step capabilities. As of 2023, Tesla's prior self-driving network is a pretty good base for E2E refinement and further development. They may well be looking at alternatives that go back to the beginning, to rebuild a system that's an even better base; I don't know.
- If you try to start from nothing, just a huge neural net with all the neurons connected to all the others and some arbitrary set of connection weights, the number of parameters is ridiculously large and the number of attempts required to get to a working solution approaches Infinity.
- Cool emergent behavior, starting from nothing, does occur if the machine and the defined goals are simple enough. But there's not enough computing power on Earth to program the formless self-driving machine with no priors.
- Also, it's not just that it will take too long on a gigantic computer farm. More fundamental is that the no-priors / unbounded-connections NN will be unstable in early training, i.e. it cannot converge and refine a solution because by far most of the possible (and eventually unneeded) parameters (interconnect weights) are not only costly to include, they are adversarial unless they are zeroed out.
Tesla (and AI developers in general) freely admit that key breakthroughs and insights lead to simpler yet better performing systems - but it's not just a case of "Why didn't you think of that before?" to achieve tomorrow's working solution.
So George Hotz, along with many others, can say he knew all along that E2E would be the winner. To me, that's not at all the same as saying that Tesla wasted time doing it wrong.
Yeah. For V12 to be possible Tesla needed the following:Yes, and there are real reasons of that can be added to clarify the word "need":
Starting from v11 isn't theoretically needed, but to start from scratch without heuristic code, you'd have to constrain the problem to a small toy-like driving problem that can converge, then constrain (protect) that network from fundamental disruption while you add next-step capabilities. As of 2023, Tesla's prior self-driving network is a pretty good base for E2E refinement and further development. They may well be looking at alternatives that go back to the beginning, to rebuild a system that's an even better base; I don't know.
- If you try to start from nothing, just a huge neural net with all the neurons connected to all the others and some arbitrary set of connection weights, the number of parameters is ridiculously large and the number of attempts required to get to a working solution approaches Infinity.
- Cool emergent behavior, starting from nothing, does occur if the machine and the defined goals are simple enough. But there's not enough computing power on Earth to program the formless self-driving machine with no priors.
- Also, it's not just that it will take too long on a gigantic computer farm. More fundamental is that the no-priors / unbounded-connections NN will be unstable in early training, i.e. it cannot converge and refine a solution because by far most of the possible (and eventually unneeded) parameters (interconnect weights) are not only costly to include, they are adversarial unless they are zeroed out.
Tesla (and AI developers in general) freely admit that key breakthroughs and insights lead to simpler yet better performing systems - but it's not just a case of "Why didn't you think of that before?" to achieve tomorrow's working solution.
So George Hotz, along with many others, can say he knew all along that E2E would be the winner. To me, that's not at all the same as saying that Tesla wasted time doing it wrong.
Now with end-to-end safe enough to be deployed to customers, is this aspect of labeled perception as important versus generally gathering examples of where 12.x is doing poorly and how it could behave better? I can understand the value in getting to a base level of functionality and safety for initial deployment such as finding specific examples of NHTSA stops or preventing accidental going on red lights. We're still in the transition phase where data of specific situations might leverage 11.x to find regional or other differences, but long-term maybe this will be relatively under utilized?A way to ask the fleet for more data of specific situations
Yeah, they will still need to gather "full stop at stop sign"-like scenarios for each jurisdiction. Not just examples of where the driver thinks the car is driving like a good driver, but also where NHTSA and their peers are being a pain in their actually smart summon. And plenty of other things, they might realize that they lack enough RHD-snow data and have to ask the fleet in China for snowy examples. And plenty of other specific things they are aware of where they are making a targeted effort in order to fix the problem.Now with end-to-end safe enough to be deployed to customers, is this aspect of labeled perception as important versus generally gathering examples of where 12.x is doing poorly and how it could behave better? I can understand the value in getting to a base level of functionality and safety for initial deployment such as finding specific examples of NHTSA stops or preventing accidental going on red lights. We're still in the transition phase where data of specific situations might leverage 11.x to find regional or other differences, but long-term maybe this will be relatively under utilized?
It stil moderates. Seems that accelerating resets the BASE speed it is determining. So same amount of slowing and speeding just levels off at the "new" base speed.I wonder if the speed will hold even on curves (like mountain highways) where 12.3 will try to modulate. On local streets I know it will reset on stops.
I mentioned above that I think the stop sign data-gathering issue probably involves a huge number of simulated clips, as they've explicitly stated that a very small set of drivers (and not very much intersecting with the set of deemed good drivers) perform the stop behavior that NHTSA wants.Yeah, they will still need to gather "full stop at stop sign"-like scenarios for each jurisdiction. Not just examples of where the driver thinks the car is driving like a good driver, but also where NHTSA and their peers are being a pain in their actually smart summon.
Based on occasional reports, I think the China data gathering and training effort may need to reside entirely in firewalled China-dedicated systems. Nonetheless your point is valid across many countries with different road and traffic patterns. We also know that Tesla has considerable experience in data Gathering campaigns to address specific scenarios; this started at least as far back as the human labeling teams around 2019.And plenty of other things, they might realize that they lack enough RHD-snow data and have to ask the fleet in China for snowy examples. And plenty of other specific things they are aware of where they are making a targeted effort in order to fix the problem.
I agree that FSD disengagements, while being a useful class of data, is only one aspect and may not even be the most important data set. One might say that a disengagement event is just one kind of flag to call attention to a possibly interesting clip. But many of these are arguably unnecessary, different users have very different disengagement thresholds, and there are probably a larger number of important clips that weren't accompanied by user disengagement at all.The main innovation of end2end is that now every problem is a data problem. But they still have the data problem and just getting driver disengaging will not be enough, even if all of those cases are still useful.
I think it’s state dependent, but the accuser is the government, the camera is just evidence.I'm curious about this, as I've understood the opposite. Red light and speed cameras were common many years ago, but have been removed due to constitutional challenges - specifically the right to face your accuser. Can't haul a camera into court.
Initially I think the rollouts tend to start around the weekend but once we’re this far into it, it usually continues at increasing rate until fully deployed regardless of day.Do new waves only happen on weekends or is it random
It’s promising to hear that v12.3 is excelling in smoothness and such, this is critical for wide release as a Level 2 driver assist.