Welcome to Tesla Motors Club
Discuss Tesla's Model S, Model 3, Model X, Model Y, Cybertruck, Roadster and More.
Register
This site may earn commission on affiliate links.
Screenshot 2024-03-25 at 19.14.56.png
 
Does v12 use anything from versions prior or were all those years a waste of time software wise?
Disclaimer: I'm not an ML engineer but I've been trying to make sense of it myself. Basically, I think the answer is that the code developed up through v11 was not at all a waste or retrospectively unnecessary.

The v12 NN is not a formless blank slate that is trained from zero on the video data; rather it has a starting-point form defined by a large number of non-zero interconnection weights, and crucially a very huge number of zero-valued weights - all of this effectively defining special-purpose localized processing sub-networks that interpret the data and make decisions. You can roughly analogize it to various cortexes of the brain, the visual and auditory etc. corresponding to perception centers, and the cerebral cortex, hippocampus for memory and spatial comprehension and so on.

Without the predecessor work developed by code and incremental ML techniques throughout the prior versions, Tesla wouldn't have the foundation networks to train. You can't just start from zero with everything connected to everything else and let it train on a giant formless network, because that has too many parameters to handle even with enormous compute resources, and it wouldn't converge to anything useful on its own. (unless the given problem - unlike driving - can be represented by a small number of well-defined rules, like the often cited Alpha Zero trained from scratch to become the world's best expert Go player).

Of course this doesn't mean that if Tesla started over again, knowing what they know now, that they would take exactly the same path to get to a trainable and capable network starting point. And it doesn't mean that they won't, sooner or later, do it over again to develop an even more capable and efficient trainable network. But for now, I believe the NN "output network" inherited from v11 is the essential basis of the trainable v12 network structure.

Below are some earlier posts I made on the same topic if you're interested, but I tried to restate and summarize the ideas in this post.

 
Maybe your vehicle in service will get priority updates to 12.3.1 in the near-term given the direction from the top: "And please do the same when cars are returned from service. This is very important." Practically software deployment team will probably be prioritized to get this out to all vehicles without additional complications for the service team.
My car was in the second wave to customers for 12.3, so I'm hoping it gets 12.3.1 early as well. Of course, Tesla will do as Tesla does!

I wonder if they'll offer to give me a demo ride when I pick up my car?
 
  • Like
Reactions: Mardak
Disclaimer: I'm not an ML engineer but I've been trying to make sense of it myself. Basically, I think the answer is that the code developed up through v11 was not at all a waste or retrospectively unnecessary.

The v12 NN is not a formless blank slate that is trained from zero on the video data; rather it has a starting-point form defined by a large number of non-zero interconnection weights, and crucially a very huge number of zero-valued weights - all of this effectively defining special-purpose localized processing sub-networks that interpret the data and make decisions. You can roughly analogize it to various cortexes of the brain, the visual and auditory etc. corresponding to perception centers, and the cerebral cortex, hippocampus for memory and spatial comprehension and so on.

Without the predecessor work developed by code and incremental ML techniques throughout the prior versions, Tesla wouldn't have the foundation networks to train. You can't just start from zero with everything connected to everything else and let it train on a giant formless network, because that has too many parameters to handle even with enormous compute resources, and it wouldn't converge to anything useful on its own. (unless the given problem - unlike driving - can be represented by a small number of well-defined rules, like the often cited Alpha Zero trained from scratch to become the world's best expert Go player).

Of course this doesn't mean that if Tesla started over again, knowing what they know now, that they would take exactly the same path to get to a trainable and capable network starting point. And it doesn't mean that they won't, sooner or later, do it over again to develop an even more capable and efficient trainable network. But for now, I believe the NN "output network" inherited from v11 is the essential basis of the trainable v12 network structure.

Below are some earlier posts I made on the same topic if you're interested, but I tried to restate and summarize the ideas in this post.

They retrain from scratch often and you can start without any foundation model. And V11 was not needed for any foundation model, according to Ashok their SORA like video generation module is the foundational model. But I think the main point is that it's very hard to do it right from the start. Karpathy said it well, "labelling is an iterative process".

Imo this is what Tesla, SpaceX and Elons companies do really well. The iterative very quickly with lots of embarrassing explosions and silly behaviours along the way until one day you have Starship, FSD V12.3.1 leaving the competition in the dust with the bears still convinced that Tesla is a joke company and Elon is a total fraud who didn't do anything except steal the company from its original founders.

Imo the main useful things from V11 is a huge dataset of complicated situations, of known failures of the previous networks and a way to classify good/bad drivers and drives. Without this doing V12 from scratch would have to do a lot more iterations before it reached V12.3.1 level performance, probably a few years of iterations.
 
We had a storm overnight and out of curiosity I turned on FSD for my drive home from work this morning. The roads had clear(ish) areas where people were driving and snow/slush on either side and in the middle. 12.3 tried to center in the lane rather than center in the wheel marks meaning it simply was driving in the snow and slush rather than in the clear parts of the road. I turned it off after a few hundred feet.

* note: I only did this on a straight section of road with no other traffic around and was (even more) ready to take over. I do not recommend using FSD in any sort of inclement weather where driving, road grip or visibility is affected. FSD is not designed for this and trusting it in these situations is foolhardy at best.

Matches my experience. It's pretty clear V12.3 is not the release where it learns how to drive in winter. That said, the dirt road performance (which is excellent) gives me some hope that the current architecture is amenable to further training and improvement on snow conditions. Dirt and snow are somewhat similar: no road markings, lower grip, bumpy, narrow, etc

100% second being much much more cautious in snow testing.

Does v12 use anything from versions prior or were all those years a waste of time software wise?

The visualizations for us humans are from the V10-era. Active safety features (automatic emergency braking and such) are based, at least in part, on the V11 architecture building blocks for perception and are used across the fleet. That alone makes all those years not a waste!
 
Below are some earlier posts I made on the same topic if you're interested, but I tried to restate and summarize the ideas in this post.
Thank you, here’s something interesting I wonder if anyone else is seeing. In past versions, what I assume was map updates, I would get different behaviors suddenly after experiencing the same mistakes over and over, then one day it would be corrected without an installed update. I made a video my first day with 12.3, but because I used a new camera and some of it was out of focus, I decided not to edit and post it. Fast forward a few days, I took my normal test route again, and lo and behold, a couple of the issue spots drove flawlessly. I ran them again and the same result. Now here’s the kicker, I grabbed our model x and ran the route, and it made the same mistakes my model s made a week prior. What would be insane is if in a few days the model x stopped making the mistakes also. I ended up recording it again and if this happens I’ll def edit a video for it.
 
Another day with more examples of 12.3 handling situations that 11.9 couldn't.
  • Exiting the freeway it had to merge across 3 lanes of traffic in about 500' to make a left turn. No problem
  • Reverse route it had to wait for an opening so it could turn right then merge left to turn onto the freeway. No problem.
I also discovered that I can turn off the auto wipers (at least for about 15 minutes.)

Finally, a general observation: There's always been variability in people's experiences but there seems to be greater variability in v12.3 than in past versions. Not sure why that is but that's what I've noticed based on reports here.

1711415041012.png
 
Definitely a lot of my 10.x/11.x issues are fixed, if not all? Still experiencing wrong lane choices that were fine in 11.4.9 (regression).

Wobbly lane choices suck. Auto speed control seems to work better in more dense traffic conditions. I live in a rural area, and 42mph on 55mph posted roads ain’t gonna cut it. Nor does 17mph on a 40mph posted road work either. Drunk drivers have better speed control on these roads. If a cop was behind me while doing 17 in a 40, I’m getting pulled over with suspicion of being the drunkest guy in 3 zip codes. No one drives like this, no one should drive like this.