Dojo is turning out to be a game changer. Creating this thread to hold some more technical discussions outside of the main investor thread.
You can install our site as a web app on your iOS device by utilizing the Add to Home Screen feature in Safari. Please see this thread for more details on this.
Note: This feature currently requires accessing the site using the built-in Safari browser.
I had a very different take-away from the discussion, but admittedly I could have misunderstood him. Rather, Jim seemed to imply that Dojo was meant as a more general solution based on a blue-sky approach as to what was possible (i.e., that Dojo would have more than one customer). He was involved at the early stages, but he did not know what Dojo is at this point because of multiple pivots.
Also, I don't believe that he said that Lattner is involved in Tenstorrent, Jim's new outfit. Rather, that Jim put on a conference last year that both Lattner and Karpathy were involved in.
MODS: This may be of general interest to genpop , but off course, feel free to kick it to a sub-forum.
Thanks DD ! - great summary: Good level of detail combined with clear explanations.
Follow up question:
In similar projects of this magnitude what is the ballpark ratio of work used for data transfer, compression or bottleneck workarounds, as you describe it, compared to solving the actual problem (FSD)?
Clarification:
Did Keller really say 10- 1 million times faster/better or are you using your own background to evaluate the upper bound? Did he compare to the (in car) FSD version 3 chip efficiency, or to other ways of doing FSD (competitors)
Wild-eyed speculation: Perhaps Tenstorrent will built some kind of specialized compute unit which Tesla can use in DOJO? Perhaps made-to-order, Tesla-only. Perhaps not a main or bulk part but supplementary? I love Teslas vertical integration, but sometimes having a trusted partner is worth a lot.
General amazement:
I still find it hard to believe that DOJO is generic enough for other tasks. Elon said it could mine bitcoin, and your summary seems to imply a generic quality. To my limited understanding, FSD is (or was) considered so freaking hard that solving that requires specialized hardware - as evidenced by Tesla doing exactly that re. their custom car chips.
It also kind of doesn't make sense financially: Solving FSD is worth so much money, that even making a lot of bitcoins wouldn't really measure up. On the other hand, if solving FSD takes a number of years, and a huge amount of compute and custom chips, having some extra income is useful.
The only way that I can make sense of DOJO being generic is if DOJO is actually a trojan horse kind of tech for solving AGI !!
Is there a chance that this is actually what Tesla is trying to do? Or am I flying of a tangent here?
(How does that rhyme with Elons continued warnings about AI?)
Or, if DOJO does not solve AGI entirely, then solves an at least a large subset of AGI. Or perhaps doesn't quite solve, but boosts other known techniques by a significant order of magnitude - a kind of AGI runway. Which in the end may be the missing link for solving AGI ?..!
Maybe Elon concluded that FSD was close enough to AGI that he might as well solve for AGI. And then get FSD 'for free'.
If that is the case, then solving for a subset of AGI is worth a ... what is the level above f***ton?
Most of us are getting used to Tesla being 10+ startups. I used to think that solving FSD was worth a lot - to mankind, but also to us fans and investors. Solving AGI (or a significant subset) Danm! That is is 'huger than huge'.
If true, then someone PM Warren Redlich - his most crazy estimates are way too conservative!
(AGI: Artificial General Intelligence)
I have this hunch we are closer to Level 5 FSD than is generally thought.
Why is Dojo needed if Level 5 is achievable without it?Dojo is turning out to be a game changer. Creating this thread to hold some more technical discussions outside of the main investor thread.
Great foresight with this thread @Discoducky! I also think Dojo will turn out very important in the long run.
From what I've read, and come to understand, about Tesla's in-house chip making they've made FPGA prototype chips, but do we have any confirmation they've moved to the ASIC stage of manufacturing now? In other words that they've landed on a specific hardware design? Do we have any info on the capacity of Dojo in for example Teraflops/second - even though such a metric may not be super relevant with a specific purpouse supercomputer?
I hope they aren't pursuing FPGA, just not enough memory and throughput. Not optimal for matrix math at scale with FP optimimization. Highly doubtful IMO.Great foresight with this thread @Discoducky! I also think Dojo will turn out very important in the long run.
From what I've read, and come to understand, about Tesla's in-house chip making they've made FPGA prototype chips, but do we have any confirmation they've moved to the ASIC stage of manufacturing now? In other words that they've landed on a specific hardware design? Do we have any info on the capacity of Dojo in for example Teraflops/second - even though such a metric may not be super relevant with a specific purpouse supercomputer?
The Robot Brains PodcastSomeone on reddit did a nice TLDR.
- Working with Elon Musk: "He's an incredible person, I'm still trying to map out his superpowers. Incredible intuition even with lack of information. Great judgement. He's a double edged sword because he wants the future yesterday. You have to have a certain attitude to tolerate that. If you can, you will thrive at Tesla."
- Hardest problem is variability. So many possible problems in the real world.
- 75% of his time is spent curating data. 25% on algorithms. Machine Learning is Programming where the computer fills in the blanks.
- Interventions are a great trigger for training the system. The team can detect disagreements too: A stop sign flickers on and off or map data is wrong.
- "Millions of images easily" on the system. Takes 2-3 weeks to train on new data.
- The neural network in cars today do not make road edge predictions based on one image from one camera anymore, but a birds-eye view with data from all 8 cameras. Known as "Software 2.0".
- Operation Vacation: For the engineers writing code, the goal is for neural networks to improve by themselves (and with help of labelers) and theoretically the engineers could go on vacation while the system continues to improve. Mentions "it's a half joke, our northstar."
- They do-not outsource their data labeling.
- Dojo: An in-house (vertical integration) chip designed for training mass amounts of data. Active project currently at Tesla.
- Currently do a lot of manual labeling but looking to train data based on sensors. For example: Running some cars on Lidar/Radar to verify and train visual data.
- Waymo vs Tesla: Waymo does a ton of HD mapping with many sensors and many humans before a car can enter any area. Tesla is not using HD maps, but "Low definition" maps that simply give "Left turn, Right turn" directions. Tesla relies on inexpensive parts and expensive software to drive on any road you give it. "We don't know to centimeter level accuracy where a curb is. The car needs to look at images and decide where it should be." A much "higher bar, harder to design" problem but a lot less expensive.
- Implies that Waymo doesn't have enough cars to collect the data to solve FSD but doesn't outright say it. "Scale is incredibly important for dataset curation. I would rather trade sensors for data."
- Expect exponential improvements in automation everywhere, from cars to drones to warehouses, in just the next few years. The growth in the AI space in the past 4 years has not been close to linear and Andrej is excited to see what the next 4, 10, even 20 years will bring us.
I'm halfway through and it is fun to listen to. I've seen some notes already from others and the talk seems very interesting until the very end.
Tangential question for you @Discoducky given AI day was a recruiting event.Very detailed breakdown of D1: Tesla Dojo – Unique Packaging and Chip Design Allow An Order Magnitude Advantage Over Competing AI Hardware
Super cool stuff!