Please post. You didn't say you're a domain expert. Sorry my bad - seriously, please contribute. I mean it.Fine. You want a domain expert - me - to not post, I won't. Good luck defining L5 to suit yourself.
You can install our site as a web app on your iOS device by utilizing the Add to Home Screen feature in Safari. Please see this thread for more details on this.
Note: This feature may not be available in some browsers.
Please post. You didn't say you're a domain expert. Sorry my bad - seriously, please contribute. I mean it.Fine. You want a domain expert - me - to not post, I won't. Good luck defining L5 to suit yourself.
Also sorry I didn't see the last paragraph where you put your opinion upon first read. I only saw the definition of L5.For starters, SAE autonomous driving grades need to be stated and understood up front:
SAE J3016 Automated Driving Standards
Level: 5
Name: Full Automation
Narrative Definition: the full-time performance by an automated driving system of all aspects of the dynamic driving task under all roadway and environmental conditions that can be managed by a human driver.
Execution of Steering and Acceleration/ Deceleration: System
Monitoring of Driving Environment : System
Fallback Performance of Dynamic Driving Task : System
System Capability (Driving Modes) : System
Definition of driving mode: Driving mode is a type of driving scenario with characteristic dynamic driving task requirements (e.g., expressway merging, high speed cruising, low speed traffic jam, closed-campus operations, etc.).
The above are all SAE definitions, verbatim.
In short, L5 requires full 360 degree situational awareness without requiring human intervention. For example, lane changes cannot be executed without keeping track of cars, bikes, bicycles or pedestrians in the rear flanks. Front visibility is not sufficient.
In my opinion, upto L3 is quite straightforward to implement. AP2 should be able to accomplish it. The jump from L3 to L4 is very tricky, not just technologically, but in terms of regulatory oversight. L4 to L5 is even harder.
...vision ...should be good enough for software, with enough resolution and processing power. Of course, it may be easier with more input sources.
Of course, this leaves open the questions of how much code it will take and how long it will take to develop and exactly how much processing power is needed to run that code and is it reasonable for a car without using too much energy. All of this remains unanswered. If your question meant to imply considering these factors, I don't yet think anyone can say. If you question was a more simple question of whether or not it's technically possible regardless of those factors, then I would say yes.
Okay please elaborate.Hello,
Computer BS + MS here. Not my major profession but I also took Computer Vision and AI courses and had done projects in the university.
Long story short, no I don't think it's possible to achieve Level 5 using current hardware.
Elon says we will be able to sleep in the car in 2 years. If that happens, I'd eat my hat.
...The jump from L3 to L4 is very tricky, not just technologically, but in terms of regulatory oversight. L4 to L5 is even harder.
It's called a screen/sort. Yes - when looking for info in a field one doesn't understand, one uses imperfect signaling devices (such as degrees/employment field) to make a quick judgement on who to listen to. It's imperfect - invariably you throw out good sources of information, but you also likely/hopefully throw out more bad sources than good ones when time is limited. Sorry if it upsets you. I have no agenda - I just want to hear what actual domain experts think because the forum is thick with uninformed opinions (including mine).Please understand what he truly expects: a rational discussion of something he doesn't understand in a loose domain such that should anyone respond that doesn't suit his agenda allows for a full tantrum reminiscent of DJ Trump. Queue Mr. A. Jackson please...
AI expert here, but more in the natural language processing area, but I can extrapolate autopilot.
Short answer - Not possible.
Long answer -
You need 4 ingredients to make the L5 autopilot pizza,
- a) 360 degree situational awareness
- b) Fleet learning
- c) Data crunching both in real time (like your head does), and learned (like your head does)
- d) A computer and a powersource to support all this.
360 degree situational awareness
Compare it with your head, you have stereoscopic vision, but it's mounted on an axis (your neck). AP camera is not, radar does not have enough resolution.
Cameras are vastly inferior to eyes, except we can make cameras good in a single dimension.
Military drone cameras cost millions, and they get around the whole problem with brute force (very clear lenses, huge aperture, massive CCDs) ~ but then they produce so much data. A car with a 90kwh battery can't run a computer powerful enough to process all that in real time.
Fleet Learning
I don't think they are taking advantage of their 'high resolution maps' just given how the system behaves so far, and that is the 'learning' bit that your head is so good at, and Tesla just isn't. And I don't think they will be able to take advantage of the high-rez-maps either - not to the extent they'd like you to believe, simply because the car neither has enough data storage, nor enough computational power, nor a power source to support that kind of computation. Basically that "fleet learning" thing - that's way oversold than what Tesla can realistically ever deliver. However, no way to measure the success of that ;-) so they can get away with that bluff. Story of corporate America, the judge (you) is dumber than the criminals (Tesla). Anyway, so what they CAN do is to improve their algorithms based on data. I don't think they are doing that greatly yet either. i.e. they are not considering every car's data. That may actually not be necessary even for what they are trying to achieve for now.
What they cannot do you drive 15 mins before me, and swerve a pothole, and my Tesla magically learns from your experience without a programmer in freakmont writing a line of code - BS! Not happening.
Data crunching both in real time (like your head does), and learned (like your head does)
Data crunching in real time ~ oh crap that bickhead is just swerved into my lane cuz he was texting and driving.
Learned - it's friday night, better be extra careful of honda civics with coke can exhausts and underbody lights.
With AI, we could extrapolate missing bits of info, but this won't happen for 3 reasons,
-- Human beings do not possess the ability to write complex software, not to that level on that hardware on a mobile platform.
-- Neither do we have the necessary hardware to process all that in a small enough power efficient enough package to mount on a car.
-- Your head is a computer, with stereoscopic forward vision, but it has a ridiculous amount of computational power and far superior algorithms to what Tesla is writing (okay that was below the belt, snicker).
-- Your head cannot do L5 under all situations either. Can you drive in fog? In a downpour? snowstorm?
Now with AP2 hardware, can we do it? Sorry nope!
We have the necessary 'cameras' and 'sensors' but not the necessary software, or the hardware to run that software.
However, we can get pretty damn close, for 90% of the time.
Tesla is taking the logical path here, to use Linux/GCC/graphics card acceleration repurposed for AP computation. Basically they are brute forcing computation at the problem and reacting as fast as they can to give you the illusion of FSD .. which is good enough for majority of the situations.
A computer and a powersource to support all this.
So the 'learned' portion - is some bonehead in freakmont learning for the computer and writing out code. But the real time learning replacing that bonehead - well, its possible, one day. But we need vastly superior hardware than we have today. If we did, countries wouldn't be racing each other in making the best super computer possible, while not really explaining what the hell they use them for.
Summary,
With AP2 and the current hardware Tesla will be able to give you the illusion of FSD that will actually work in 80-90% situations, which is plenty good and fairly impressive. However, it's not as sci-fi as Elon is trying to sell you. And yes, it's a shame that no other auto-manufacturer can figure this out. PS: This is the internet, so I guess I pulled off being an expert pretty well, no?
To start with, I insist on the use of SAE definitions from the PDF posted earlier. Any other basis will not work for purposes of discussion, because I refuse to discuss on the basis of arbitrarily defined parameters of what L5 may be.Can you unpack this more?
I have a BS in Computer Science and currently work in the implementation(inference/forward pass) side with deep neural networks for computer vision....The trick here isn't the sensors. Tesla's AP2 sensor suite has better coverage and tracking than you do, by far. The hard part is being able to train a network to use those inputs to perform at least as well as a human can, while still being able to process those inputs fast enough.
Do you agree with @Sir Guacamolaf guacamolf and @thegruf in their characterizations that training the neural network to successfully process the input from the cameras is years off. If I've misinterpreted your positions @thegruf and @Sir Guacamolaf please correct me.
And to all three of you - and any other experts here - does the addition of lidar and/or more radar sensors reduce the computation required?
...They're clearly bringing in some of the vision networks(evidence of this in that 8.1 appears to use parallax between the two cameras since it can identify stopped cars)
I thought the improved detection of stationary/partially obstructing objects in path was down to the new active reconstruction mode of the radar signal with whitelisting that Tesla have developed?
Can you say more - about yourself and also about your thoughts on the subject?I am an expert and I say yes.