You can install our site as a web app on your iOS device by utilizing the Add to Home Screen feature in Safari. Please see this thread for more details on this.
Note: This feature may not be available in some browsers.
So he is. There were supposed to be one million of them on the road by the end of 2020Elon is going all out for robotaxi. Nothing less which could slow things down.
While an automated driving system might prevent you having 99% of accidents…it would actually cause the manufacturer to have 100% responsibility that it never had beforeNot really.. if the car is truly self driving the human 'driver' is just a passenger and not resonsible for any accidents.
Therefore the responsibility falls onto the manufacturer of the computer system that is driving. If the manufacturer isn't prepared to take on that responsibility it shows they don't believe their system is safe enough for general use.
While an automated driving system might prevent you having 99% of accidents…it would actually cause the manufacturer to have 100% responsibility that it never had before
If you break your neck in a car accident because the software got it wrong, are you going to just think “damn but never mind, statistically - couple of people somewhere haven’t been in an accident and I’ve taken one for Team Tesla”?The onus of the car manufacturer having liability is dragging the whole project down.
Which begs the question, if FSD disengages 1 second before impact because it has run out of options, then technically you are the one 'in control' when the accident happens.It would only be liable when in full AV mode and the vehicle was found to be at fault.
That’s why even L3 has a 7-10 second required handover where the car is responsible and something Tesla aren’t even trying to implement which is why they’re stuck at L2.Which begs the question, if FSD disengages 1 second before impact because it has run out of options, then technically you are the one 'in control' when the accident happens.
Where do we draw the line?
[FSD] ... and even happily power slide around a roundabout for fun
either Tesla pay or your insurance will have to take on the extra obligation and of course for them to do so they’ll want to be satisfied on the quality and appropriate use
Because it shouldn’t - so issue is with manufacturer.Which begs the question, if FSD disengages 1 second before impact because it has run out of options, then technically you are the one 'in control' when the accident happens.
Where do we draw the line?
So far we have only one contender in the L3 stakes…Mercedes. But only under extremely strict conditions. Imagine you are in an accident and someone else is hurt…are Mercedes going to lie down and pay up…or will they walk that section of motorway until they find the one thing that is outside their remit. And what then ? your car insurance will drop you instantly and suddenly you are personally liable…that’s ok, you’re a Mercedes owner…you can afford it. But what about us poor Tesla owners ?If you break your neck in a car accident because the software got it wrong, are you going to just think “damn but never mind, statistically - couple of people somewhere haven’t been in an accident and I’ve taken one for Team Tesla”?
The choice is simple, either Tesla pay or your insurance will have to take on the extra obligation and of course for them to do so they’ll want to be satisfied on the quality and appropriate use.
They can walk up and down the motorway until the cows come home, if their L3 assistant was active they will have to pay up. Simples.or will they walk that section of motorway until they find the one thing that is outside their remit.
My Insurance (Churchill) already has a whole section of terms around autonomous vehicles, the TLDR is that they cover anything that is approved for use on UK roads.So far we have only one contender in the L3 stakes…Mercedes. But only under extremely strict conditions. Imagine you are in an accident and someone else is hurt…are Mercedes going to lie down and pay up…or will they walk that section of motorway until they find the one thing that is outside their remit. And what then ? your car insurance will drop you instantly and suddenly you are personally liable…that’s ok, you’re a Mercedes owner…you can afford it. But what about us poor Tesla owners ?
Funny how your assumption is Mercedes will do anything to get off the hook yet they’re the only ones who have even started on the L3 journeySo far we have only one contender in the L3 stakes…Mercedes. But only under extremely strict conditions. Imagine you are in an accident and someone else is hurt…are Mercedes going to lie down and pay up…or will they walk that section of motorway until they find the one thing that is outside their remit. And what then ? your car insurance will drop you instantly and suddenly you are personally liable…that’s ok, you’re a Mercedes owner…you can afford it. But what about us poor Tesla owners ?
On the other side, keeping at Level 2 means that Tesla get what they want : training data.People totally misunderstand why Merc have limits on L3. Its an inconvenient truth for the Tesla hardcore that the limits are not because of software capability, the cars can continue to drive along at L2 at higher speeds and without the restrictions, the limits are imposed as a safety limit while confidence with them, the regulators, the police, the insurers, all the interested parties gain confidence and understanding in a true L3 scenario.
I suspect we'll find out the software on our cars is just as badly written in the future.At least we aren’t running on Fujitsu software…
With Level 3, the car sees and understands something is not right and asks the driver to take over.Why can't they get training data while offering L3? Remember, it's not L3 or nothing, you go back down to L2 when you're not able to use L3.
Hopefully V12 will see the true "march of the 9's" commence, and if V12 is what we hope it is, then improvement could be rapid from here, potentially skipping level 3 all together, which in its present guise at Mercedes IMHO is not really of much practical help to the user.
With Level 3, the car sees and understands something is not right and asks the driver to take over.
On the other side, keeping at Level 2 means that Tesla get what they want : training data.
Every intervention is a training video that can improve the software - you don't get that with level 3.
To be fair, from a control perspective we really haven't seen much of this improvement yet, however with end to end nets now in play with V12, this is a particular reason for Tesla to hold out at level 2 for longer.
Hopefully V12 will see the true "march of the 9's" commence, and if V12 is what we hope it is, then improvement could be rapid from here, potentially skipping level 3 all together, which in its present guise at Mercedes IMHO is not really of much practical help to the user.