... dealing with people directing traffic is entirely separate from road sign handling, and has to deal with lots of cases, including people pointing in a specific direction with flashlights, etc. That's likely the sort of situation where the car would have to stop and ask a human to tell it what to do...
This sub-thread, the discussion of unusual stop signs and multiple edge-case examples of humans directing traffic etc, underlines a relevant generalized question:
Does L4/L5 driving require the ability to understand and react to every possible scenario in the way that experienced humans do?
As in:
- Every odd object appearing in the road (What is it? How bad is it to hit it? Is it worse to hit it than to run off the road or sideswipe the adjacent car or slam the brakes and get hit?)
- Every odd sign (Is it important? Is it official? Is it a joke or an ad or a prank? How bad is it to guess wrong?)
- Every interaction with humans through voice or hand signals (Are they talking to me? Are they officially directing or impromptu assisting or just angry? What are they trying to say? Is it someone in trouble, a hitchhiker, a panhandler, a protester, a carjacker?
Probably most everyone here agrees that these judgments, difficult enough for humans, are beyond any AI that can't pass the Turing test 24/7. (If it could, I'd say that self-driving would be the least of its accomplishments.) And indeed some members predict widespread L4+ is at least decades away - based on AI being unlikely to mimic, much less surpass, human judgment in these cases.
Many others here (myself included) think L4 is doable fairly soon, and yet we continually get mired in discussions of challenging edge-case examples -
which can never end if AV response must be
human-like in nature yet unquestionably
super-human in safety performance.
It's a logical trap, and the way out is to change the problem's boundary conditions. That's why I think it's clear that AVs must be given a leg up over past human-only systems, in some aspect(s) of the their operational tasks.
Forgive me if the list is a little off, but it's usually understood that Self-Driving system it's some variant of:
Perception Localization Decision Planning Control
I don't hear too much debate about feasibility of
Localization*, more about the relative necessity of detailed pre-mapping. Robotically
Planning the trajectory and
Control of its execution (once the decision has been made) are still imperfect based on watching Tesla FSD videos, but that has nothing to do with feasibility or compute and machine-design limitations, and in my view need no technical help beyond better-quality driving expertise encoded into the software.
So this leaves
Perception and
Decision as the most challenging Intelligence-based aspects, and the argument about AI's chance of success is mostly around edge cases there. These are the areas that could really use a leg up over humans, because we're very doubtful that AI is on the trajectory to solve them as well as adaptable humans, especially not to the desired mistake-free level that everyone wants.
Now, what could we do to advantage the machine in a way that could make up for its acknowledged AI deficiencies? Well, what if we could show it not just what's around it, but what's outside the view and/or hard-to-interpret perceptually, what's coming up or what to do right here at these cones?
I'm saying that a key leg up, quite realizable today, is V2X. (I'm certainly not the only one, but there's surprisingly low discussion and a lot of scoffing and that's why I'm posting again.)
Are you not sure in 0.01% of cases that thing there is a stop sign or not, or whether it applies to you at the moment? Well then 99.x% of those remaining uncertainties will be solved because it will tell you itself, and so will the other cars nearby. Not sure you can blaze through that shadowed overpass or tunnel? It can tell you and so can the cars that went through just now. How do I know if there's a little kid hidden between parked cars? Various other cars, including the parked ones, and maybe also the little safety beacon her mom clipped onto her jacket. Not by legal mandate please, but by common sense - the same as why the child is wearing the jacket without legal requirement.
And finally, I reiterate that this will not only serve to mitigate the AV edge-case challenges, but will, over a few years' time, dramatically reduce non-AV accidents as well, because cars whether in AV or manual mode will increasingly be equipped. Even if yours isn't, some and later many and later most will be, and that's huge. X2V on fixed infrastructure would happen even faster and that's also huge.
*By the way, we already give AVs a leg up with GPS and mapping. But it's not controversial simply because its use for driving started before any serious AV deployment. Despite that it's not essential and we can and did drive without it for a century, it's a big help, an existing leg up that humans can use. So AFAIK no one is on here claiming that GPS and computer-mapping technology are unnecessary, imperfect, jammable, hackable or sometimes unavailable, even though it's arguably all of those.