You can install our site as a web app on your iOS device by utilizing the Add to Home Screen feature in Safari. Please see this thread for more details on this.
Note: This feature may not be available in some browsers.
I think you’re exactly right. Foot on accelerator overrides TACC braking. I had an issue the other day that I was ready to blame on the car but it was my operator error. I was in AS and the car came to a stop behind traffic at a light. After the stop the AP went wonky with an error alert so I flipped the right control stop up to cancel AP. Light turns green...traffic starts moving forward but my car went reverse....LMAO. Foot on brake then into drive and all was good...it’s a good thing I didn’t get on the accelerator pulling away because cars were behind me. Car only moved in reverse a foot or so before I simply stopped...then to FWD ...and all ops normal. Careful disengaging automation using the control stalk slightly up because you can get Reverse accidentally. Cheers.It looks like autosteer was on until the last minute because it's pretty solid as far as steering goes. The car isn't accelerating or decelerating when you compare it to your surroundings. Unless you were steering. Another possibility, without autopilot cancelling out was if you had your foot slightly on the accelerator. This would have overridden any braking from TACC.
Also, 2019.5.15 and 2019.8.3 are pretty buggy, so who knows.
I always try for the half press. I've done the full press accidentally once myself.I think you’re exactly right. Foot on accelerator overrides TACC braking. I had an issue the other day that I was ready to blame on the car but it was my operator error. I was in AS and the car came to a stop behind traffic at a light. After the stop the AP went wonky with an error alert so I flipped the right control stop up to cancel AP. Light turns green...traffic starts moving forward but my car went reverse....LMAO. Foot on brake then into drive and all was good...it’s a good thing I didn’t get on the accelerator pulling away because cars were behind me. Car only moved in reverse a foot or so before I simply stopped...then to FWD ...and all ops normal. Careful disengaging automation using the control stalk slightly up because you can get Reverse accidentally. Cheers.
Sounds high, but since it's a new car, you're obviously getting a new deck lid and since it'll require a repaint, it means blending paint on all the surrounding panels. I assume the quote provides a cost breakdown.@Fernand Thanks for the catch. Insurance is covering it, i'm just paying from my deduction. Don't want to give info on this body shop since my car is still in their possession. But its the East Bay Contra Coasta. I'm just glad my spoiler wasn't attached when this happened.
As far as I know, Tesla has never paid out in such situations, nor ever "admitted fault" - out of the lawyerly fear that it would set an un-affordable precedent that could wipe out the company. And The Answer to what we are all asking here, namely "what exactly happened" in this and in other accidents, will, for years to come, likely be hidden behind the "proprietary" and "beta" pleas. I'd be pleasantly surprised if Tesla told us, unless it was to call it pilot error. Would it help to know for sure?
Now, I use my Model 3 in NavAP and AP a lot, on freeways, highways, and city boulevards. Sure, I take a reasonably defensive approach. But honestly, how I feel doing it depends mostly on a rising or falling wave of trust or faith. Some days I happily entrust my life to it. On other days, I'm nervous, as I consider the full consequences of error. And to remind us, we have the two recent Boeing crashes.
As a software and hardware developer, I know how buggy all code is, and how it's always pushed out before it's fully tested. Because it's impossible. There isn't enough time to test every possible combination of events. So I know it will fail spectacularly on occasion. But I'm also very impressed by how well it works most of the time. So I'm torn. And I know that MY own driving is pretty fair but will fail big time now and then. I'm not sure if "rather better than a human" is comforting.
This thread is extremely interesting in revealing, between the lines, how emotionally difficult this whole robot car business is. Some will unequivocally blame the system. Others the operator. But we're not really sure. Anybody here have similar thoughts?
This is a fantastic post and exactly what I think as well with a similar background.
For me the biggest issue at the moment is the phantom braking which really feels solvable in the very short term.
I also think if they used rear camera to verify it's safe to brake (lowering the "score" to do it or not, or how much is safe) I'd feel much better for when it does glitch. We may need HW3 before this phantom braking is gone. They must know the issue is there and there seems no progress on it. I suspect they know they need to "process more neural net cases" (what ever they call them) to rule out mistakes. More cases processed per second the more precise "scoring" will be. I think an order of magnitude more processing may solve a lot of issues.
The erratic lane changes I don't mind because I'm ready for it, when I confirm change lane. Not good of course but the braking out of the blue can happy any time. You could go 2 hours with out a glitch and wham.
I'm really shocked they are considering no confirmation NoA without HW3. It seems unnecessary risk. It better be optional, and defaulted off. Or those one in 10 million spectacular failures will become 1 in a million (which is a lot). It may damage their reputation and harm the eventual FSD regulation process to come.
Does anyone know of a phantom brake that has caused an accident? I'd be shocked if it hasn't.
I don't think this is true. Human stereopsis is useless beyond 10 meters. Your brain determines distance by looking at many other clues.It is very difficult (and therefore inaccurate) to predict distance from a single visual image source, like that from a single rear camera. If you've ever known anyone who has vision problems or lacks one of their two eyes, you might be aware that they've got spacial and depth perception challenges.
This - human eyes are spaced so that stereo works for objects within reach.I don't think this is true. Human stereopsis is useless beyond 10 meters. Your brain determines distance by looking at many other clues.
https://human-factors.arc.nasa.gov/publications/AIAA.2011.DepthPerceptionCueCntrl.pdf
My uncle was killed by one of the sleepy drivers, as was the other driver. Please pull over and use dog mode for a quick nap.I'm glad that you and many others feel the same way. I've had this problem before (especially after long hours of work and little sleep the night before) and it's one of the main reason's that led to me being a Tesla owner because I don't want to endanger anyone else on the road. I recently read something from the CDC that states that 1 in 25 people on the road doze off behind the wheel. Hopefully, Tesla and other manufacturers can make this technology more reliable going forward.
If the driver didn't initiate a lane change and the car didn't signal a lane change, then AP didn't initiate a lane change. Now, it might have followed the wrong lane markings, inexplicably drifted out of lane or what seems more likely in the case you've described tried to center itself in an apparently wider lane caused by an unmarked merge. I'm especially paranoid of using AP past unmarked merges because of that behavior, it should keep tracking the left (or right) line if the other line suddenly terminates and the lane widens.I don't personally know of a phantom braking incident causing an accident, but I do have knowledge of an AP initiated lane-change that was made by a Tesla Model X that caused an accident. The vehicle should have continued to travel in it its current lane, but for reasons yet unexplained, the Tesla moved over one lane and struck a slower moving vehicle in the left lane while merging. The driver did not signal a lane change, and the vehicle did not indicate or suggest a lane change.
It could be that Tesla has determined that lane change confirmation does not require the processing power of HW3 and that most of the requirements are solved with their deep-learning models that can be computed with their older generation of hardware.
I feel its a bit early to introduce - not so much because of the danger involved, but more because it's never really done a great job of suggesting lane changes. Tesla probably feels that it's worth making some poor decisions now concerning lane changes and gather the data so that they can improve on it as their data-set size increases substantially.
It is very difficult (and therefore inaccurate) to predict distance from a single visual image source, like that from a single rear camera. If you've ever known anyone who has vision problems or lacks one of their two eyes, you might be aware that they've got spacial and depth perception challenges. Taking into account what is occurring behind you is of course very important, but I think Tesla is going to be challenged to utilize this information accurately without the presence of additional radar sensors. There is so much to take into account in a situation like this, including the reaction and the stopping power of the car behind you.
I regularly accelerate through AP directed speed reductions, less to avoid an accident, and more so to prevent other drivers from becoming upset.
I don't personally know of a phantom braking incident causing an accident, but I do have knowledge of an AP initiated lane-change that was made by a Tesla Model X that caused an accident. The vehicle should have continued to travel in it its current lane, but for reasons yet unexplained, the Tesla moved over one lane and struck a slower moving vehicle in the left lane while merging. The driver did not signal a lane change, and the vehicle did not indicate or suggest a lane change.
t could be that Tesla has determined that lane change confirmation does not require the processing power of HW3 and that most of the requirements are solved with their deep-learning models that can be computed with their older generation of hardware.
I feel its a bit early to introduce - not so much because of the danger involved, but more because it's never really done a great job of suggesting lane changes. Tesla probably feels that it's worth making some poor decisions now concerning lane changes and gather the data so that they can improve on it as their data-set size increases substantially.
Good point about the rear not having "Stereo" vision like the front does. Perhaps it really should. Or Radar.
I wonder if the two sides with rear could form "most" of a "depth view". Maybe even the inside rear facing as well
Sure, you can drive through the deceleration, if you catch it early enough. I'm usually in WTF mode for a second or two, trying to assess if it's right. I personally have less issue with lane change issues. But that is certainly #2 on my list.
I have not had the car change lanes on it's own, which is what I think you said. I understand that it could (under the collision avoidance umbrella). But not for NoA. As far as I understood.
If the driver didn't initiate a lane change and the car didn't signal a lane change, then AP didn't initiate a lane change. Now, it might have followed the wrong lane markings, inexplicably drifted out of lane or what seems more likely in the case you've described tried to center itself in an apparently wider lane caused by an unmarked merge. I'm especially paranoid of using AP past unmarked merges because of that behavior, it should keep tracking the left (or right) line if the other line suddenly terminates and the lane widens.
How sure is your friend that he was actually using autosteer? I know that if I was next to someone driving erratically like the driver in the left lane I would have disabled autopilot. Maybe he disabled autosteer by putting pressure on the wheel and then let go of the steering wheel after he thought he was clear of the car in the left lane?Have a look at the video.
The lane markings appear to be clear. It does not seem to represent the case you have described (and for which I encounter daily on my drive home of a widening lane but it is possible that it got confused).
I'm very interested in your opinion, and the others in the forum.
Here is the narrative of what we think happened, after speaking in more detail with the driver this afternoon, and reviewing the footage in slow-motion.
- Model X was driving in the center lane and passing an adjacent car positioned in the left lane.
- The Adjacent vehicle was starting to drift slightly into the center lane.
- As the Model X was passing this vehicle, it also drifted slightly over the lane marker to the left as if it had an intention to occupy the lane.
- The Model X did not appear to be fully aware of the adjacent vehicle, or its distance.
- At this point, the two vehicles collided - the rear left wheel of the Tesla came into contact with the right wheel of the adjacent vehicle.
- Driver corrected for a moment after the impact by briefly steering right, and then manually took the Tesla into the left lane, and stopped.
Tesla gathered all of the data, escalated it internally, made some calls back to the driver, and promised to be in touch within 7-21 days later. There was no additional meaningful follow up from Tesla other than to say that they're AP team reviewed the footage and are currently prioritizing other AP related investigations and would eventually respond with more detail. There have been weekly communication since then (initiated by the driver) but no answers. I think we are hoping to better understand AP's intention just prior to, and during the impact.
The driver took immediate responsibility for the event and settled the damage. He did not end up involving insurance and was able to repair the damage to his model X for relatively little cost. Fortunately, no one was injured in the event.
Here are the images post impact:
View attachment 393749 View attachment 393750
How sure is your friend that he was actually using autosteer? I know that if I was next to someone driving erratically like the driver in the left lane I would have disabled autopilot. Maybe he disabled autosteer by putting pressure on the wheel and then let go of the steering wheel after he thought he was clear of the car in the left lane?
Dunno. Maybe the car driving over the lane line repeatedly made it hard for AP to see the lane line. I bet neural nets are hard to debug! That's why it's beta software I guess.Affirmative on the question of whether or not autosteer was still enabled. At this point in his AP driving experience, he was still extremely trusting and let it do its thing. I'm a bit more cautious and have a tendency to disable AP on any questionable driving condition. I'm fairly certain I would have disengaged in this scenario.
Stereo imaging for the rear is better than the front, at the moment - though with the challenge of different focal lengths. The fender cameras see directly to the rear, so any object will be in the sight from at least one fender camera and the rear fisheyebl unless it's very close.
That gives a minimum of a three foot separation between cameras, compared to the couple inches forward.