Last year, two Waymo robotaxis in Phoenix “made contact” with the same pickup truck that was in the midst of being towed, which prompted the Alphabet subsidiary to issue a recall on its vehicles’ software. A “recall” in this case meant rolling out a software update after investigating the issue and determining its root cause.
In a blog post, Waymo has revealed that on December 11, 2023, one of its robotaxis collided with a backwards-facing pickup truck being towed ahead of it. The company says the truck was being towed improperly and was angled across a center turn lane and a traffic lane. Apparently, the tow truck didn’t pull over after the incident, and another Waymo vehicle came into contact with the pickup truck a few minutes later. Waymo didn’t elaborate on what it meant by saying that its robotaxis “made contact” with the pickup truck, but it did say that the incidents resulted in no injuries and only minor vehicle damage. The self-driving vehicles involved in the collisions weren’t carrying any passenger.
After an investigation, Waymo found that its software had incorrectly predicted the future movements of the pickup truck due to “persistent orientation mismatch” between the towed vehicle and the one towing it. The company developed and validated a fix for its software to prevent similar incidents in the future and started deploying the update to its fleet on December 20.
It was in an orientation our devs didn’t account for and we don’t want liability.
“Towed improperly”
At least they are consistent
I love the corpospeak. why say “crashed into” when you can use “made contact” which sounds futuristic and implies that your product belongs to an alien civilization?
By “made contact”, it means that they “smashed”.
it means that they “smashed”.
So are we gonna have some baby robotaxi trucks driving around in a few month’s time?
Make contact with that like button!
Next they’re going to add passive voice to further confuse the issue. “A pickup truck was made contact with by two vehicles…”
Developers are not testing all of the edge cases properly.
Don’t assume a vehicle was under its own power, as like in this case, as it could be towed, so the towing vehicles parameter should be considered.
Check those tires! Make sure they are all on the ground.
“made contact” “towed improperly”. What a pathetic excuse. Wasn’t the entire point of self driving cars the ability to deal with unpredictable situations? The ones that happen all the time every day?
Considering the driving habits differ from town to town, the current approaches do not seem to be viable for the long term anyway.
It’s as if they are still in testing. This is many years away from being safe, but it will happen
It’s a rare edge case that slipped through because the circumstances to cause it are obscure, from the description it was a minor bump and the software was updated to try and ensure it doesn’t happen again - and it probably won’t.
Testing for things like this is difficult but looking at the numbers from these projects testing is going incredibly well and we’re likely to see moves towards legal acceptance soon
I still don’t understand how these are allowed. One is not allowed to let a Tesla drive without being 100% in control and ready to take the wheel at all times, but these cars are allowed to drive around autonomously?
If I am driving my car, and I hit a pedestrian, they have legal recourse against me. What happens when it was an AI or a company or a car?
You have legal recourse against the owner of the car, presumably the company that is profiting from the taxi service.
You see these all the time in San Francisco. I’d imagine the vast majority of the time, there are no issues. It’s just going to be big headlines whenever some accident does happen.
Nobody seems to care about the nearly 50,000 people dying every year from human-caused car accidents
Nobody seems to care about the nearly 50,000 people dying every year from human-caused car accidents
I would actually wager that’s not true, it’s just that the people we elect tend to favor the corporations and look after their interests moreso than the people who elected them, so we end up being powerless to do anything about it.
sure, but why do these accidents caused by AI drivers get on the news consistently and yet we rarely see news about human-caused accidents? it’s because news reports what is most interesting - not exactly accurate or representative of the real problems of the country
Yeah same reason why a single EV fire is national news but an ICE fire is just an unnoteworthy, everyday occurrence.
The company is at fault. I don’t think there’s laws currently in place that say a vehicle has to be manned on the street, just that it uses the correct signals and responds correctly to traffic, but I may be wrong. It may also be local laws.
The company says the truck was being towed improperly
Shit happens on the road. It’s still not a great idea to drive into it.
The company developed and validated a fix for its software to prevent similar incidents
So their plan is to fix one accident at a time…
Honestly, I think only trial and error will let us get a proper autonomous car.
And I still think autonomous cars will save many more lives than it endangered once it become reliable.
But for now this is bound to happen…
To be clear, they still are responsible of these car and the safety of others. They didn’t test properly.
They should be trying every edge case they can think about.
A large screen on the side of a truck ? What if a car is displayed on it ? Would the car sensor notice the difference?
A farmer dropped a hay bale on the road ? It got flattened by rain ? Does the car understand that this might not be safe to drive on or to brake on ?
There is hundreds of unique situations that they should be trying before an autonomous car gets even close to a public road.
But even if you try everything there will be mistakes and fatalities.
There is hundreds of unique situations that they should be trying before an autonomous car gets even close to a public road.
Do you think “better than human drivers” is sufficient for deployment on public roads, or do you think the bar should be higher?
Honestly, I’m pragmatic, if less people die in accidents involving autonomous car, then yes.
The thing is we shouldn’t be trusting the manufacturers for these stats. It has to be reported by a government agency or something.
Similarly Autonomous car software should have to be certified by an independent organization before being deployed. Same thing for updates to the software. Otherwise we would get deadly updates from time to time.
If we deploy and handle autonomous car with the same safety approach as in aviation I’m sure this transition can be done fairly safely.
So their plan is to fix one accident at a time…
Well how else would you do it?
You drive a car and can’t quite figure out what is happening in front of you.
Do you:
- A: Turn up the music and plow right through.
- B: Slow down (potentially to a full stop) and assess the situation.
- C : Slow down, close your eyes and continue driving slowly into the obstacle
- D: Sound the horn and flash the lights
From the description offered in the article the car chose C, which is wrong.
Given the millions of global road deaths annually I think B is probably the least popular answer.
Honestly slowing down too much can easily create an accident that didn’t exist in the first place.
Not every situation can be handled by slowing down.
If that’s the default behavior on high speed road this could be deadly for the car behind you.
Rules are written in blood. Once you figure out all the standard cases, you can only try and predict as many edge cases that you can think of. You can’t make something fool proof because there will always be a greater fool that will come by.
Unexpected or not, it should do its best to stop or avoid the obstacle, not drive into it.
An autonomous vehicle shouldn’t ever be able to actively drive forward into anything. It’s basic collision detection that ought to brake the car here. If something is in the position the car wants to drive to, it simply shouldn’t drive there. There’s no reason to blame the obstacle for being towed incorrectly…
In this case it thought the vehicle had a different trajectory due to how it was improperly set up.
The car probably thought it wasn’t going to hit it until it was too late and the trajectory calculation proved incorrect.
Every vehicle on the road is few moments away from crashing if we calculate that incorrectly. It doesn’t matter if it knows its there.
Same thing applies to a human driver. Most accidents happen because the driver makes a wrong assumption. The key to safe driving is not getting in situations where driving is based on assumptions.
Trajectory calculation is definitely an assumption and shouldn’t be allowed to override whatever sensor is checking for obstructions ahead of the car.
The car can’t move without trajectory calculations though.
If the car ahead of you pulls forward when the light goes green, your car can start moving forward as well keeping in mind the lead cars trajectory and speed.
If it was just don’t hit an object in its path, the car wouldn’t move forward until the lead was half way down the block.
The car knew the truck was there in this case, it wasn’t a failure to detect. Due to a programming failure it thought it was safe to move because the truck wouldn’t be there.
If you’re following a vehicle with proper distance and it slams the brakes you should be able to stop in time as you’ve calculated their trajectory and a safe speed behind. But if that same vehicle slams on the brakes and goes into reverse, well… Goodluck.
It’s all assumptions assuming the detection is accurate in the first place.
If you’re following a vehicle with proper distance and it slams the brakes you should be able to stop in time as you’ve calculated their trajectory and a safe speed behind.
You dont need to calculate their trajectory. It’s enough to know your own.
If a heavy box falls off a truck and stops dead in front of you, you need to be able to stop. That box has no trajectory, so it’s an error to include other vehicles trajectories in the safe distance calculation.
Traffic can move through an intersection closely by calculating a safe distance, which may be smaller than the legal definition, but still large enough to stop for anything suddenly appearing on the road. The only thing needed is that the distance is calculated based on your own speed and a visually confirmed position of other things. It can absolutely be done regardless of the speed or direction of other vehicles.
Anyway. A backwards facing truck is a weird thing to misinterpret. Trucks sometimes face backwards for whatever reasons.
It would be interesting to know how the self driving car would react to a ghost driver.
You dont need to calculate their trajectory. It’s enough to know your own.
This doesn’t make sense. It’s why I was saying the car won’t move at a stop light when it goes green until the car is half way down the street.
If the car is 2.5 seconds ahead of me at 60mph on the highway, it’s only 2.5 seconds ahead of me if the other car is doing 60 mph. If the car is doing 0mph then I’m going to crash into it.
It needs to know how fast and what direction the obstacle is going, and how to calculate the rate of acceleration/deceleration and extrapolate from there.
Just like Tesla! And people wonder why they are a hated company.
In a blog post, Waymo has revealed that on December 11, 2023, one of its robotaxis collided with a backwards-facing pickup truck being towed ahead of it. The company says the truck was being towed improperly and was angled across a center turn lane and a traffic lane.
See? Waymo robotaxis don’t just take you where you need to go, they also dispense swift road justice.
they also dispense swift road justice.
They should launch shurikens out the front like a James Bond vehicle.
Hmm, so it’s only designed to handle expected scenarios?
That’s not how driving works… at all. 😐
Face it, that’s actually better than many drivers can do
The description of an unexpected/(impossible) orientation for an on road obstacle works as an excuse, right up to the point where you realize that the software should, explicitly, not run into anything at all. That’s got to be, like, the first law of (robotic) vehicle piloting.
It was just lucky that it happened twice as, otherwise, Alphabet likely would have shrugged it off as some unimportant, random event.
I didn’t read it as them saying “therefore this isn’t a problem,” it was an explanation for why it happened. Think about human explanations for accidents: “they pulled out in front of me” “they stopped abruptly”. Those don’t make it ok that an accident happened either.
Billionaires get to alpha test their software on public roads and everyone is at risk.
It’s great though - that’s how you get amazing services and technological advancement.
I wish we had that. In Europe you’re just stuck paying 50 euros for a taxi in major cities (who block the roads, etc. to maintain their monopolies).
Meanwhile in the USA you guys have VR headsets, bioluminescent houseplants and self-driving cars (not to mention the $100k+ salaries!), it’s incredible.
Lol I appreciate your enthusiasm for the USA but grass is always greener.
Bruh in the US of A the grass is greener because it’s made of polypropylene and spray painted green. Just don’t smell it, or look too hard.
Most of us are in poverty, I dont know when but we’re in another gilded age and just like the last was underneath the gold is rusty iron.
Yeah it’s $40 for an Uber in Columbus or Cleveland as well. There isn’t a monopoly on taxis creating that price, thats just how much it actually costs to rent a car for cross city travel.
If you want a no regulations/free market at the helm, you want to move to India. They have all the rules you love.
After an investigation, Waymo found that its software had incorrectly predicted the future movements of the pickup truck due to “persistent orientation mismatch” between the towed vehicle and the one towing it.
Having worked at Waymo for a year troubleshooting daily builds of the software, this sounds to me like they may be trying to test riskier, “human” behaviors. Normally, the cars won’t accelerate at all if the lidar detects an object in front of it, no matter what it thinks the object is or what direction it’s moving in. So the fact that this failsafe was overridden somehow makes me think they’re trying to add more “What would a human driver do in this situation?” options to the car’s decision-making process. I’m guessing somebody added something along the lines of “assume the object will have started moving by the time you’re closer to that position” and forgot to set a backup safety mechanism for the event that the object doesn’t start moving.
I’m pretty sure the dev team also has safety checklists that they go through before pushing out any build, to make sure that every failsafe is accounted for, so that’s a pretty major fuckup to have slipped through the cracks (if my theory is even close to accurate). But luckily, a very easily-fixed fuckup. They’re lucky this situation was just “comically stupid” instead of “harrowing tragedy”.
Get your beta tests off my tax dollar funded roads pls. Feel free to beta test on a closed track.
Full releases have plenty of bugs.
Why is an update called a recall?
What typically happens when a recall is issued for other vehicles? Don’t they either remove and replace the bad part or add extra parts to fix the issue?
How is removing bad code and replacing it with good code or just adding extra code to fix the issue any different?
Do you want to physically go somewhere?
Kinda, as the word implies. If it’s a software update, call it that; the car’s not going back to the shop/manufacturer.
It sounds like location is important for some reason.
Here’s an example of why I don’t like that they’re called recalls when it’s just a system update, if you have a recall on a food item, is there some way to fix it aside from taking it back (to be replaced) or throwing it away?
When there’s a security patch released on your phone, do we call it a recall on the phone? Or is that reserved for when there a major hardware defect (like the Samsung Note fiasco)
I think the difference in the case you mentioned is that with a car they use recall because it could be dangerous to keep using it as is.
Fair, it just seems like there should maybe be a new word for this era where an OTA update is all that’s needed.
That pickup truck was asking for it I tell ya. He was looking at me sideways, he was.
It said RAM om the side!
Do we have a fuck you in particular group yet?
yay
🚘🔥