Waymo issued a recall after two robotaxis crashed into the same pickup truck

Last year, two Waymo robotaxis in Phoenix “made contact” with the same pickup truck that was in the midst of being towed, which prompted the Alphabet subsidiary to issue a recall on its vehicles’ software. A “recall” in this case meant rolling out a software update after investigating the issue and determining its root cause.

In a blog post, Waymo has revealed that on December 11, 2023, one of its robotaxis collided with a backwards-facing pickup truck being towed ahead of it. The company says the truck was being towed improperly and was angled across a center turn lane and a traffic lane. Apparently, the tow truck didn’t pull over after the incident, and another Waymo vehicle came into contact with the pickup truck a few minutes later. Waymo didn’t elaborate on what it meant by saying that its robotaxis “made contact” with the pickup truck, but it did say that the incidents resulted in no injuries and only minor vehicle damage. The self-driving vehicles involved in the collisions weren’t carrying any passenger.

After an investigation, Waymo found that its software had incorrectly predicted the future movements of the pickup truck due to “persistent orientation mismatch” between the towed vehicle and the one towing it. The company developed and validated a fix for its software to prevent similar incidents in the future and started deploying the update to its fleet on December 20.

JCreazy,

I’m getting tired of implementing technology before it’s finished and all the bugs are worked out. Driverless cars are still not ready for prime time yet. The same thing is happening currently with AI or companies are utilizing it without having any idea what it can do.

nooeh,

How will they encounter these edge cases without real world testing?

JCreazy,

Fair point

drivepiler,

I agree, but testing with a supervisory driver should be required in case of emergency situations. Both safer and creates job opportunities.

corsicanguppy,

tired of implementing technology before it’s finished

That’s is every single programme you’ve ever used.

Software will be built, sold, used, maintained and finally obsoleted and it will still not be ‘complete’. It will have bugs, sometimes lots, sometimes huge, and those will not be fixed. Our biggest accomplishment as a society may be the case where we patched software on Mars or in the voyager probe still speeding away from earth.

Self-driving cars, though, don’t need to have perfectly ‘complete’ software, though; they just need to work better than humans. That’s already been accomplished, long ago.

And with each fix applied to every one of them, it’s a situation they all shouldn’t ever repeat. Can we say the same about humans? I can’t even get my beautiful, stubborn wife to slow down, leave more space, and quit turning the steering wheel in that rope-climbing way like a farmer on a tractor does (because the airbag will take her hand off).

dsemy,

That’s is every single programme you’ve ever used.

No software is perfect, but anybody who uses a computer knows that some software is much less complete. This currently seems to be the case when it comes autonomous driving tech.

And with each fix applied to every one of them, it’s a situation they all shouldn’t ever repeat.

First, there are many companies developing autonomous driving tech, and if there’s one thing tech companies like to do is re-invent the wheel (ffs Tesla did this literally). Second, have you ever used modern software? A bug fix guarantees nothing. Third, you completely ignore the opposite possibility - what if they push a serious bug in an update, which drives you off a cliff and kills you? It doesn’t matter if they push a fix 2 hours later (and let’s be honest, many of these cars will likely stop getting updates pretty fast anyway once this tech gets really popular, just look at the state of software updates in other industries).

daed,

I understand your issue with these cars - they’re dangerous, and could kill people with incomplete or buggy software. I believe the person you are responding to was pointing out that even with the bugs, these are already safer than human drivers. This is already better when looking at data rather than headlines and going off of how things seem.

Personally, I would prefer to be in control of the vehicle at all times. I don’t like the idea of driverless tech either.

redfox,

Well, has anyone done good statistics to show all the self driving cars are more dangerous than regular distracted humans as a whole?

We can always point to numerous self driving car errors and accidents, but I am under the impression that compared to the number of accidents involving people on a daily basis, self driving cars might be safer even now?

I’m thinking of how many crashes took place in the time it took me to type this out. I’m also curious about the fatality rate between self or assisted driving vs not.

I think we tend to be super critical of new things, especially tech things, which is understandable and appropriate, but it would be nice to see some holistic context. I wish government regulators would publish that data for us, to help us form informed opinions instead of having to rely on manufacturers (conflict of interest) or journalists who need a good story to tell, and some clicks.

dsemy,

Currently there are many edge cases which haven’t even been considered yet, so maybe statistically it is safer, but it doesn’t change anything if your car makes a dumb mistake you wouldn’t have and gets you into an accident (or someone else’s car does and they don’t stop it cause they weren’t watching the road).

long_chicken_boat,

I’m against driverless cars, but I don’t think this type of errors can be detected in a lab environment. It’s just impossible to test with every single car model or real world situations that it will find in actual usage.

An optimal solution would be to have a backup driver with every car that keeps an eye on the road in case of software failure. But, of course, this isn’t profitable, so they’d rather put lives at risk.

nivenkos,

That’s how you get technological advancement.

Bureaucracy just leads to monopolies and little to any progress.

LesserAbe,

You’re right there should be a minimum safety threshold before tech is deployed. Waymo has had pretty extensive testing (unlike say, Tesla). As I understand it their safety record is pretty good.

How many accidents have you had in your life? I’ve been responsible for a couple rear ends and I collided with a guard rail (no one ever injured). Ideally we want incidents per mile driven to be lower for these driverless cars than when people drive. Waymos have driven a lot of miles (and millions more in a virtual environment) and supposedly their number is better than human driving, but the question is if they’ve driven enough and in enough varied situations to really be an accurate stat.

Doof,

A slightly tapped a car a first day driving, that’s it. No damage. Not exact a good question.

Look at how data is collected with self driving vehicles and tell me it’s truly safer.

LesserAbe,

My point asking about personal car incidents is that each of those, like your car tap, show we can make mistakes, and they didn’t merit a news story. There is a level of error we accept right now, and it comes from humans instead of computers.

It’s appropriate that there are stories about waymo, because it’s new and needs to be scrutinized and proven. Still it would benefit us to read these stories with a critical mind, not to reflexively think “one accident, that means they’re totally unsafe!” At the same time, not accepting at face value information from companies who have a vested interest in portraying the technology as safe.

Doof,

I obviously do since I said look at how the data is collected, what is counted and what is not. Take your own advice and look into that. It’s not this one accident that makes me think it’s unsafe, and certainly not ready to be out there driving.

LesserAbe,

Here’s an article saying that based on data so far, waymo is safer than human drivers. If you have other information on the subject I’d be interested to read it.

Doof,

www.youtube.com/watch?v=pmGOjHi-7MM&t=129s This is a good and entertaining video on it but if you prefer to read here is the sources docs.google.com/document/d/…/edit

Also your own article “But it’s going to be another couple of years—if not longer—before we can be confident about whether Waymo vehicles are helping to reduce the risk of fatal crashes.”

overzeetop,
@overzeetop@lemmy.world avatar

The description of an unexpected/(impossible) orientation for an on road obstacle works as an excuse, right up to the point where you realize that the software should, explicitly, not run into anything at all. That’s got to be, like, the first law of (robotic) vehicle piloting.

It was just lucky that it happened twice as, otherwise, Alphabet likely would have shrugged it off as some unimportant, random event.

dan1101,

Billionaires get to alpha test their software on public roads and everyone is at risk.

nivenkos,

It’s great though - that’s how you get amazing services and technological advancement.

I wish we had that. In Europe you’re just stuck paying 50 euros for a taxi in major cities (who block the roads, etc. to maintain their monopolies).

Meanwhile in the USA you guys have VR headsets, bioluminescent houseplants and self-driving cars (not to mention the $100k+ salaries!), it’s incredible.

vaultdweller013,

Most of us are in poverty, I dont know when but we’re in another gilded age and just like the last was underneath the gold is rusty iron.

HappyRedditRefugee,

We have something like that here too: MOIA in Hamburg.

BakerBagel,

Yeah it’s $40 for an Uber in Columbus or Cleveland as well. There isn’t a monopoly on taxis creating that price, thats just how much it actually costs to rent a car for cross city travel.

If you want a no regulations/free market at the helm, you want to move to India. They have all the rules you love.

LesserAbe,

Lol I appreciate your enthusiasm for the USA but grass is always greener.

Patches,

Bruh in the US of A the grass is greener because it’s made of polypropylene and spray painted green. Just don’t smell it, or look too hard.

JungleJim,

Bioluminescent house plants are cool but as an American I can tell you right now that my luxury bones hurt.

nivenkos,

I can tell you right now that my luxury bones hurt.

That’s the same in Europe though, dentistry isn’t covered on public insurance in the UK, Spain, Sweden, etc.

But we have even less net salary to cover it when there are problems.

JungleJim,

True, but your savings on non-luxury bones helps with the fees associated with luxury ones, I’m sure. I can’t do anything for my bones with a $30 glowing petunia.

redfox, (edited )

I appreciate/understand your envy. I’m not sure why everyone disagrees so much unless they have also lived under similar constraints.

Unless sarcasm.

Also agree with it might be perception or grass is greener like other comment 😉

LesserAbe,

I didn’t read it as them saying “therefore this isn’t a problem,” it was an explanation for why it happened. Think about human explanations for accidents: “they pulled out in front of me” “they stopped abruptly”. Those don’t make it ok that an accident happened either.

___,

It would have been a different article if two waymos decided to take a wrong turn off a cliff.

bizzle,
@bizzle@lemmy.world avatar

It should of course not run into anything, but it does need to be able to identify obstacles at the very least for crash priority when crazy shit inevitably happens. For instance, maybe it hits a nice squishy Pomeranian that won’t cause any damage to itself instead of swerving to avoid it and possibly totalling itself by hitting a fire hydrant.

Or maybe it hits the fire hydrant instead of a toddler.

At any rate, being able to identify an obstacle and react to unexpected orientations of those obstacles is something I think a human driver does pretty well most of the time. Autonomous cars are irresponsible and frankly I can’t believe they’re legal to operate.

Patches,

I can’t believe they’re legal to operate.

That’s the neat part. They aren’t always legal. It doesn’t stop them.

wsj.com/…/california-dmv-calls-ubers-autonomous-a…

Blackmist,

This is our future isn’t it? This is it. Spending our days wondering if we’re going to be mown down by a clumsy Johnnycab because it was fractionally cheaper than paying somebody to drive.

Argonne,

I’d take my chances with that rather than all the crazies out in the road now

cm0002,

Fr, I’d still trust a self driving vehicle over a human driver any day of the week.

Humans are terrible drivers, this could have easily been just another person driving distracted or something and then we wouldn’t even know about it because it wouldn’t be news worthy.

Mango,

It was in an orientation our devs didn’t account for and we don’t want liability.

“Towed improperly”

deafboy,
@deafboy@lemmy.world avatar

“made contact” “towed improperly”. What a pathetic excuse. Wasn’t the entire point of self driving cars the ability to deal with unpredictable situations? The ones that happen all the time every day?

Considering the driving habits differ from town to town, the current approaches do not seem to be viable for the long term anyway.

Meowoem,

It’s a rare edge case that slipped through because the circumstances to cause it are obscure, from the description it was a minor bump and the software was updated to try and ensure it doesn’t happen again - and it probably won’t.

Testing for things like this is difficult but looking at the numbers from these projects testing is going incredibly well and we’re likely to see moves towards legal acceptance soon

Argonne,

It’s as if they are still in testing. This is many years away from being safe, but it will happen

indomara,

I still don’t understand how these are allowed. One is not allowed to let a Tesla drive without being 100% in control and ready to take the wheel at all times, but these cars are allowed to drive around autonomously?

If I am driving my car, and I hit a pedestrian, they have legal recourse against me. What happens when it was an AI or a company or a car?

kava,

You have legal recourse against the owner of the car, presumably the company that is profiting from the taxi service.

You see these all the time in San Francisco. I’d imagine the vast majority of the time, there are no issues. It’s just going to be big headlines whenever some accident does happen.

Nobody seems to care about the nearly 50,000 people dying every year from human-caused car accidents

CosmicCleric,
@CosmicCleric@lemmy.world avatar

Nobody seems to care about the nearly 50,000 people dying every year from human-caused car accidents

I would actually wager that’s not true, it’s just that the people we elect tend to favor the corporations and look after their interests moreso than the people who elected them, so we end up being powerless to do anything about it.

kava,

sure, but why do these accidents caused by AI drivers get on the news consistently and yet we rarely see news about human-caused accidents? it’s because news reports what is most interesting - not exactly accurate or representative of the real problems of the country

ShepherdPie,

Yeah same reason why a single EV fire is national news but an ICE fire is just an unnoteworthy, everyday occurrence.

I_Fart_Glitter,

And they wonder why we set them on fire…

rsuri,

In a blog post, Waymo has revealed that on December 11, 2023, one of its robotaxis collided with a backwards-facing pickup truck being towed ahead of it. The company says the truck was being towed improperly and was angled across a center turn lane and a traffic lane.

See? Waymo robotaxis don’t just take you where you need to go, they also dispense swift road justice.

Sculptor9157,

Maybe it was a cybertruck and the super stealth design made it’s signature very small.

EdibleFriend,
@EdibleFriend@lemmy.world avatar

Do we have a fuck you in particular group yet?

baseless_discourse,
EdibleFriend,
@EdibleFriend@lemmy.world avatar

yay

dog_,

🚘🔥

Aatube,
Aatube avatar

Why is an update called a recall?

Chozo,
Chozo avatar

The fleet of cars is summoned back to the HQ to have the update installed, so it causes a temporary service shutdown until cars are able to start leaving the garage with the new software. They can't do major updates over the air due to the file size; pushing out a mutli-gigabyte update to a few hundred cars at once isn't great on the cellular network.

Jakeroxs,

Actually there have been several Tesla “recalls” that were just simply OTA updates.

twack,

Because Tesla was fixing significant safety issues without reporting it to the NHTSA in a way that they could track the problems and source of the issue. The two of them got into a pissing match, and the result is that now all OTA’s are recalls. After this, the media realized that “recall” generates more views than “OTA”, and here we are.

Dlayknee,

I think it’s slightly more nuanced - not all OTAs are recalls, and not all recalls are OTAs (for Tesla). Depending on the issue (for Teslas), the solution may be pushed via an OTA in which case they “issue a recall” with a software update. They’re actually going through this right now. For some other issues though, it’s a hardware problem that an OTA won’t fix so they issue a recall to repair the problem (ex: when the wiring harness for their cameras was fraying the cables).

This is 100% from the NHTSA shenanigans, though.

Kbobabob,

What typically happens when a recall is issued for other vehicles? Don’t they either remove and replace the bad part or add extra parts to fix the issue?

How is removing bad code and replacing it with good code or just adding extra code to fix the issue any different?

Do you want to physically go somewhere?

filcuk,

Kinda, as the word implies. If it’s a software update, call it that; the car’s not going back to the shop/manufacturer.

Kbobabob,

It sounds like location is important for some reason.

Jakeroxs, (edited )

Here’s an example of why I don’t like that they’re called recalls when it’s just a system update, if you have a recall on a food item, is there some way to fix it aside from taking it back (to be replaced) or throwing it away?

When there’s a security patch released on your phone, do we call it a recall on the phone? Or is that reserved for when there a major hardware defect (like the Samsung Note fiasco)

Kbobabob,

I think the difference in the case you mentioned is that with a car they use recall because it could be dangerous to keep using it as is.

Jakeroxs,

Fair, it just seems like there should maybe be a new word for this era where an OTA update is all that’s needed.

MNByChoice,

They often are. Many recalls for other manufacturers are similar. They don’t actually buy back the cars and crush them.

nxdefiant,
cestvrai,

Hmm, so it’s only designed to handle expected scenarios?

That’s not how driving works… at all. 😐

wahming,

Face it, that’s actually better than many drivers can do

samus12345,
@samus12345@lemmy.world avatar

They thought the truck was being driven by Sarah Conner.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • technology@lemmy.world
  • DreamBathrooms
  • ethstaker
  • khanakhh
  • magazineikmin
  • InstantRegret
  • rosin
  • Youngstown
  • slotface
  • thenastyranch
  • everett
  • kavyap
  • tacticalgear
  • mdbf
  • Durango
  • provamag3
  • tester
  • cisconetworking
  • osvaldo12
  • ngwrru68w68
  • cubers
  • GTA5RPClips
  • modclub
  • anitta
  • megavids
  • normalnudes
  • Leos
  • JUstTest
  • lostlight
  • All magazines