this post was submitted on 05 Oct 2023
231 points (95.3% liked)

Technology

58143 readers
5397 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Victim in critical condition

all 20 comments
sorted by: hot top controversial new old
[–] blewit@lemm.ee 44 points 11 months ago (1 children)

“And we know what you're thinking: would the second car have still hit her if it was being driven by a person? It's a good question. Did being under AI control really make a difference? ”

No, what I was thinking was: would the first car have still hit her if it was being driven by AI?

For sure we know it wouldn’t have driven away like the person that hit her initially did.

[–] abhibeckert@lemmy.world 30 points 11 months ago* (last edited 11 months ago) (2 children)

Cruise avoiding pedestrian

In that screenshot, you can see a pedestrian is moving towards the road (but not yet on the road proper, he is still between two parked cars) and the Cruise has already decided to illegally drive with two wheels on the wrong side of the road, over a double yellow line, to avoid driving close to the pedestrian.

Two seconds later in that video the pedestrian has run into the path of the driverless Cruise car, and the car has stopped. You can view the full video here:

https://twitter.com/kvogt/status/1602766709806747648

[–] Turun@feddit.de 14 points 11 months ago

This is a completely different incident.

[–] Pika@sh.itjust.works 10 points 11 months ago* (last edited 11 months ago) (1 children)

Not to downplay this situation because it is a bad situation regardless but looking at the multiple articles that they have released on this and the lack of video that they have provided regarding it, without seeing the video it's hard to say who's at fault here. Well aside from the driver who hit and ran obviously.

It sounds like both vehicles had a green light at the intersection and I expect the second lane wouldn't have been able to see the pedestrian that was crossing in the first Lane to begin with. The article States the vehicle when it saw the pedestrian "braked aggressively" in order to try to not hit the person. I don't think that this is as simple as a oh there was a person waiting at the crosswalk so it shouldn't have gone, a lot of these intersections also have pedestrian lights on both sides of the crosswalk. Every article is blaming the autonomous vehicle, but I really don't think an actual driver would have done things differently given what has been released. In fact they might have actually made things worse by immediately driving off the person's leg/ankle instead of waiting for the lift.

I'll be interested to see footage when it's released cuz I'm curious this was an actual mistake on the autonomous vehicles part or if this was a it did what it could not actually being able to be avoided

[–] Pseu@kbin.social 15 points 11 months ago (1 children)

A car driven by a human is unlikely to need firefighters to lift the vehicle up to get at the woman pinned by its tire. Even if they're good at general driving they have an unfortunate habit of making emergencies worse.

[–] MeanEYE@lemmy.world -1 points 11 months ago

Then again, car driven by a machine won't text and do makeup at the same time. I've seen so many idiots in traffic it hurts to try and remember them all. That said humans are pretty good at driving, unless they start acting like idiots and do something else while driving.

[–] tigeruppercut@lemmy.zip 7 points 11 months ago (1 children)

Just today the Some More News folks put out a great breakdown of a lot of the current issues and shady shit going on in the AI car industry.

https://www.youtube.com/watch?v=pmGOjHi-7MM

[–] PipedLinkBot@feddit.rocks 3 points 11 months ago

Here is an alternative Piped link(s):

https://www.piped.video/watch?v=pmGOjHi-7MM

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I'm open-source; check me out at GitHub.

[–] swope@kbin.social 3 points 11 months ago

I think most computer vision cameras are fairly low resolution, so I'm not expecting the hit-and-run vehicle will be identified.

[–] fiat_lux@kbin.social 1 points 11 months ago (1 children)

One of the big problems with how organisations often work, especially private businesses, is the extremely casual attitude to product testing and risk assessment. It's only after they spend shitloads on lawyers and public relations that they are suddenly able to prioritise creating jobs dedicated explicitly to preventing the damage they cause with that attitude.

But because law-making is slow, weighed down by flawed human power structures combined with legitimately necessary procedures, the only thing businesses need to do is to outpace the speed of law change to avoid being punished. Outpacing the law has been easy enough to do at the best of times, but with half-assed exploitative software development in a rapidly progressing robotics and 'AI Boom' environment, it will only get easier and hurt more people.

And then the executives who allowed their shoddy products to hurt people will just change employers, likely for a pay raise or just selling the business outright. The only consequences for their reckless management personally are a few late nights in a bad mood. All because limited liability meant they might as well have just been an innocent bystander.

Meanwhile the victims - if they survive, are left in lifelong pain and misery, because courts ruled that the law doesn't cover their novel situation. Not to mention the damage to their families and communities.

Globally, we need to start holding individual organisation decision-makers to personal account for the damage their decisions cause. Both financial and prison-time, for both environmental and human damage. I mean like "Board of Directors and all Chief Officers of Cruise on trial for negligent homicide" levels of responsibility. It's the only way to prevent this kind of unnecessary suffering.

tl;dr
1. Risk of personal loss is the only way people in power will prioritise building safer products.
2. We need the law to catch up faster to a world where humans can offload more life-changing decisions to computers.
3. Law-makers should start assuming we live on the Star Trek holodeck in a Q episode instead of the Unix epoch, if they are going to catch up on their huge backlog.
4. People need to start assuming their code is imprecise and dangerous and build in graceful failures. Yes, it will be expensive in a time-sensitive environment, important things often are.

[–] PersnickityPenguin@lemm.ee -1 points 11 months ago* (last edited 11 months ago) (2 children)

I've read about the cruse team, and "extremely casual attitude to product testing" does not accurately describe what they are doing. The cruse vehicles have a much lower and less severe accident rating than human drivers, and have logged millions of road miles without seriously injuring anyone (until now).

Unlike a certain narcissistic auto manufacturers owner...

[–] Turun@feddit.de 1 points 11 months ago

Oh, did they actually release data and had an independent research group analyse it? Or is this a statement from their PR department? It's easy to be better than the average human driver if you only drive in good weather and well built roads.

Tesla always makes big claims about how safe it is, but to the best of my knowledge never actually released any usable data about it. It would be awesome if cruise did that.

[–] fiat_lux@kbin.social 0 points 11 months ago (1 children)

'Less shit than the average human at it' is a really low bar to set for modern computers, even if Tesla fails at that poor standard and Cruise is currently top of the game. We still need much higher bars when we're talking about entirely automated systems which are controlling speedy large chunks of metal, or even other smaller-scale-impact-and-damage systems. Systems which can't just hop out, ask if the victims are OK, render appropriate first aid, accurately inform emergency services, etc.

The more automation, the higher the standards should be, which means we need to set legal requirements that at least try to scale with he development of technology.

[–] philodendron@lemdro.id 0 points 11 months ago (1 children)

I disagree. Human drivers kill over 40,000 Americans a year. If there’s an alternative that kills less than 40,000 a year we should take it. Ideally mass transit but America seems to like cars.

[–] fiat_lux@kbin.social 1 points 11 months ago

I wasn't suggesting stopping the development of automated vehicles because it's impossible to have 0 damage. I was advocating having high standards for software/hardware development and real consequences for decision-makers trying to find shortcuts.

Progress and standards are not mutually exclusive.

[–] PHPete@sh.itjust.works -4 points 11 months ago

I find this bizarrely reminiscent of Brandon Lee's death.