Tesla driver who killed 2 people while using autopilot must pay $23,000 in restitution without having to serve any jail time (fortune.com)
from sanqueue@lemmy.world to technology@lemmy.world on 16 Dec 2023 12:53
https://lemmy.world/post/9635151

#technology

threaded - newest

autotldr@lemmings.world on 16 Dec 2023 12:55 next collapse

This is the best summary I could come up with:


A Tesla driver will pay more than $23,000 in restitution for the deaths of two people during a 2019 car crash in a Los Angeles suburb, a decision announced the same day that the automaker recalled nearly all vehicles sold in the U.S.

Wednesday’s court hearing wrapped up a case believed to be the first time in the U.S. prosecutors brought felony charges against a motorist who was using a partially automated driving system.

The recall affects more than 2 million Tesla vehicles and will update software and fix a defective system that’s supposed to ensure drivers are paying attention when using Autopilot.

The Tesla driver in the Los Angeles case, Kevin Aziz Riad, pleaded no contest to two counts of vehicular manslaughter with gross negligence.

Authorities say Aziz Riad, a limousine service driver, was at the wheel of a Tesla Model S that was moving at 74 mph (119 kph) when it left a freeway and ran a red light on a local street in Gardena, California, on Dec. 29, 2019.

The Tesla, which was using Autopilot at the time, struck a Honda Civic at an intersection, and the car’s occupants, Gilberto Alcazar Lopez and Maria Guadalupe Nieves-Lopez, died at the scene.


The original article contains 364 words, the summary contains 198 words. Saved 46%. I’m a bot and I’m open source!

qooqie@lemmy.world on 16 Dec 2023 12:59 next collapse

Wow the value of a life I guess. I don’t really know what can come close to the value of a life, but this doesn’t seem like it.

burliman@lemm.ee on 16 Dec 2023 13:10 next collapse

What would be the value of life then? I’ll save you the answer: no matter how big the number you say, someone else will say bigger. Until it becomes priceless, which is the answer.

However death and accidental death isn’t always avoidable. And when we pin the fault on someone we cannot expect to say “priceless” is what they owe the victim’s family. So we assign an amount of money or time that hurts, and call it good.

Doesn’t mean life is worth that. And saying so doesn’t help anyone.

a4ng3l@lemmy.world on 16 Dec 2023 13:18 next collapse

Sure but even looking a only the financial produce of one person for a family dwarfs the comical 23k here. And that’s not even looking at the emotional side of things. 23k is straight insulting imho.

ghastly_03_startup@infosec.pub on 16 Dec 2023 13:38 next collapse

Seeing as they were using the now being recalled Tesla auto pilot during the auto accident, it may not entirely be the drivers fault. This may be part of the rational behind the judgement.

AFaithfulNihilist@lemmy.world on 16 Dec 2023 14:00 collapse

Tesla should be out millions for this. The autopilot feature is a gimmick and not at all transparent. They’re beta testing on the public and people are dying because of it. This is a corporate decision that needs to have corporate consequences over and above legal ones. People shouldn’t just be getting minor fines, they should be going to prison and losing absolutely everything.

MNByChoice@midwest.social on 16 Dec 2023 14:40 collapse

Consistent and clear pattern of lying to everyone, minimizing, and shifting blame with clear motivation of personal profits.

Fully agree on prison time.

mindlessscrollingparrot@kbin.social on 16 Dec 2023 13:39 next collapse

Two people were killed, so you're really talking 11.5k.

Tarquinn2049@lemmy.world on 16 Dec 2023 17:42 collapse

That’s life insurances job, this would be on top of life insurance, and is more about where the money comes than where it is going.

TenderfootGungi@lemmy.world on 16 Dec 2023 14:01 next collapse

True. But what if Tesla has to pay a billion for producing software that runs people over? They probably would not have beta software on the road.

PeepinGoodArgs@reddthat.com on 16 Dec 2023 14:23 collapse

The U.S. uses the value of statistical life VSL. Here are the numbers from the Department of Transportation over the last 10 years or so.

So, it is interesting and egregious that the driver needs only pay $23K and Tesla pays nothing at all!

catastrophicblues@lemmy.ca on 16 Dec 2023 19:46 collapse

So according to this, the DoT values a life at $12.5M in 2022? I’m curious about their methodology.

onion@feddit.de on 17 Dec 2023 12:25 collapse

You look at different jobs, how high the risk of dying is and how much they pay, and work out from that how much more pay people demand for say a 1% risk increase. Then you scale that up to 100% risk.

So if you were to work an average job no one has ever survived, and you died on the day you retire, you would’ve earned those 12mil

catastrophicblues@lemmy.ca on 18 Dec 2023 22:16 collapse

Huh that’s kinda neat. Thanks!

[deleted] on 16 Dec 2023 13:15 next collapse

.

Neato@kbin.social on 16 Dec 2023 14:34 next collapse

That was the penalty for the felony charge for the driver of the car that ran off the highway into a surface street. It's almost certain that drivers insurance also paid out their maximum.

In addition, Tesla is recalling all those cars to change the system that pretends to ensure a driver using autopilot is actually paying attention.

And a civil suit will likely follow from the 2 victims families.

[deleted] on 16 Dec 2023 17:51 next collapse

.

dylanmorgan@slrpnk.net on 16 Dec 2023 18:36 collapse

That is just the fine, the families are suing the driver and Tesla. Here’s hoping the Tesla suit gives them the real prize: the death of a company.

(I know it won’t happen but a guy can dream.)

samus7070@programming.dev on 16 Dec 2023 13:56 next collapse

The real crime is marketing the driver assist capability under the name autopilot when it is anything but that.

TheGrandNagus@lemmy.world on 16 Dec 2023 14:05 next collapse

Oh no, it’s even worse than that.

It’s the CEO and other staff repeatedly speaking of the system as if it’s basically fully capable and it’s only for legal reasons why a driver is even required. Even saying that the car could drive from one side of the US to the other without driver interaction (only to not actually do that, of course).

It’s the company never correcting people when they call it a self driving system.

It’s the company saying they’re ready for autonomous taxis and saying owner’s cars will make money for them while they aren’t driving it.

It’s calling their software subscription Full Self Driving

It’s honestly staggering to me that they’re able to get away with this shit.

meleecrits@lemmy.world on 16 Dec 2023 14:24 collapse

I love my Model 3, but everything you said is spot on. Autopilot is a great driver assist, but it is nowhere near autonomous driving. I was using it on the highway and was passing a truck on the left. The road veered left and the truck did as well, keeping in its lane the entire time. The car interpreted this as the truck merging over into my lane and slammed the brakes. Fortunately, I was able to figure out what went wrong and quickly accelerated myself so as to not become a hazard to the cars behind me.

Using Autopilot as anything more than a nice dynamic cruise control setting is putting your life, and other lives, in danger.

Neato@kbin.social on 16 Dec 2023 14:31 next collapse

Holy shit. My car doing that once and I'd be a nervous wreck just thinking about using it again.

burliman@lemm.ee on 16 Dec 2023 14:45 next collapse

That’s the bar that automatic driving has. It messes up once and you never trust it again and the news spins the failure far and wide.

Your uncle doing the same thing just triggers you to yell at him, the guy behind him flips you off, he apologizes, you’re nervous for a while, and you continue your road trip. Even if he killed someone we would blame the one uncle, or some may blame his entire class at worst. But we would not say that no human should drive again until it is fixed like we do with automated cars.

I do get the difference between those, and I do think that they should try to make automated drivers better, but we can at least agree about that premise: automated cars have a seriously unreasonable bar to maintain. Maybe that’s fair, and we will never accept anything but perfect, but then we may never have automated cars. And as someone who drives with humans every day, that makes me very sad.

maynarkh@feddit.nl on 16 Dec 2023 15:20 next collapse

There is a big difference between Autopilot and that hypotethical uncle. If the uncle causes an accident or breaks shit, he or his insurance pays. Autopilot doesn’t.

By your analogy, it’s like putting a ton of learner drivers on the road with unqualified instructors, and not telling the instructors that they are supposed to be instructors, but that they are actually taking a taxi service. Except it’s somehow their responsibility. And of course pocketing both the instruction and taxi fees.

The bar is not incredibly high for self driving cars to be accepted. The only thing is that they should take the blame if they mess up, like all other drivers.

burliman@lemm.ee on 16 Dec 2023 23:25 collapse

Yeah, for sure. Like I said, I get the difference. But ultimately we are talking about injury prevention. If automated cars prevented one less death per mile than human drivers, we would think they are terrible. Even though they saved one life.

And even if they only caused one death per year we’d hear about it and we might still think they are terrible.

Neato@kbin.social on 16 Dec 2023 15:38 next collapse

The difference is that Tesla said it was autopilot when it's really not. It's also clearly not ready for primetime. And auto regulators have pretty strict requirements about reliability and safety.

While that's true that autonomous cars kill FAR less people than human drivers, ever human is different. If an autonomous driver is subpar and that AI is rolled out to millions of cars, we've vastly lowered safety of cars. We need autonomous cars to be better than the best driver because, frankly, humans are shit drivers.

I'm 100% for autonomous cars taking over entirely. But Tesla isn't really trying to get there. They are trying to sell cars and lying about their capabilities. And because of that, Tesla should be liable for the deaths. We already have them partially liable: this case caused a recall of this feature.

Staiden@lemmy.dbzer0.com on 16 Dec 2023 18:02 collapse

But the vaporware salesman said fully automatic driving was 1 year away! In 2018, 2019, 2020, 2021… he should be held responsible. The guy once said to further technology some people will die and that’s just the price we pay. It was in a comment about going to Mars, but we should take that in to accout for everything he does. If I owned a business and one of my workers died or killed someone because of gross negligence I’d be held responsible why does he get away with it.

SlopppyEngineer@discuss.tchncs.de on 16 Dec 2023 14:50 collapse

Except Tesla’s uncle had brain damage and doesn’t really learn from the situation so will go it again, and had clones of him driving thousands of other cars.

KpntAutismus@lemmy.world on 16 Dec 2023 15:37 next collapse

i barely trust the lane-keeping assistant in my friend’s car. imagine going 70+km/h and suddenly the car decides to jerk the steering to the left/right because you weren’t exactly in the middle of your lane.

fuck modern assistants IMO. i can use the steering wheel just fine, and people have been able to for a hundred years.

pennomi@lemmy.world on 16 Dec 2023 16:35 next collapse

Considering that driving is (statistically) the most dangerous thing the average person does, I wouldn’t really say that people use the steering wheel just fine.

It’s just that computers are currently worse at it than humans.

KpntAutismus@lemmy.world on 16 Dec 2023 17:01 next collapse

agreed. if “autopilot” becomes a better driver than the average person, then it has a right to exist.

PlutoParty@programming.dev on 16 Dec 2023 17:24 collapse

Despite autopilot’s flaws, this is already true, if we are speaking statistically.

pennomi@lemmy.world on 16 Dec 2023 17:51 collapse

In not entirely sure if I trust the statistics that are available for a couple reasons (and feel free to correct me if I’m wrong):

  1. They are self reported by the manufacturer
  2. Systems like autopilot will revert to manual control when it detects a situation it can’t handle, which means it has the luxury of “not being at fault during the crash” when it may have caused the situation 5 seconds before
  3. It’s comparing to all vehicles instead of just vehicles that have similar non-self-driving but effective safety features
FaceDeer@kbin.social on 16 Dec 2023 20:46 collapse

I wouldn't even say that without seeing statistics to back it up. The news doesn't cover routine traffic accidents, but one Tesla screws up one thing and that story is front page. Don't rely on anecdotes and emotions.

merc@sh.itjust.works on 17 Dec 2023 21:44 collapse

i can use the steering wheel just fine, and people have been able to for a hundred years.

People have been bad at it for a hundred years. I’m not saying that people should necessarily be using auto-steering that keeps them in the middle of their lanes, but they should at least be using systems that beep at them when they stray out of their lane.

The bar for self-driving technology isn’t some amazing perfect computer that never makes a mistake. It’s the average driver. The average driver is bad.

KpntAutismus@lemmy.world on 18 Dec 2023 00:50 collapse

we can do two things (these are not mutually exclusive):

-take further control away from the drivers and make them dependent on a computer, which can always misunderstand a situation and make the driver responsible for it.

-educate drivers properly, at least in the US. americans have been historically bad at driving and have also been known to be undereducated.

merc@sh.itjust.works on 18 Dec 2023 04:24 collapse

I’m all for more driver education, and for stricter licensing requirements like they have in Europe. Having said that, eventually computers are going to have to take over.

It’s pretty absurd that we’re handing control over multi-ton devices traveling at tens of meters per second to fallible, bored, easily distracted humans. The safer cars get, the safer drivers feel. The safer drivers feel, the less they feel they need to concentrate on driving.

Safe driving just will never be a skill that humans will be good at. The tasks that humans are good at that require concentration are tasks that are challenging and remain challenging. Think playing a sport where there’s always action and you have to react. Humans are bad at tasks that are mostly routine and boring, but if your concentration lapses you can cause a catastrophe. Those are the kinds of tasks where people get bored so they start glancing away, reading a book or looking at a smartphone, or whatever. For driving to be engaging, it has to be non-boring, which means non-safe. The safer it gets, the more boring it gets, so people stop paying the required attention. There’s just no winning.

snooggums@kbin.social on 16 Dec 2023 16:05 next collapse

I have had the adaptive cruise control brake on multiple Hondas and Subarus in similar situations. Not like slamming on the brakes, but firm enough to confuse the hell out of me.

Every time it was confusing and now I just don't use it if the road is anything but open and clear.

buran@lemmy.world on 16 Dec 2023 17:38 collapse

Honda’s sensing system will read shadows from bridges as obstructions in the road that it needs to brake for. It’s easy enough to accelerate out of the slowdown, but I was surprised to find that there is apparently no radar check to see if the obstruction is real.

My current vehicle doesn’t have that issue, so either the programming has been improved or the vendor for the sensing systems is a different one (different vehicle make, so it’s entirely possible).

Wrench@lemmy.world on 16 Dec 2023 17:43 collapse

I give teslas more room because I have been brake checked by them on empty roads before. These ghost brake problems are prevalent.

Alchemy@lemmy.world on 16 Dec 2023 14:32 next collapse

Your cars actions could kill someone.

Speculater@lemmy.world on 16 Dec 2023 17:34 collapse

That’s only a $11.5k fine though.

Damage@slrpnk.net on 16 Dec 2023 18:17 next collapse

Something like that happened to me while using adaptive cruise control on a rental Jeep Renegade, it slammed the brakes twice on the highway but for no clear reason. I deactivated it before it tried a third one.

NeoNachtwaechter@lemmy.world on 16 Dec 2023 19:41 next collapse

The road veered left and the truck did as well, keeping in its lane the entire time. The car interpreted this as the truck merging over into my lane and slammed the brakes.

Even dynamic cruise control must never do such dangerous mistakes!

You should claim that they fix this on warranty, and they should prove that this is never going to happen again.

LordKitsuna@lemmy.world on 17 Dec 2023 22:53 collapse

Almost all of them do it, the one most fresh in my mind is the Prius because my work uses them as base cards so I drive them a lot. If the highway curves kind of hard to either the left or the right sometimes it will panic and think you’re about to hit the car in the lane next to you because they’re technically in front of you and so it will try to brake.

Thankfully there is an option to turn off the automatic braking it will just start screaming instead

LordKitsuna@lemmy.world on 17 Dec 2023 22:52 collapse

The auto cruise on the Priuses at work do this a lot. If the freeway curves to the left or something it will panic and think I’m about to hit the cars in the lane next to me also going through the Curve

Neato@kbin.social on 16 Dec 2023 14:29 next collapse

Tesla should be playing wrongful death suits every time autopilot kills someone. Their excuses don't excuse the blatant marketing that leads people to believe it's a self driving car.

800XL@lemmy.world on 16 Dec 2023 14:44 collapse

But you see that wasn’t the vehicle’s fault. It’s been programmed perfectly. What happened was the fault of the pedestrians and driver for not properly predicting what the car would do.

maybe /s maybe not.

Goferking0@ttrpg.network on 16 Dec 2023 15:35 collapse

no you see the issue is that the auto pilot stopped right before the accident so obviously it was entirely drivers fault, please don’t check how much time was between it stopping and the accident

raptir@lemdro.id on 16 Dec 2023 15:52 next collapse

Do we need to go through what autopilot in a plane or boat actually does again?

dexa_scantron@lemmy.world on 16 Dec 2023 16:32 next collapse

If we do, then they shouldn’t have picked a name that most people think does something it doesn’t.

RushingSquirrel@lemm.ee on 17 Dec 2023 11:57 collapse

When you drive a Tesla, it’s pretty clear what autopilot is. The name is a marketing term, you can’t engage it everywhere and anytime, you’ve got to keep your hands on the wheel or it disables itself, won’t stop at stop signs and red lights, won’t do line changes, etc.

PatFussy@lemm.ee on 16 Dec 2023 16:42 next collapse

<img alt="" src="https://pbs.twimg.com/media/Eh10gEKVoAIZkU7.jpg:large">

kool_newt@lemm.ee on 16 Dec 2023 17:13 next collapse

It doesn’t matter, Tesla cars are marketed to the public which isn’t expected to know these things. To probably 90% of people “autopilot” means “drive automatically”.

CmdrShepard@lemmy.one on 17 Dec 2023 03:24 collapse

To probably 90% of people “autopilot” means “drive automatically”.

Based on what?

poopkins@lemmy.world on 17 Dec 2023 11:32 next collapse

Tesla markets this feature as “Full Self-Driving Capability.” Maybe I’m poorly informed, but to me that means that the car is fully capable of driving itself without human interaction.

CmdrShepard@lemmy.one on 17 Dec 2023 11:36 collapse

FSD is an entirely separate thing. Autopilot is just an LKAS system, or adaptive cruise control.

poopkins@lemmy.world on 17 Dec 2023 11:46 collapse

Aha, today I learned that Autopilot is just lane-keeping and adaptive cruise control. I feel that it must be a common misunderstanding to confuse the terms “Autopilot” and “Fully Self-Driving” in the vernacular.

Many other manufacturers refer to lane-keeping systems as “driver assistance,” and I believe Tesla is intentionally misleading consumers with the impression that their system is more capable and allows the driver to pay less attention.

RushingSquirrel@lemm.ee on 17 Dec 2023 11:54 collapse

Until you drive it. You know the capabilities, you know when you can and cannot activate it, you know how often it tells you to look at the road and if you don’t prove you’ve got your hands on the wheel, it disables itself for the drive (you need to park to reactivate it). No Tesla driver thinks autopilot is more than a lane and distance keeping assistance.

Autopilot is a marketing name, that’s it.

kool_newt@lemm.ee on 17 Dec 2023 17:28 collapse

Based on my usage and understanding of the word being a lay person.

I’m an engineer myself, sometimes there are words that you have to be cognizant of the differences in meaning to other engineers vs lay people or even engineers in other fields. Some words are heavily overloaded, and “autopilot” is kinda one of them (others being “domain”, “node”, “artificial intelligence”, etc.).

doublejay1999@lemmy.world on 16 Dec 2023 20:17 next collapse

What does full self driving mean ?

CmdrShepard@lemmy.one on 17 Dec 2023 03:23 collapse

Full Self Driving and Autopilot are two totally separate systems.

fiah@discuss.tchncs.de on 16 Dec 2023 16:03 next collapse

do we need to go through the differences in training, aptitude and intelligence between pilots, captains and your neighbor Greg again? Marketing it as “autopilot” to anyone who can sign a car loan is reckless and has killed people and will continue to kill people until they stop

CmdrShepard@lemmy.one on 17 Dec 2023 03:32 collapse

Yep, just like “cruise control” made tons of people drive their car into the ocean thinking they could sail it to popular island destinations.

merc@sh.itjust.works on 17 Dec 2023 21:44 collapse

Depends entirely on the type of autopilot.

iamtherealwalrus@lemmy.world on 16 Dec 2023 16:32 next collapse

Way to ignore the death of two people, and hijack the discussion for your own opinions. Good job /s

dylanmorgan@slrpnk.net on 16 Dec 2023 18:29 next collapse

I think the real crime is vehicular manslaughter, especially the SECOND one.

Fox@pawb.social on 17 Dec 2023 16:31 collapse

It’s a common misunderstanding that an autopilot system in an airplane does everything or even a lot of things. The most basic ones keep the wings level and nothing else. Of course Tesla is probably counting on that misconception to sell this feature, but actual pilots using any kind of autopilot are still on the hook to pay attention 100% of the time.

menemen@lemmy.world on 17 Dec 2023 23:17 collapse

In an airplane that is fine as pilots are specifically trained on the planes they fly (at least in theory). No one gets a special course in how to drive a specific (non industrial) car…

TenderfootGungi@lemmy.world on 16 Dec 2023 13:59 next collapse

If you want to kill someone in the US with little consequences, run them over with a car.

ladicius@lemmy.world on 16 Dec 2023 14:48 collapse

Germany the same. Small fine, three month without license, that’s it for killing a human being.

Pechente@feddit.de on 16 Dec 2023 15:39 next collapse

Holy shit, really? Never looked into it but judging by how people drive here (lots of people on their phones while driving, missing red lights all the time) it certainly doesn’t seem like there are severe consequences for any wrongdoings.

Thorny_Insight@lemm.ee on 16 Dec 2023 15:46 collapse

If we’re talking about an honest accident then how long do you think the jail term should be?

Pyr_Pressure@lemmy.ca on 16 Dec 2023 16:14 next collapse

If it were an honest accident then nothing. If it were due to neglect or lack of due diligence then maybe a few months of of weekend jail or month of full time jail.

MisterFrog@lemmy.world on 18 Dec 2023 04:26 collapse

For killing someone? Causing someone’s death due to negligence is only worth a month of jail to you?

sbv@sh.itjust.works on 16 Dec 2023 18:21 next collapse

“honest accident” is the crux of the question. If the driver was doing everything perfectly and some other party was entirely responsible for the accident, not much (maybe none?).

But, at least in my corner of Canada, most drivers are not behaving responsibly or adhering to the law. Speeding, following too closely, illegally passing, and using phones while driving are common. If a driver kills someone while doing something overtly dangerous, they deserve jail time.

Dark_Arc@social.packetloss.gg on 18 Dec 2023 06:04 collapse

I hate that speed(ing) always gets lumped in whenever “dangerous behavior” comes up. Going faster than an arbitrary road sign says you should isn’t inherently dangerous.

  • Going faster than the arbitrary road sign can be dangerous.
  • Going the speed the arbitrary road sign says can be dangerous.
  • Going slower than the arbitrary road sign can be dangerous.

It’s about the conditions of the road, paying attention, signaling to other drivers what you’re trying to do, and being prepared for people and animals to do something dumb.

  • following too closely
  • using phones while driving

These things are on a whole other level than speeding or “illegally” passing. But the person who can’t keep their car centered in the lane, wrecks every other winter, doesn’t use their turn signal, doesn’t notice an ambulance right behind them, and drives too close to the car in front of them will say “I’m a GOOD driver because I don’t speed. Shame on all these bad drivers that pass me!”

Even worse some of those “GOOD” non-speeding drivers will try to “police the roads” and prevent people from getting around them which has literally resulted in completely pointless deaths during emergencies.

sbv@sh.itjust.works on 18 Dec 2023 12:10 collapse

Googling around, it looks like there’s a strong relationship between increased speed and accident severity. The reasons cited are increased kinetic energy of the vehicle, decreased effectiveness of the built-in safety equipment, and a higher risk of rollover. It’s particularly dangerous for pedestrians and cyclists.

Increased speed also increases the risk of an accident, since it reduces the amount of time drivers have to react, and increases the vehicles stopping distance.

From

(Interestingly, speeding decreases fuel efficiency, but that isn’t what the original post was about)

Dark_Arc@social.packetloss.gg on 18 Dec 2023 16:27 collapse

Yes, but in Texas you’ll find the interstate speed limit is 85 and in Ohio it was 65, now it’s 70.

There’s not some fundamentally crazy difference between Ohio and Texas roads that results in Texas accepting 20mph higher speed limits than the rate Ohio was using within the last 10 years or so.

There is a history of speed limits for fuel efficiency and scarcity en.wikipedia.org/wiki/National_Maximum_Speed_Law. Which is more evidence that speed limits do not inherently correspond to safety. Driving the limit, over the limit, or under the limit doesn’t “magically” make you safer.

Sure, if you’re in an accident two objects traveling faster are going to cause more damage to each other than two objects traveling slower. However, getting into that accident in the first place has little to do with speed and a lot to do with situational awareness and giving yourself enough time to stop in case of emergency (i.e. how close you’re following the car in front of you, how fast you’re passing kids on the sidewalk, etc).

You’ll note speed is always blamed for increasing severity (i.e., it’s a factor in the severity of the accident). Unfortunately the links don’t work anymore to get to the underlying source but sites.psu.edu/…/is-driving-faster-safer/ states “A study conducted by the Florida Department of Transportation says that accidents that were caused by speeding is actually 2.2%.” The Autobahn is another great example. Speed is very rarely cited as the primary cause … because speed isn’t really the issue.

We should be focusing on issues that actually cause accidents like tailgating, blocking the left lane, failing to signal, etc.

grue@lemmy.world on 17 Dec 2023 17:58 next collapse

If we’re talking about an honest accident

There is no such thing.

helenslunch@feddit.nl on 17 Dec 2023 22:13 collapse

In the US that’s called “involuntary manslaughter” and presumably the sentences are shorter, but also still exist so people can’t throw up their hands and go “whoops! That was totes an accident” and get off scot-free.

Alchemy@lemmy.world on 16 Dec 2023 14:00 next collapse

Anyone else tired of beta testing Tesla’s garbage just by being outside on the roads near these vehicles?

RushingSquirrel@lemm.ee on 17 Dec 2023 12:07 collapse

Human beings controlling cars are extremely dangerous. Drunk drivers, racing, going through red lights and stops, speeding, not paying attention, etc. No need for autopilot for the streets to be dangerous for pedestrians. Autopilot keeps the car in line, which is already way safer than most 100% human-controlled accidents.

And again, the driver is responsible to keep their eyes on the road, even when using cruise-control or any sort of driving assistance.

[deleted] on 17 Dec 2023 15:56 collapse

.

LEDZeppelin@lemmy.world on 16 Dec 2023 14:51 next collapse

Why so low penalties? Were the victims black or brown?

14th_cylon@lemm.ee on 16 Dec 2023 15:25 next collapse

oh, it is another crash…

the driver was really a victim of the tesla/uber experiment. first - someone in tesla or uber decided to turn off the emergency brake assistent because it was giving out what they considered too many false positive alarms. the car at the moment of the accident knew about the pedestrian and tracked him, but the emergency brake which should have engaged was turned off. and then the driver was thrown under the bus by uber.

it is really hard to pay attention as a driver when there is really nothing to pay attention to. hard to blame the driver.

Thorny_Insight@lemm.ee on 16 Dec 2023 15:48 next collapse

Do you think higher penalties would make people less prone for accidents?

RushingSquirrel@lemm.ee on 17 Dec 2023 12:13 collapse

When the accident is caused by a blatant lack of paying attention, I believe it would. Not paying attention, causing death of someone? 2 years in jail. You’re responsible of what you do.

Rai@lemmy.dbzer0.com on 16 Dec 2023 22:01 next collapse

the car’s occupants, Gilberto Alcazar Lopez and Maria Guadalupe Nieves-Lopez, died at the scene.

Yeaaaaaah…

lolcatnip@reddthat.com on 17 Dec 2023 00:20 collapse

What do you consider an appropriate fine for someone who’s using a product in the manner approved by laws and the manufacturer when the product malfunctions and kills someone?

RushingSquirrel@lemm.ee on 17 Dec 2023 12:11 collapse

They’re not using the product in the manner approved by the manufacturer at all. Driver is 100% responsible in this case and 23k is an absolute insult to the victims and the judicial system.

We are missing some info regarding why the penalty was so low, though. With the details from the article, the sentence doesn’t make any sense.

Pxtl@lemmy.ca on 16 Dec 2023 15:43 next collapse

Same as it ever was Kill a Pedestrian, Pay a $500 Fine

Come join the war on cars. !fuckcars@lemmy.world

KpntAutismus@lemmy.world on 16 Dec 2023 15:50 collapse

the problem here is the law. there should be actual consequences, not fines. jail time for murder.

aeharding@lemmy.world on 16 Dec 2023 17:11 collapse

Part of the reason why you don’t lose your license for killing someone with a car in the US is because it’s much more of a ‘punishment’ because of how car dependent the US is.

Also, keep in mind a lot of trips are 3 miles or less in the US, and most drive it, despite wanting alternatives to driving.

If someone is trying to get from A to B in a 2 mile trip and the government basically mandates people to drive that, can you really blame them if they end up killing someone accidentally? What if they accidentally kill themselves smashing into a tree? You might assign some of the blame to their driving, but would that solve anything in the long term? a large part of the blame should be assigned to this insane transportation system we’ve built where everyone needs to drive 2 miles to pick up a bag of milk.

TLDR prevention, not blame will reduce traffic violence.

kool_newt@lemm.ee on 16 Dec 2023 17:20 next collapse

I think many don’t use the alternatives because there are significant challenges vs using a car you’re already paying for.

I’d love to bicycle but it’s just not safe.

my_hat_stinks@programming.dev on 17 Dec 2023 09:16 collapse

You’re already paying for a car because your infrastructure demands one. I’m not in the US and get by just fine without one. Saves me a shitload of money too.

Cycling is unsafe because your infrastructure was built exclusively for cars. Your infrastructure is built exclusively for cars because most people have to use cars because your infrastructure was built exclusively for cars. It’s circular reasoning.

Both of those issues are caused by infrastructure and would be solved by building slightly less stupid infrastructure.

vaultdweller013@sh.itjust.works on 17 Dec 2023 12:27 next collapse

Also theres the edge cases like mine where walking and biking are vastly slower than driving even though the city is somewhat walkable, because I live at one of the highest elevated roads in my city. Its a 15 min walk or a 5 min bike to places, its a 1 hr walk back and a 30 min bike back.

kool_newt@lemm.ee on 17 Dec 2023 17:04 collapse

Which is a good reason for people that want bicycle infra to vote and run for office!

KpntAutismus@lemmy.world on 16 Dec 2023 18:07 collapse

proper education for drivers and decent bike lanes are key. absolutely agree.

PatFussy@lemm.ee on 16 Dec 2023 16:39 next collapse

The guy was going through a suburb at 75 mph blowing through stop lights. Ofcourse he has to pay, im surprised hes not getting jail time. This has nothing to do with the car, thats just gross negligence

eltrain123@lemmy.world on 16 Dec 2023 17:26 next collapse

My tesla doesn’t let me use autopilot or FSD if I set it over 5% of the posted speed limit. How is this guy going 75 in the burbs?

Tarquinn2049@lemmy.world on 16 Dec 2023 17:43 next collapse

That is unfortunately a configurable option.

BossDj@lemm.ee on 16 Dec 2023 17:44 collapse

I wonder if Tesla had this section of road mapped as freeway. Especially since it rolled through a red.

“Suburb” in LA is a very loose term

I would think that the guy is just lying, but Tesla would call that out REAL quick.

asdfasdfasdf@lemmy.world on 17 Dec 2023 03:42 collapse

My interpretation of the title is “only has to pay…”. 23K is nothing.

Nacktmull@lemm.ee on 16 Dec 2023 18:30 next collapse

Well, he didn´t do anything … /s

doublejay1999@lemmy.world on 16 Dec 2023 20:19 next collapse

2 murders, 23 grand. The mafia charge more.

If you wanna kill somebody, use a car.

Potatos_are_not_friends@lemmy.world on 16 Dec 2023 23:05 next collapse

The $11,500 “murder” add-on

lolcatnip@reddthat.com on 17 Dec 2023 00:12 collapse

Involuntary manslaughter ≠ murder

psycho_driver@lemmy.world on 16 Dec 2023 21:04 next collapse

For fuck’s sake I doubt if that would cover funeral expenses.

pHr34kY@lemmy.world on 18 Dec 2023 08:02 collapse

It won’t even fix their car.

maryjayjay@lemmy.world on 17 Dec 2023 16:08 next collapse

Civil suit. He’s already been proven guilty

Commiunism@lemmy.wtf on 17 Dec 2023 16:59 next collapse

There’s this saying about how if something is punishable by a fine, then it’s only illegal for poor people.

I don’t even have to finish this do I

fosforus@sopuli.xyz on 17 Dec 2023 17:05 next collapse

Finland’s fine system at least tries: some fines scale based on the perp’s monthly income.

Example: theguardian.com/…/finnish-businessman-hit-with-12…

I’m unshamedly proud of this. Apparently Switzerland has the same system.

isVeryLoud@lemmy.ca on 17 Dec 2023 17:19 collapse

I love this

PM_Your_Nudes_Please@lemmy.world on 17 Dec 2023 22:18 collapse

There’s a joke that if you want to murder someone in America, make sure you do it in a car. Our courts are specifically tailored to avoid penalizing drivers for “accidentally” killing people.

BlackNo1@lemmy.world on 18 Dec 2023 00:57 next collapse

uh is that it?

werefreeatlast@lemmy.world on 18 Dec 2023 03:38 next collapse

You honor, I actually didn’t wack anyone with this self actuating axe. I bought it and I told it to go chop wood. The people just happened to be too close to the axe. Yeah I was holding the axe but I wasn’t actually putting any pressure. The tail was wagging the dog in other words.

Ok so $10,000.00. Fine? Oh alright I guess that’ll teach me not to buy autonomous axes.

jabjoe@feddit.uk on 18 Dec 2023 07:06 next collapse

Fines = legal for a price.

JadenSmith@sh.itjust.works on 18 Dec 2023 12:22 collapse

Fines only exist to punish the poor.

Cannacheques@slrpnk.net on 18 Dec 2023 07:30 next collapse

Oh come on, only two people?

testuserpleaseupvote@lemmy.world on 18 Dec 2023 12:36 next collapse

American taxpayers will pick up the rest of the bill. Nice subsidy for the rich.

Roflmasterbigpimp@lemmy.world on 18 Dec 2023 12:54 next collapse

So $11.500 per Person. Huh. I would have guessed it that american Lives would be more expensive.

helenslunch@feddit.nl on 17 Dec 2023 22:09 collapse

Wait, I’m confused, I thought this was Tesla’s fault?