Tesla braces for its first trial involving Autopilot fatality (www.reuters.com)
from L4s@lemmy.world to technology@lemmy.world on 28 Aug 2023 18:00
https://lemmy.world/post/4018382

Tesla braces for its first trial involving Autopilot fatality::Tesla Inc is set to defend itself for the first time at trial against allegations that failure of its Autopilot driver assistant feature led to death, in what will likely be a major test of Chief Executive Elon Musk’s assertions about the technology.

#technology

threaded - newest

autotldr@lemmings.world on 28 Aug 2023 18:00 next collapse

This is the best summary I could come up with:


SAN FRANCISCO, Aug 28 (Reuters) - Tesla Inc (TSLA.O) is set to defend itself for the first time at trial against allegations that failure of its Autopilot driver assistant feature led to death, in what will likely be a major test of Chief Executive Elon Musk’s assertions about the technology.

Self-driving capability is central to Tesla’s financial future, according to Musk, whose own reputation as an engineering leader is being challenged with allegations by plaintiffs in one of two lawsuits that he personally leads the group behind technology that failed.

The first, scheduled for mid-September in a California state court, is a civil lawsuit containing allegations that the Autopilot system caused owner Micah Lee’s Model 3 to suddenly veer off a highway east of Los Angeles at 65 miles per hour, strike a palm tree and burst into flames, all in the span of seconds.

Banner’s attorneys, for instance, argue in a pretrial court filing that internal emails show Musk is the Autopilot team’s “de facto leader”.

Tesla won a bellwether trial in Los Angeles in April with a strategy of saying that it tells drivers that its technology requires human monitoring, despite the “Autopilot” and “Full Self-Driving” names.

In one deposition, former executive Christopher Moore testified there are limitations to Autopilot, saying it “is not designed to detect every possible hazard or every possible obstacle or vehicle that could be on the road,” according to a transcript reviewed by Reuters.


The original article contains 986 words, the summary contains 241 words. Saved 76%. I’m a bot and I’m open source!

tmRgwnM9b87eJUPq@lemmy.world on 28 Aug 2023 18:17 next collapse

Although it’s far from perfect, autopilot gets into a lot less accidents per mile than drivers without autopilot.

They have some statistics here: www.tesla.com/VehicleSafetyReport

EDIT: As pointed out by commenters in this thread, autopilot is mainly used on high ways, whereas the crash average is on all roads. Also Tesla only counts a crash if the airbag was deployed, but the numbers they compared against count every crash, including the ones without deployed airbags.

MoonlitSanguine@lemmy.one on 28 Aug 2023 18:29 next collapse

Do you have statistics not by Tesla?

Thorny_Thicket@sopuli.xyz on 28 Aug 2023 19:02 collapse

They’re probably the only ones who even has access to such statistics. If you’re simply just going to refute the stats because of the source then atleast provide some credible counter evidence.

bernieecclestoned@sh.itjust.works on 28 Aug 2023 19:08 next collapse

forbes.com/…/tesla-again-paints-a-very-misleading…

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 03:31 collapse

Even according to that article autopilot and FSD seems to be about at the level as human driver. I’m willing to accept that - many others arent.

The narrative here is that these systems are dangerous and shouldn’t be allowed to be used on public roads. My argument is that they’re not as dangerous as reading stories about these individual incidents might make them seem like and they’re getting better all the time. If they’re not significantly better than human drivers now they will be soon and Tesla most likely is going to lead the way.

silvercove@lemdro.id on 28 Aug 2023 20:13 collapse

If you’re simply just going to refute the stats because of the source then atleast provide some credible counter evidence.

Tesla’s numbers are trash. Tesla have been caught again and again lying.

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 03:18 collapse

…then provide some more trustworthy stats because you just saying that is not it. This is literally like debating a climate change denier or flat earther.

"Here’s a picture of the earth from space"

  • Lies! Nasa cannot be trusted. CGI.
silvercove@lemdro.id on 29 Aug 2023 06:38 collapse

Why would Tesla release any numbers that would make it look bad?

NocturnalMorning@lemmy.world on 28 Aug 2023 18:34 next collapse

Oh yeah, potentially cherrypicked statistics straight from Tesla. I’ll believe those statistics when they come from someone not with a horse in the race to adopt autonomous vehicles.

1rre@discuss.tchncs.de on 28 Aug 2023 19:30 next collapse

What’s the motivation to cherrypick though?

Human drivers are bad enough that I don’t think there’s any doubt that autopilot puts them to shame with regards to safety, so they can either look way better and not be suspicious, or look way better and be suspicious… Sounds like an obvious choice to me

NocturnalMorning@lemmy.world on 28 Aug 2023 20:08 next collapse

They have a financial motivation. You ucould also just Google self driving car safety, and one of the first Google hits is an article that calculated the safety of human drivers from data collected in 2021. Turns out humans are already pretty damn safe, there’s roughly 99.9998 of driving with zero accidents.

1rre@discuss.tchncs.de on 28 Aug 2023 20:42 collapse

there’s roughly 99.9998 of driving with zero accidents

I assume you mean accidents with a fatal injury, given there is a ~1% chance that any given death will be from a car accident (17.4 deaths per 100k per year * 70 years = 1.2%) - using your statistic yields closer to 2.5% however this works with only one driver dying.

Turns out humans are pretty damn safe

Turns out you’ve been tricked by statistics, driving is fucking lethal and chances are most people know or are friends with someone who has died or will die in a car accident (assuming ~80 friends/acquaintances per person)

Honytawk@lemmy.zip on 29 Aug 2023 13:33 collapse

Why the arbitrary number of 70 years?

If you calculate the chance of having an accident per route traveled (about 2.57 per person per day), you get a number much closer to NocturnalMorning’s statistics.

1rre@discuss.tchncs.de on 29 Aug 2023 15:37 collapse

A rough estimate for global life expectancy. It’s actually slightly over 73, so the chances of dying in a car accident are marginally higher than I said.

The data I used wasn’t related to driving frequency or age, it was purely the number of people in a random global sample of 100,000 people you would expect to die in a car accident in a given year. That of course includes people of all ages and people who never drive at all, but also taxi & HGV drivers. Even if we say people aren’t in cars so much under the age of 5 or over the age of 60, that would push up the deaths per 100,000 people per year between 5 and 60 by the exact amount to keep the chance per year over a human lifetime at 17.4/100000.

silvercove@lemdro.id on 28 Aug 2023 20:12 collapse

I don’t think there’s any doubt that autopilot puts them to shame with regards to safety

Where are the numbers to back this up?

1rre@discuss.tchncs.de on 28 Aug 2023 20:32 collapse

Start with the numbers on humans driving drunk, tired, on their phone, while having a conversation, bored or in practically other state and work backwards. Driving is dangerous as fuck and it’s pretty much universally accepted that the biggest challenge for autonomous vehicles is humans doing unpredictable and stupid shit

silvercove@lemdro.id on 28 Aug 2023 20:55 collapse

that the biggest challenge for autonomous vehicles is humans doing unpredictable and stupid shit

The biggest challenge when I’m driving is humans or AI doing unpredictable and stupid shit.

You still have not given any numbers to back up your claim. While we all expect that AI will one day be much better than humans in driving, there is no data to say that it currently is.

1rre@discuss.tchncs.de on 28 Aug 2023 21:45 collapse

Ok so sure there’s nothing on Tesla’s autopilot, however that’s not to say there’s nothing on autonomous systems…

www.ncbi.nlm.nih.gov/pmc/articles/PMC8431415/

In 2018 and 2017, 6,735,000 and 6,453,000 traffic crashes occurred in the United States, which resulted in 33,919 and 34,560 deaths, respectively.

orsa.org.uk/…/reducing-driver-error-accidents/

In reality, car crashes aren’t accidents and 94% are due to human error In 2011, British police officers attended 118,404 road traffic collisions (figures from the Department of Transport). In 42% of these crashes, the most frequently reported factor was that the driver ‘failed to look properly’. The second most commonly listed factor for 21% of the crashes was the driver ‘failing to judge the other person’s path or speed’. The third most common contributing factor was the driver being actually ‘careless, reckless or in a hurry’ and this accounted for 16% of the crashes.

There’s your stats on humans being reckless and dangerous when driving cars, and of course there’s nothing concrete for fully autonomous cars because they aren’t legal anywhere, but here’s some stats on pretty much every existing driver assist - notably they all prevent accidents compared to just a human driving: www.ncbi.nlm.nih.gov/pmc/articles/PMC8431415/

It really isn’t a stretch from the 3 most frequent crash causes being human error and human assistance tools reducing accident frequency a bunch to say that all these systems coming together (as they cover near enough everything to do with driving a car) would be safer than a human driver, but I don’t doubt you’ll deny it as you’re asking for something impossible to give (as governments haven’t allowed full autonomous driving cars yet, so there’s no statistics on their use) and so aren’t actually looking for information but to confirm your biases and feel like you’ve “won”, despite the fact there’s no objectively unsuspicious data on the exact situation you’re asking for meaning that you can’t prove yourself right either beyond “I’m a little suspicious of this company so I must be right”

underisk@lemmy.ml on 28 Aug 2023 19:38 collapse

I think it’s been reported that the FSD statistics they put out are worthless because it tends to disable itself right before collisions.

decerian@lemmy.world on 28 Aug 2023 18:38 next collapse

Those stats are misleading though. Autopilot only runs on highways, which are much safer per mile even for human drivers.

Tesla are basically comparing their system, which only runs in pristine, ideal conditions, against an average human that has to deal with the real world.

As far as I’m aware they haven’t released safety per mile data from the FSD cars yet, and until they do I will remain skeptical about how much safer it currently is.

skymtf@pricefield.org on 28 Aug 2023 18:40 next collapse

Doesn’t auto pilot kinda work on normal roads now? Not saying I trust the stat either.

ephemeral_gibbon@aussie.zone on 28 Aug 2023 21:48 collapse

That’d be the fsd stats, not autopilot

Thorny_Thicket@sopuli.xyz on 28 Aug 2023 19:00 next collapse

According to this report, the average Tesla equipped with FSD Beta, driven on predominantly non-highway sections of road, crashes 0.31 times per million miles, a dramatic decrease from the average American, who crashes 1.53 times every million miles.

Source

fuckwit_mcbumcrumble@lemmy.world on 28 Aug 2023 19:07 collapse

Does that report from Tesla include when autopilot turns off shortly before crashing into something?

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 03:37 collapse

That’s literally the only data we have so that’s what I’m basing my opinion to while being fully aware that while I doubt that these stats lie they may however be misleading as statistics often are.

My key argument still stands; autopilot/FSD is not as bad/dangerous as people here make them to be and they’re getting better all the time.

If one is going to make the claim that these systems are more dangerous than human driver then show me the data you’re basing it on. People surely don’t think that just because they don’t like the company/CEO, right?

tmRgwnM9b87eJUPq@lemmy.world on 28 Aug 2023 19:52 next collapse

Hmm you’re right about autopilot mainly being used on highways and those roads are a lot safer. I’ll edit my main comment

Asifall@lemmy.world on 28 Aug 2023 20:34 collapse

It actually would be really hard to get an unbiased estimate of safety given the current systems, because the data is inherently cherry picked by drivers who can switch the feature on/off depending on how complex the driving task is. What a simple number like crashes per mile really measures is really how likely FSD drivers are to overestimate the system’s ability plus some unknown base rate of unavoidable accidents.

Probably the only way to control for this is looking at cars that are fully autonomous door to door and aren’t limited to pre-selected roads/areas. I don’t know that anyone is even doing that sort of testing.

MajesticSloth@lemmy.world on 28 Aug 2023 19:00 next collapse

This reminds me when you google if a certain company or product is good or legit and the top one is posted from the companies website.

RecallMadness@lemmy.nz on 28 Aug 2023 19:03 next collapse

And when autopilot is at fault for an accident or fatality, who should be held responsible?

Just because it’s better, shouldn’t absolutely them of responsibility when it fails.

severien@lemmy.world on 28 Aug 2023 19:26 next collapse

It’s an interesting question. But I would be disappointed if the self-driving was basically killed by the legal questions, since it has a huge potential to save lives.

excel@lemmy.megumin.org on 28 Aug 2023 19:40 collapse

The driver is always responsible for using the tools within the car correctly and maintaining control of the vehicle at all times.

Either way the driver would be at fault. However, the driver might be able to make a (completely separate) case that the car’s defects made control impossible, but since the driver always had the option to disable self-driving, I doubt that would go anywhere.

Just like you don’t get off the hook if your cruise control causes an accident… and it doesn’t matter how much Tesla lied about what it may or may not be capable of, because at the end of the day it’s always the driver’s responsibility to know the limitations of the vehicle and disable the feature and take control when necessary.

RecallMadness@lemmy.nz on 28 Aug 2023 19:58 next collapse

Which is exactly what this case is claiming, that the software is defective.

And what happens when we progress beyond Level 2 or 3 automation? Then the car is making choices for the driver, choices the driver may not have any say in or realistically be capable of reacting to in an emergency?

Deferring responsibility to the driver under any scenario is a cop-out. We have a long history of engineering qualifications and regulations to ensure safety of the populace, engineers and architects design structures to be safe, plumbers have to plumb to code, heck even cars themselves have a mile long list of compliance requirements. All to ensure the thing that companies build aren’t killing the population, and when they do someone is responsible.

Yet as soon as we start talking about software, “not my problem dawg.”.

tony@lemmy.hoyle.me.uk on 29 Aug 2023 07:19 collapse

This is a guy who was using a glorified cruise control (which is all AP is) at high speed whilst watching a DVD instead of looking at the road.

The software can only help so much. There’s a reason why there are laws requiring attentiveness checks now… people are reckless

Honytawk@lemmy.zip on 29 Aug 2023 13:37 collapse

People are only reckless because they believe Teslas false marketing claims.

The car doesn’t “just drive itself”, it isn’t even close to “just driving itself”. The advertising claiming so is much more at fault than the driving watching a movie.

theneverfox@pawb.social on 30 Aug 2023 10:38 collapse

So you’re correct to call it a tool, with this level of automation the driver is ultimately the operator. But you’re missing something

Did you misuse the tool, did they sell you a bad tool, or did their instructions cause the tool to be misused?

The first is as you said - if I make and sell you a circular saw and you cut your finger off being an idiot, that’s on you.

If the thing flew apart under normal use, that’s on me - it’s likely my responsibility, and possibly negligence.

If the box or user manual said it is for wood and metal use, and it’s actually entirely unsafe for metal use, that’s probably negligence on my part

Cruise control doesn’t unexpectedly jerk your wheel to the side, if it did and you could prove you were using it reasonably and in the recommended way, you’d almost definitely get off the hook

silvercove@lemdro.id on 28 Aug 2023 20:12 collapse

Why should we trust any numbers that comes from Tesla?

silvercove@lemdro.id on 28 Aug 2023 20:14 next collapse

Autopilot is not safe.

washingtonpost.com/…/tesla-autopilot-crashes-elon…

CmdrShepard@lemmy.one on 29 Aug 2023 00:13 next collapse

Driving is not safe. These systems could be improved upon, but they’ve also saved numerous lives by preventing accidents from occurring in the first place. The example in the OP happened while this driver was sitting behind the wheel watching a movie. The first example in your article occurred with a driver behind the wheel. If either of them had been driving a 1995 Honda Civic, these accidents would have occurred just the same, but would anyone be demanding that Honda is to blame?

pup_atlas@pawb.social on 29 Aug 2023 04:06 next collapse

No, we would (rightfully so) blame the driver for merging into a semi truck that from my understanding was clearly visible.

silvercove@lemdro.id on 29 Aug 2023 06:37 collapse

but they’ve also saved numerous lives by preventing accidents from occurring in the first place.

There is no data to make this claim. You’re just making this up.

CmdrShepard@lemmy.one on 29 Aug 2023 11:09 collapse

Give me a break. You think all these companies are dumping billions of dollars into technology that doesn’t work? You’re making stuff up. Go watch some dashcam videos on YouTube if you want some proof.

silvercove@lemdro.id on 29 Aug 2023 11:34 next collapse

Are you kidding me? I never said it will never work. But that does not mean its current state is safe to trust your life.

CmdrShepard@lemmy.one on 29 Aug 2023 12:01 collapse

You did in fact just say that by saying that I was making up the fact that these systems have saved lives. Moving the goalposts to “you can’t trust your life to it” doesn’t make your original argument anymore accurate nor does it reference anything in dispute. Nobody said you should trust your life to cruise control.

Honytawk@lemmy.zip on 29 Aug 2023 13:28 next collapse

Nobody did indeed say you should trust your life to cruise control.

But Tesla did claim you could trust your life to autopilot because “the car basically drives itself”, which it obviously doesn’t.

CmdrShepard@lemmy.one on 29 Aug 2023 16:39 collapse

Tesla didn’t claim that. Musk claimed their early FSD “basically drove itself” in what appears to have been a staged demonstration. This accident and lawsuit are about Autopilot, which is a completely different system.

silvercove@lemdro.id on 29 Aug 2023 15:18 collapse

There is no doubt that one day these systems will be so good that they will make transportation much safer. But there is no data that shows that we’re already there.

CmdrShepard@lemmy.one on 29 Aug 2023 16:34 next collapse

You mean you’ve done zero research on the topic before injecting your opinions, so you simply haven’t seen any data?

thedriven.io/…/accident-rate-for-tesla-80-lower-t…

New data released in its Impact Report show that Tesla vehicles with Autopilot engaged (mostly highway miles) had just 0.18 accidents per million miles driven, compared to the US vehicle average of 1.53 accidents per million miles.

monash.edu/…/The-Potential-Benefits-of-LKAS-in-Au…

A statistically significant 16% reduction in the risk of involvement in all casualty crashes of these types and a 22% reduction estimated for fatal and serious injury crashes was associated with LKA fitment to Australian light vehicle was estimated.

pubmed.ncbi.nlm.nih.gov/27624313/

The analysis showed a positive effect of the LDW/LKA systems in reducing lane departure crashes. The LDW/LKA systems were estimated to reduce head-on and single-vehicle injury crashes on Swedish roads with speed limits between 70 and 120 km/h and with dry or wet road surfaces (i.e., not covered by ice or snow) by 53% with a lower limit of 11% (95% confidence interval [CI]). This reduction corresponded to a reduction of 30% with a lower limit of 6% (95% CI) for all head-on and single-vehicle driver injury crashes (including all speed limits and all road surface conditions).

forbes.com/…/vehicle-safety-features-accidents/

ADAS functionalities can change the driving experience. According to research by LexisNexis Risk Solutions, ADAS vehicles showed a 27% reduction in bodily injury claim frequency and a 19% reduction in property damage frequency.

silvercove@lemdro.id on 29 Aug 2023 19:51 next collapse

We’re not talking about ADAS in general, we’re talking about what Tesla is selling.

CmdrShepard@lemmy.one on 29 Aug 2023 21:27 collapse

What Tesla is selling was covered in the first link. If you disagree, either provide proof to the contrary or quit “making things up.”

silvercove@lemdro.id on 29 Aug 2023 22:04 collapse

That first link is Tesla advertisement, not independently acquired data. It’s worthless.

CmdrShepard@lemmy.one on 29 Aug 2023 22:43 collapse

I thought you were all about sourced information though. Where’s your evidence that it’s inaccurate?

silvercove@lemdro.id on 30 Aug 2023 05:54 collapse

What is the evidence that is it accurate? Is there any reason to trust Tesla’s advertisements?

CmdrShepard@lemmy.one on 30 Aug 2023 15:00 collapse

Ah, you have no evidence to back your argument up so you’re just going to dig your heels in and quite hypocritically make unfounded claims. Got it.

Autopilot has been around for nearly a decade now but “it doesn’t work” and you’re the only person on the planet to figure it out in all that time, which is why you can’t provide any proof. I’ll bet you acquired this super secret knowledge without ever setting foot inside a Tesla too, right?

silvercove@lemdro.id on 30 Aug 2023 15:17 collapse

You’re the one making unfounded claims. Tesla advertisements are not data.

CmdrShepard@lemmy.one on 30 Aug 2023 18:04 collapse

So prove it or stop making things up. Show us data proving that Teslas crash at the same rate as all those vehicles without ADAS systems. Should be easy for someone so confident in their (thus far) completely unfounded and unsourced opinion. You haven’t been able to provide a single shred of evidence backing your claim. Why is that? Where is the independent verification for your claim since that’s something important to you?

silvercove@lemdro.id on 30 Aug 2023 18:32 collapse

Seeing cars pile up on the San Francisco tunnel due to Tesla’s phantom breaking crap is good enough reason.

Yet you have failed to provide a single drop of evidence to support the disinformation you’re spewing.

CmdrShepard@lemmy.one on 30 Aug 2023 21:08 collapse

Lol so when you rear-end someone, it’s the fault of the person in front? Not sure where you learned how to drive, but they failed you.

The phantom braking is an issue, but again that’s another example of the driver failing to control their vehicle. Like all other systems, Autopilot warns you multiple times when it’s going to disengage.

Funny you claim I’ve failed to provide a single drop of evidence when I’ve posted numerous links. You’re, what, ten comments in now, having demanded evidence numerous times, yet have failed to provide even a single link to back any of your made-up claims. That’s pretty telling and a bit embarrassing if you ask me.

silvercove@lemdro.id on 31 Aug 2023 06:31 collapse

Haha, go away troll. Your links are Tesla advertisements, and yes if you break in fast flowing traffic you are at fault and you will cause an accident.

rambaroo@lemmy.world on 30 Aug 2023 08:43 collapse

Lol the only relevant link is the first one, which comes from Teslas cherry picked and thoroughly disproven data set.

rambaroo@lemmy.world on 30 Aug 2023 08:41 collapse

Actually there is some doubt about that. Completely irrelevant to the present either way though.

chakan2@lemmy.world on 29 Aug 2023 21:52 collapse

billions of dollars into technology that doesn’t work?

Absolutely. Heard of the F22?

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 04:03 next collapse

Driving a car is not safe. 40000 people die on car crashes every year in the US alone. Nothing in that article indicates that autopilot/FSD is more dangerous than a human driver. Just that they’re flawed systems as is expected. It’s good to keep in mind that 99.99% safety rating means 33000 accidents a year in the US alone.

silvercove@lemdro.id on 29 Aug 2023 06:37 next collapse

You can’t just put something on the streets without first verifying it’s safe and working as intended. This is missing for Autopilot. And the data that’s piling up is showing that Autopilot is deadly.

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 06:48 next collapse

First of all what is it that you consider safe? I’m sure you realize that 100% safety rating is just fantasy so what is the acceptable rate of accidents for you?

Secondly would you mind sharing the data “that’s piling up is showing that Autopilot is deadly” ? Reports of individual incidents is not what I’m asking for because as I stated above; you’re not going to get 100% safety so there’s always going to be individual incidents to talk about.

You also seem to be talking about FSD beta and autopilot interchangeably thought they’re a different thing. Hope you realize this.

silvercove@lemdro.id on 29 Aug 2023 06:53 collapse

There are very strict regulations around what is allowed to be in the streets and what isn’t. This is what protects us from sloppy companies releasing unsafe stuff in the streets.

Driver assist features like the Autopilot are operating in a regulatory grey zone. The regulation has not caught up with technology and this allows companies like Tesla to release unsafe software in the streets, killing people.

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 06:59 collapse

Exactly. Driver assist features. These aren’t something to be blindly relied on and everyone knows this and the vehicle will remind you. Every crash is fault of the driver - not the system.

Now if you don’t mind showing me the data that’s “piling up is showing that Autopilot is deadly”

silvercove@lemdro.id on 29 Aug 2023 07:07 collapse

Exactly. Driver assist features.

Except Tesla isn’t selling them as such. Theid advertisement videos as early as 2016 say “the driver is not necessary, the car is driving itself”. This is false marketing in its purest and simplest form: theguardian.com/…/tesla-self-driving-video-staged…

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 07:36 collapse

I’m still waiting for the data that you said is piling up. You also did not specify what number of accidents you find acceptable for a self driving system. It’s almost like you’re trying to evade my questions…

silvercove@lemdro.id on 29 Aug 2023 11:36 next collapse

Give me a breakn The WaPo article is linked above. Also, when it comes to safety, the burden of proof is on those arguing that something is safe.

Thorny_Thicket@sopuli.xyz on 29 Aug 2023 18:50 collapse

If there’s piles of data it shouldn’t be difficult to prove it’s unsafe.

You still haven’t even specified what is considered safe.

silvercove@lemdro.id on 29 Aug 2023 19:52 next collapse

Stop talking nonsense. The anecdotes that are piling up clearly indicate there is a problem with Tesla’s autopilot.

rambaroo@lemmy.world on 30 Aug 2023 08:40 collapse

It’s in the weapon article you were already linked, you just keep choosing to ignore it. Another user on my blocked list.

Honytawk@lemmy.zip on 29 Aug 2023 13:27 collapse

Do you think Tesla would get sued if the data wasn’t piling up?

jimbolauski@lemmy.world on 29 Aug 2023 13:42 collapse

You can say the exact same thing for people.

firadin@lemmy.world on 29 Aug 2023 17:36 next collapse

Exactly, you can’t just drive without verifying that you’re a safe driver. That’s why we have a process to get a driver’s license. Has Autopilot passed licensing?

jimbolauski@lemmy.world on 30 Aug 2023 01:02 collapse

So if an autonomous car can drive around the block and parallel park it’s licensed?

rambaroo@lemmy.world on 30 Aug 2023 08:38 next collapse

Humans have a lower accident rate than Tesla’s autopilot, it says so in the article itself.

jimbolauski@lemmy.world on 30 Aug 2023 12:15 collapse

I don’t see that claim anywhere in the article.

In fact any comparisons I’ve found show Tesla’s autopilot performing better than humans. One crash 4.41 million miles driven on autopilot in a tesla vs one crash for every 1.2 million miles in a tesla without autopilot, NHTSA’s most recent data shows that in the United States there is an automobile crash every 484,000 miles.

cleantechnica.com/…/tesla-1-crash-per-4-41-millio…

AssholeDestroyer@lemmy.ml on 30 Aug 2023 13:03 collapse

Yeah thats why we make drivers get licenced.

jimbolauski@lemmy.world on 31 Aug 2023 10:18 collapse

Being able to drive around the block and parallel park is enough for self driving to be certified?

paintbucketholder@lemmy.world on 29 Aug 2023 19:28 next collapse

Former NHTSA senior safety adviser Missy Cummings, a professor at George Mason University’s College of Engineering and Computing, said the surge in Tesla crashes is troubling.

“Tesla is having more severe — and fatal — crashes than people in a normal data set,” she said in response to the figures analyzed by The Post.

This would indicate that FSD is more dangerous than a human driver, would it not?

chakan2@lemmy.world on 29 Aug 2023 21:50 next collapse

Depends…did you read that study on Twitter or another source?

rambaroo@lemmy.world on 30 Aug 2023 08:35 collapse

It’s from the Washington Post article linked in the parent comment. Come tf on dude. You look like a douche accusing people of using Twitter as a source when the actual source is literally in the same thread.

chakan2@lemmy.world on 30 Aug 2023 14:18 collapse

It was a joke about Twitter users. Of course FSD is more dangerous than a human. It took all 0f 20 minutes for it to try to run a red on Musk.

Thorny_Thicket@sopuli.xyz on 30 Aug 2023 03:18 collapse

That still doesn’t tell are those accidents happening more compared to normal cars. If you have good driver assist systems which are able to prevent majority of minor crashes but not the severe ones then the total number of crashes goes down but the kinds that remain are the bad ones.

rambaroo@lemmy.world on 30 Aug 2023 08:32 collapse

They are in accidents at higher rates than the normal data set so that’s exactly what it says.

dustyData@lemmy.world on 30 Aug 2023 12:32 collapse

Humans my friend. We can hold humans accountable. We can’t hold hunks of semi-sentient sand and nebulous transient configurations of electrons liable of anything. So, it has to be better than humans, which is not. If it isn’t better than humans, then we’ll rather just have a human in control. Because we can argue with and hold the human accountable for their actions and decisions.

T156@lemmy.world on 29 Aug 2023 07:27 collapse

Isn’t it a glorified cruise control/lane guidance system, rather than an actual automated driving system? So it would be about as safe as those are, rather than being something that you can just leave along to handle its own business, like a robotic vacuum cleaner.

fat_stig@lemmy.world on 29 Aug 2023 13:32 next collapse

It is just a shit load of if then else statements. If the inputs don’t have a corresponding if then it just defaults to doing nothing.

skyspydude1@lemmy.world on 29 Aug 2023 20:54 collapse

The main issue is that they market it like a fully autonomous system, and made it just good enough that it lulls people into a false sense of security that they don’t need to pay attention, while also having no way to verify they are, unlike other systems from BMW, GM, or Ford.

Other systems have their capabilities intentionally hampered to insure that you’re not going to feel it’s okay to hop in the passenger seat and let your dog drive.

They are hands-on driver assists, and so they are generally calibrated in a way that they’ll guide you in the lane, but will drift/sway just a bit if you completely take your hands off the wheel, which is intended to keep you, y’know, actually driving.

Tesla didn’t want to do that. They wanted to be the “best” system, with zero safety considerations at any step other than what was basically forced upon them by the supplier so they wouldn’t completely back out. The company is so insanely reckless that I feel shame for ever wanting to work for them at one point, until I saw and heard many stories about just how bad they were.

I got to experience it firsthand too working at a supplier, where production numbers were prioritized over key safety equipment, and while everyone else was willing to suck it up for a couple of bad quarters, they pushed it and I’m sure it’s indirectly resulted in further injuries and potentially deaths because of it.

RojoSanIchiban@lemmy.world on 30 Aug 2023 00:57 next collapse

This is an absolutely bald-faced lie. Tesla absolutely does NOT market Autopilot as fully autonomous system. Autopilot is nothing other than lane-centering and adaptive cruise control with emergency braking, and that’s it. There is zero ambiguity about it on the vehicle and in documentation. Plus, it specifically requires the driver to maintain control of the wheel.

You need to stop, drop, and roll or jump in the nearest pool before your pants burn you to a crisp.

skyspydude1@lemmy.world on 30 Aug 2023 01:57 next collapse

Oh really? Is that why for years now, on the front page for Autopilot on Tesla’s site, was the infamous “Paint it Black” demo, where in the first 10 seconds it says “The driver only there for legal reasons, the car is driving itself”? What do you think is going to stick in the mind of a potential buyer: that video of the car “driving itself” right on the Tesla website, or the generic 5 line page that you’ll see in basically every single car with a satnav these days saying, “Please operate the car safely”?

Regardless of how much people like you love to get into the technicalities and differences between Autopilot and Full Self Driving and chime in with “ACKSHUALLY” and insert any number of the same tired responses about how autopilot works on aircraft or what it says in the documentation, it changes nothing about how they’ve shaped the public perception of their system and how people are going to attempt and use it.

Stop defending their shitty practices. Literally everyone else has figured out how to prevent people from abusing these systems, Tesla won’t even bother, because people like you will step in and defend it every time for some fucking reason, and as a bonus it saves them money.

themajesticdodo@lemmy.world on 30 Aug 2023 02:07 next collapse

Autopilot is nothing other than lane-centering and adaptive cruise control with emergency braking, and that’s it.

When you put it that way, the term Autopilot does sound really misleading.

MataVatnik@lemmy.world on 30 Aug 2023 05:33 next collapse

In have this product named Telephone. I absolutely do NOT market Telephone as a remote long distance voice chat system. Telephone is nothing other than a voice-recording and adaptive voice control with emergency saving features, and that’s it.

rambaroo@lemmy.world on 30 Aug 2023 08:28 next collapse

They literally named it “autopilot”. Give me a break. You have zero business calling other people dishonest.

joel_feila@lemmy.world on 30 Aug 2023 08:54 collapse

Well there have people accusing Tesla of advertising their cars as much more self driving then they are. Specifically Teals has been accused of false advertising because it is what you describe, but they sell it as a self driving car.

fushuan@lemm.ee on 30 Aug 2023 07:24 collapse

hey wanted to be the “best” system, with zero safety considerations at any step other than what was basically forced upon them by the supplier so they wouldn’t completely back out. The company is so insanely reckless that I feel shame for ever wanting to work for them at one point

What does this remind me of… Oh yeah right, OceanGate

sugartits@lemmy.world on 28 Aug 2023 21:09 next collapse

The second trial, set for early October in a Florida state court, arose out of a 2019 crash north of Miami where owner Stephen Banner’s Model 3 drove under the trailer of an 18-wheeler big rig truck that had pulled into the road, shearing off the Tesla’s roof and killing Banner. Autopilot failed to brake, steer or do anything to avoid the collision, according to the lawsuit filed by Banner’s wife.

Is this the guy who was literally paying no attention to the road at all and was watching a movie whilst the car was in motion?

I legit can’t find information on it now as every result I can find online is word for word identical to that small snippet. Such is modern journalism.

I know people like to get a hard on with the word “autopilot”, but even real pilots with real autopilot still need to “keep an eye on things” when the system is engaged. This is why we have two humans in the cockpit on those big commercial jets.

ephemeral_gibbon@aussie.zone on 28 Aug 2023 21:43 next collapse

The way musk marketed it was as a “self driving” feature, not a driving assist. Yes with all current smart assists you need to be carefully watching what it’s doing, but that’s not what it was made out to be. Because of that I’d still say tesla is responsible.

CmdrShepard@lemmy.one on 29 Aug 2023 00:08 next collapse

I think you’re referring to FSD beta and not Autopilot. One is supposed to be the self driving feature at some point while the other is simply lane keeping/cruise control. FSD wasn’t even available when this crash happened.

ephemeral_gibbon@aussie.zone on 29 Aug 2023 22:14 collapse

No I was referring to autopilot, just look at the name of it. It’s I know it’s not capable of self driving (and neither is the even more absurd name of “full self driving”) but to your average person it intentionally sounds as if the car is driving itself instead of it being a driving assist.

CmdrShepard@lemmy.one on 29 Aug 2023 22:42 collapse

No you were referring to what you think Tesla said about Autopilot and I pointed out that you were mistaken.

I think it sounds like autopilot in a plane or ship, where it maintains course but still needs a pilot sitting at the controls. Regardless of what you think it is or isn’t, it’s your duty as a driver to understand the capabilities of the two ton machine that you’re voluntarily operating.

eager_eagle@lemmy.world on 29 Aug 2023 04:00 next collapse

Tesla’s Autopilot is driving assistance. I don’t know where you saw Musk marketing it as a self driving feature. Hell, even for the misnomer “full self driving” they note:

The currently enabled features require a fully attentive driver, who has their hands on the wheel and is prepared to take over at any moment.

pup_atlas@pawb.social on 29 Aug 2023 04:04 collapse

The feature is called “Autopilot”, meaning that the car automatically pilots itself, rather than using a human pilot. The definition of autopilot is literally “a device for keeping an aircraft or other vehicle on a set course without the intervention of the pilot.” I’m not sure how he could have more explicitly misrepresented the product.

eager_eagle@lemmy.world on 29 Aug 2023 04:06 collapse

meaning that the car automatically pilots itself, rather than using a human pilot

No it doesn’t. Even an airplane autopilot only maintain the course set by the pilot and it’s not capable of making decisions and navigating autonomously.

All technologies in publicly sold vehicles today and in recent years are of driving assistance and require driver’s attention. Anybody using the tech without paying attention is being negligent.

pup_atlas@pawb.social on 29 Aug 2023 04:08 next collapse

Autopilot is capable of navigating though, and it does make decisions like when to merge and when to execute a turn, by design. I don’t think it’s adequately equipped to make those decisions, but by design, it does. They even advertise it on their official YouTube channel, with a clip of them just plugging in a destination and letting the car get them there in their video. Tesla is responsible for advertising they do, and claims they make of their product that simply aren’t true.

eager_eagle@lemmy.world on 29 Aug 2023 07:36 collapse

This is FSD, not autopilot. Also note the driver is paying attention.

pup_atlas@pawb.social on 29 Aug 2023 11:54 collapse

It is two different modes of the same system, one just has more features enabled than the other. You also can’t tell if the driver is paying attention, as they are mostly out of frame. Even if they are, their hands are entirely off the wheel, and it’s unlikely that they would be able to react in time to prevent an accident even if they are paying attention.

Auli@lemmy.ca on 29 Aug 2023 12:13 collapse

Autopilot is cable of basically ying the plane itself. A human is there for when shit goes wrong.

eager_eagle@lemmy.world on 29 Aug 2023 12:22 next collapse

Only if you ignore traffic. Autopilot doesn’t take in ATC directions. But it’s not a useful comparison, air traffic and navigation is much simpler compared to ground.

candybrie@lemmy.world on 29 Aug 2023 13:17 collapse

Like if another plane is nearby? It’s not exactly just “shit going wrong.” Autopilot doesn’t follow TCAS or ATC commands for instance.

So to be similar, driver autopilot only needs to work while there are no other cars that might be in your path. Which is why we’ve had some degree of plane autopilot for nearly a century and are just starting to get some degree of car autopilot–the assumption that no cars might be in your path is pretty much always false.

jimbolauski@lemmy.world on 29 Aug 2023 13:33 collapse

Self driving is not a defined standard, it is a buzz word like increase your vitality. The SAE standards for autonomous vehicles do not have a self driving category

Auli@lemmy.ca on 29 Aug 2023 12:10 collapse

There are also two pilots. Because they know people are people. And don’t brand it a self driving and full self driving then.

dustyData@lemmy.world on 30 Aug 2023 12:38 collapse

It sends shivers down my spine to think that airlines want to eliminate the co-piloting requirement in order to reduce costs. It would be increasingly stressful for the pilots, increasing turn-over, burnout and the risk of errors during flights. I would never fly with an airline that makes a single pilot take the brunt of a flight longer than 1 hour. Hell, even quality long-distance bus travel and truck hauling companies have drivers work in tandem, switching every so many hours.

whataboutshutup@discuss.online on 28 Aug 2023 22:22 next collapse

It seems like an obvious flaw that’s pretty simple to explain. Car is learnt to operate the infromation about collisions on a set height. The opening between the wheels of a truck’s trailer thus could be treated by it as a free space. It’s a rare situation, but if it’s confirmed and reproduceable, that, at least, raises concerns, how many other glitches would drivers learn by surprise.

tony@lemmy.hoyle.me.uk on 29 Aug 2023 12:40 collapse

In most countries trucks have bars between the trailer wheels, precisely because too many car drivers got an unwelcome haircut by not paying attention.

luthis@lemmy.nz on 28 Aug 2023 23:22 next collapse

The headline makes it sound like Tesla is trialing a new ‘fatality’ feature for it’s autopilot.

CmdrShepard@lemmy.one on 29 Aug 2023 00:06 next collapse

Well, someone has to invent the suicide booths featured in Futurama. Might as well be him.

qyron@sopuli.xyz on 29 Aug 2023 07:21 collapse

I really want to trust you’re throwing a dark joke up but the sheer concept of suicide booths is a very harsh critique at a failed society. A very failed society. For it to become a joke…Call me square but that is a joke haimed to who laughs on it.

Marsupial@quokk.au on 29 Aug 2023 09:25 collapse

youtu.be/EbmQxZkSswI?si=0lcguQyWQxUggaB5

It’s a joke but a suicide booth isn’t that bad, assisted pain free death is a right everyone should have.

But having it on a street corner for ease of access is pretty fucked

qyron@sopuli.xyz on 29 Aug 2023 10:29 next collapse

My country is going through a very disputed approval over legislation for medically assisted death, for incurable conditions.

It was sent to the Constitutional Court three times and twice vetoed by the president, one for political reasons.

The majority of the population supports it.

Marsupial@quokk.au on 29 Aug 2023 10:34 next collapse

Good luck with that. Hope it can alleviate some people’s suffering.

We’ve basically got it all legal in Australia now, last state ratifies their laws in November.

qyron@sopuli.xyz on 29 Aug 2023 12:38 collapse

Oh, it’s going forward, regardless the president personal dislike (devout catholic) and the cries from the church and religious groups.

CheeseNoodle@lemmy.world on 29 Aug 2023 11:54 collapse

Making it accessalbe on what might be a fleeting impulse would be a huge problem though in the case of futurama style suicide booths.

qyron@sopuli.xyz on 29 Aug 2023 15:18 collapse

I remember reading an article about an open sourced 3D printable “suicide pod” anyone could build by themselves.

ours@lemmy.film on 29 Aug 2023 12:21 next collapse

And Futurama likes to reference many works of science fiction. Many of these cover the subject of dystopian/utopian societies where suicide is facilitated/promoted/mandated.

Futurama makes at least one direct reference to Soylent Green for one (Soylent Soda).

AssholeDestroyer@lemmy.ml on 30 Aug 2023 12:57 next collapse

The episode where Bender will explode if he says “Ass” is based of a Phillip K Dick short story.

joel_feila@lemmy.world on 30 Aug 2023 14:29 collapse

What’s it taste like

limelight79@lemm.ee on 30 Aug 2023 14:58 collapse

It varies from person to person.

joel_feila@lemmy.world on 30 Aug 2023 08:46 next collapse

plus you can select clumsy bludgeoning as a method of suicide.

Absolutemehperson@lemmy.world on 30 Aug 2023 12:38 collapse

When you say “clumsy”… how clumsy? I could go for it.

joel_feila@lemmy.world on 30 Aug 2023 14:19 collapse

I believe it was a wooden mallet just swung around followed by two ice cream scoop that go for the eyes

noughtnaut@feddit.dk on 30 Aug 2023 09:24 collapse

I take it you haven’t watched Futurama? For one, the depicted um, procedure looks rather painless-free, but also it fails entirely and the protagonist(s!) step out unscathed.

FlyingSquid@lemmy.world on 29 Aug 2023 12:18 next collapse

Mortal Kombat: Vehicle Edition

mxcory@lemmy.blahaj.zone on 29 Aug 2023 13:39 next collapse

The packet says to fight a Honda. (I know, Street Fighter, but still.)

piped.video/watch?v=3vPtn1StzA4&t=1

space@lemmy.dbzer0.com on 29 Aug 2023 18:45 collapse

Carmaggedon

Lucidlethargy@sh.itjust.works on 30 Aug 2023 00:44 next collapse

The reality is that they didn’t trial it at all, they just sent straight to production. In this case, it successfully achieved a fatality.

HiddenLayer5@lemmy.ml on 30 Aug 2023 09:17 next collapse

With how Elon has been acting this is a distinct possibility.

It would probably scream “Xterminate!” before running you over.

torpak@discuss.tchncs.de on 30 Aug 2023 09:39 collapse

I’m literally waiting for the moment when a disproportionate ammount of Musk-critics die in car crashes.

tslnox@reddthat.com on 29 Aug 2023 21:50 next collapse

I can’t understand how anyone is even able to let the car do something on its own. I drive old Dacia Logan and Renault Scénic, but at work we have Škoda Karoq and I can’t even fully trust its beeping backing sensors or automatic handbrake. I can’t imagine if the car steered, accelerated or braked without me telling it to.

Cethin@lemmy.zip on 30 Aug 2023 06:30 next collapse

I think it’s fine at the level where you are there and ready to take control, but you need to be paying attention still. Humans aren’t flawless and we shouldn’t expect our automated systems to be either. This doesn’t excuse Tesla, because they’ve been marketing it as something it’s not for a long time now. They’re driver assist features, not self driving features. It can keep you in a lane and maintain speed well, but you shouldn’t fully trust it. If it’s better than humans at some tasks, it should be used for those regardless of if it will fail at it sometimes. People shouldn’t be lied to and convinced it’s more than it is though.

[deleted] on 30 Aug 2023 08:02 next collapse

.

limelight79@lemm.ee on 30 Aug 2023 14:56 collapse

I actually think that the less a driver has to do, the worse they’ll be at reacting when a situation does come up.

If I’m actually driving and someone, say, runs out in front of me, I’ll slam on the brakes. I’ve had this happen, actually - it was scary as hell because my brain froze up, but…fortunately for us and the guy, my foot still knew what to do, and we stopped in time.

But if I’m sitting in the seat, just monitoring, not actively doing something, my attention is much more likely to wander, and when that incident happens, my reaction time is likely going to be a LOT slower, because I have to “mode shift” back into operating a car, whereas I was already in that mode in the incident above. I don’t think the manufacturers are adequately considering this factor.

(I recognize this might not be a perfect example with automatic brakes, but I think the point is clear.)

Sphks@lemmy.dbzer0.com on 30 Aug 2023 09:12 collapse

Aviation is now mostly full automatic. On the otehr hand, there are tons of beacons to help it.

TimeNaan@lemmy.world on 30 Aug 2023 09:39 next collapse

And it requires way more training and attention from the operator because that way they can react quickly. Not so much for cars, especially on “autopilot”.

Barack_Embalmer@lemmy.world on 30 Aug 2023 11:09 next collapse

It’s a difficult comparison to make because planes are maintaining level flight or making smooth wide-arcing turns or gradual changes in altitude, not quickly responding to imminent obstacles and traffic. Even in an autoland situation, it’s supposed to follow a gentle descent slope that’s planned long in advance. This type of operation isn’t really possible with cars, so they require a whole other set of considerations and techniques.

Rawdogg@lemm.ee on 30 Aug 2023 12:20 next collapse

No it’s not at all, there’s still a ton of work for the pilot and first officer despite autopilot

dustyData@lemmy.world on 30 Aug 2023 12:26 next collapse

And even private aviation requires hundreds of hours of experience and deep understanding of physics and extensive training before even being allowed in the air on your own. Let alone to fly others that’s a different training and license. Using those fancy “it flights itself” autopilots require several extra thousand hours of experience and specialized training, a commercial license and to be under the supervision and employment of an airline. Otherwise you are barely allowed to use the plane version of cruise control. Even after all that, you are still required to maintain your training with regular recertifications every few years, and a set of several hours of practice flight every year. Miss either condition and you lose your license.

AssholeDestroyer@lemmy.ml on 30 Aug 2023 12:55 collapse

There’s less stuff to hit in the air.

Mdotaut801@lemmy.world on 30 Aug 2023 11:05 collapse

Why do people buy Teslas? Sure. Tesla is at fault to a point but surely consumers have enough data at this point to know that Teslas are overpriced hunks of shit and the CEO is a total right wing snowflake. Why? Why buy one? I don’t fucking get it.

Sylvartas@lemmy.world on 30 Aug 2023 11:06 collapse

I mean, yeah, but I doubt that it’s Tesla’s official stance on the matter