Remix.run Logo
jasoncartwright 17 hours ago

If it autonomous or self-driving then why is the person in the car paying for the insurance? Surely if it's Tesla making the decisions, they need the insurance?

gizmo686 14 hours ago | parent | next [-]

Generally speaking, liability for a thing falls on the owner/operator. That person can sue the manufacturer to recover the damages if they want. At some point, I expect it to become somewhat routine for insurures to pay out, then sue the manufacturer to recover.

PunchyHamster 2 hours ago | parent | next [-]

but tesla is the operator

amelius 13 hours ago | parent | prev | next [-]

Or at some point subscribing to a service may be easier than owning the damn thing.

DaSHacka 13 hours ago | parent [-]

All according to plan

koakuma-chan 12 hours ago | parent [-]

It already doesn't make sense to own a car for me. It's cheaper to just call an Uber.

ghaff 9 hours ago | parent | next [-]

I'm guessing that's a fairly city viewpoint. My car is setup with roofrack and carries a lot of other gear I want. I'm regularly in places without reliable cell etc. Visiting friends can easily be an hour drive.

koakuma-chan 7 hours ago | parent [-]

Yes, a city viewport. I usually just walk, but when I don't I most often take the subway, not even Uber. Though I feel like in Toronto the subway or some part thereof is closed or under maintenance or whatever way too often. It's not very reliable.

gffrd 11 hours ago | parent | prev [-]

For some this is the case. For others, this is not the case.

amelius 10 hours ago | parent [-]

some -> most ?

einpoklum 11 hours ago | parent | prev [-]

Ah, but could one not argue that the owner of the self-driving car is _not_ the operator, and it is the car, or perhaps Tesla, which operates it?

kube-system 7 hours ago | parent | next [-]

All Tesla vehicles require the person behind the steering wheel to supervise the operations of the vehicle and avoid accidents at all times.

Also, even if a system is fully automated, that doesn’t necessarily legally isolate the person who owns it or set it into motion from liability. Vehicle law would generally need to be updated to change this.

sroussey 10 hours ago | parent | prev [-]

Mercedes agrees. They take on liability when their system is operated appropriately.

kube-system 7 hours ago | parent | next [-]

They say they will, but until relevant laws are updated, this is mostly contractual and not a change to legal liability. It is similar to how an insurance company takes responsibility for the way you operate your car.

If your local legal system does not absolve you from liability when operating an autonomous vehicle, you can still be sued, and Mercedes has no say in this… even though they could reimburse you.

iknowstuff an hour ago | parent | prev [-]

No. They don’t. It was vaporware made to fool people including you. You could never actually order it and it’s canceled now in favor of an L2 system.

kjksf 16 hours ago | parent | prev | next [-]

Because that's the law of the land currently.

The product you buy is called "FSD Supervised". It clearly states you're liable and must supervise the system.

I don't think there's law that would allow Tesla (or anyone else) to sell a passenger car with unsupervised system.

If you take Waymo or Tesla Robotaxi in Austin, you are not liable for accidents, Google or Tesla is.

That's because they operate on limited state laws that allow them to provide such service but the law doesn't allow selling such cars to people.

That's changing. Quite likely this year we will have federal law that will allow selling cars with fully unsupervised self-driving, in which case the insurance/liability will obviously land on the maker of the system, not person present in the car.

throwaway2037 5 minutes ago | parent | next [-]

    > Quite likely this year we will have federal law that will allow selling cars with fully unsupervised self-driving, in which case the insurance/liability will obviously land on the maker of the system, not person present in the car.
You raise an important point here. Is it economically feasible for system makers to bear the responsibility of self-driving car accidents? It seems impossible, unless the cars are much more expensive to cover the potential future costs. I'm very curious how Waymo insures their cars today. I assume they have a bespoke insurance contract negotiated with a major insurer. Also, do we know the initial cost of each Waymo car (to say nothing of ongoing costs from compute/mapping/etc.)? It must be very high (2x?) given all of the special navigation equipment that is added to each car.
paulryanrogers 6 hours ago | parent | prev | next [-]

Tacking "Supervised" on the end of "Full Self Driving" is just contradictory. Perhaps if it was "Partial Self Driving" then it wouldn't be so confusing.

pests 5 hours ago | parent [-]

Its only to differentiate it from their "Unsupervised FSD" which is what they call it now.

arijun 13 hours ago | parent | prev | next [-]

I imagine insurance would be split in two in that case. Carmakers would not want to be liable for e.g. someone striking you in a hit-and-run.

smallnix 12 hours ago | parent [-]

If the car that did a hit-and-run was operated autonomously the insurance of the maker of that car should pay. Otherwise it's a human and the situation falls into the bucket of what we already have today.

So yes, carmakers would pay in a hit-and-run.

JumpCrisscross 11 hours ago | parent [-]

> If the car that did a hit-and-run was operated autonomously the insurance of the maker of that car should pay

Why? That's not their fault. If a car hits and runs my uninsured bicycle, the manufacturer isn't liable. (My personal umbrella or other insurance, on the other hand, may cover it.)

graeme 9 hours ago | parent [-]

They're describing a situation of liability, not mere damage. If yor bicycle is hit you didn't do anything wrong.

If you run into someone on your bike and are at fault then you generally would be liable.

They're talking about the hypothetical where you're on your bike, which was sold as an autobomous bike and the bike manufacturer's software fully drives the bike, and it runs into someone and is at fault.

AlotOfReading 14 hours ago | parent | prev | next [-]

You can sell autonomous vehicles to consumers all day long. There's no US federal law prohibiting that, as long as they're compliant with FMVSS as all consumer vehicles are required to be.

rubyfan 15 hours ago | parent | prev | next [-]

Waymo is also a livery service which you normally aren’t liable for as a passenger of taxi or limousine unless you have deep pockets. /IANAL

kolbe 14 hours ago | parent | prev | next [-]

> Quite likely this year we will have federal law that will allow selling cars with fully unsupervised self-driving, in which case the insurance/liability will obviously land on the maker of the system, not person present in the car.

This is news to me. This context seems important to understanding Tesla's decision to stop selling FSD. If they're on the hook for insurance, then they will need to dynamically adjust what they charge to reflect insurance costs.

jasoncartwright 15 hours ago | parent | prev | next [-]

I see. So not Tesla's product they are using to sell insurance around isn't "Full Self-Driving" or "Autonomous" like the page says.

FeloniousHam 15 hours ago | parent [-]

My current FSD usage is 90% over ~2000 miles (since v14.x). Besides driving everywhere, everyday with FSD, I have driven 4 hours garage to hotel valet without intervention. It is absolutely "Full Self-Driving" and "Autonomous".

FSD isn't perfect, but it is everyday amazing and useful.

JumpCrisscross 11 hours ago | parent | next [-]

> My current FSD usage is 90% over ~2000 miles

I'd guess my Subaru's lane-keeping utilisation is in the same ballpark. (By miles, not minutes. And yes, I'm safer when it and I are watching the road than when I'm watching the road alone.)

olyjohn 8 hours ago | parent [-]

My favorite feature of Subaru's system is when you change lanes, and it stays locked onto the car in the slower lane and slams on the brakes. People behind you love that.

wat10000 13 hours ago | parent | prev | next [-]

If it was full self driving, wouldn't your usage be 100%?

pests 5 hours ago | parent | next [-]

Sometimes a car is fun to drive.

12 hours ago | parent | prev [-]
[deleted]
jasoncartwright 15 hours ago | parent | prev [-]

Yet still on relying you to cover it with your insurance. Again, clearly not autonomous.

AlotOfReading 14 hours ago | parent | next [-]

Liability is a separate matter from autonomy. I assume you'd consider yourself autonomous, yet it's your employer's insurance that will be liable if you have an accident while driving a company vehicle.

If the company required a representative to sit in the car with you and participate in the driving (e.g. by monitoring and taking over before an accident), then there's a case to be made that you're not fully autonomous.

buran77 12 hours ago | parent [-]

> it's your employer's insurance that will be liable if you have an accident while driving a company vehicle

I think you're mixing some concepts.

There's car insurance paid by the owner of the car, for the car. There's workplace accident insurance, paid by the employer for the employee. The liability isn't assigned by default, but by determining who's responsible.

The driver is always legally responsible for accidents caused by their negligence. If you play with your phone behind the wheel and kill someone, even while working and driving a company car, the company's insurance might pay for the damage but you go to prison. The company will recover the money from you. Their work accident insurance will pay nothing.

The test you can run in your head: will you get arrested if you fall asleep at the wheel and crash? If yes, then it's not autonomous or self driving. It just has driver assistance. It's not that the car can't drive itself at all, just that it doesn't meet the bar for the entire legal concept of "driver/driving".

"Almost" self driving is like jumping over a canyon and almost making it to the other side. Good effort, bad outcome.

dzhiurgis 14 hours ago | parent | prev [-]

[flagged]

zen928 13 hours ago | parent [-]

Disagree. I appreciate their viewpoint tethering corporate claims to reality by illustrating Tesla is obfuscating the classification of their machines to be autonomous, when they actually aren't. Their comments in other thread chains proved to be fruitful when lacking agitators looking to dismiss critique by citing website rules, like the post adding additional detail to how Tesla muddles legal claims by cooking up cherry-picked evidence that work against the driver despite being the insurer.

2III7 14 hours ago | parent | prev [-]

Without LIDAR and/or additional sensors, Tesla will never be able to provide "real" FSD, no matter how wonderful their software controlling the car is.

Also, self driving is a feature of a vehicle someone owns, I don't understand how that should exempt anyone from insuring their property.

Waymo and others are providing a taxi service where the driver is not a human. You don't pay insurance when you ride Uber or Bolt or any other regular taxi service.

Marsymars 14 hours ago | parent | next [-]

> Also, self driving is a feature of a vehicle someone owns, I don't understand how that should exempt anyone from insuring their property.

Well practically speaking, there’s nothing stopping anyone from voluntarily assuming liability for arbitrary things. If Tesla assumes the liability for my car, then even if I still require my “own” insurance for legal purposes, the marginal cost of covering the remaining risk is going to be close to zero.

arijun 13 hours ago | parent | prev [-]

Never say never—it’s not physically impossible. But yes, as it stands, it seems that Tesla will not be self driving any time soon (if ever).

kjksf 12 hours ago | parent [-]

They literally just (in the last few days) started unsupervised robotaxis in Austin.

They are as self-driving as a car can be.

This is different than the one where they had a human supervisor in passenger seat (which they still do elsewhere).

And different than the one where they didn't have human supervisor but did have a follow car.

Now they have a few robotaxis that are self driving.

thg 2 hours ago | parent | next [-]

It was just another marketing stunt to pump the stock price before their terrible earnings report. One "unsupervised" unit, with the supervisor in a follow car, that nobody could actually get and drive around in.

https://electrek.co/2026/01/28/teslas-unsupervised-robotaxis...

iknowstuff an hour ago | parent [-]

Stop reading electrek.

https://x.com/davidmoss/status/2016939137031381487?s=46

amanaplanacanal 10 hours ago | parent | prev [-]

Have you actually ridden in one? It's unclear whether this is a real thing or not.

zugi 8 hours ago | parent | prev | next [-]

If your minor child breaks something, or your pet bites someone, you are liable.

This analogy may be more apt than Tesla would like to admit, but from a liability perspective it makes sense.

You could in turn try to sue Tesla for defective FSD, but the now-clearly-advertised "(supervised)" caveat, plus the lengthy agreement you clicked through, plus lots of lawyers, makes you unlikely to win.

paulryanrogers 6 hours ago | parent [-]

Can a third party reprogram my dog or child at any moment? Or even take over and control them?

JumpCrisscross 11 hours ago | parent | prev | next [-]

> Surely if it's Tesla making the decisions, they need the insurance?

Why surely? Turning on cruise control doesn't absolve motorists of their insurance requirement.

And the premise is false. While Tesla does "not maintain as much insurance coverage as many other companies do," there are "policies that [they] do have" [1]. (What it insures is a separate question.)

[1] https://www.sec.gov/ix?doc=/Archives/edgar/data/0001318605/0...

forgetfreeman 10 hours ago | parent [-]

Cruise control is hardly relevant to a discussion of liability for autonomous vehicle operation.

fragmede 10 hours ago | parent [-]

In the context of ultramodern cruise control (eg comma.ai), which has a radar to track the distance to the car (if any) in front of you, and cameras so the car can wind left or right and track the freeway, I think it does.

sroussey 10 hours ago | parent [-]

Not unless they are marketing it as “autopilot” or some such that a random consumer would reasonably assume meant autopilot.

And I’d include “AI driver” as an example.

olyjohn 8 hours ago | parent [-]

A random consumer doesn't actually understand what Autopilot means. Most people don't have pilot's licenses. And cars don't fly. Did you not see all the debacles around it when it first came out?

davidhunter 16 hours ago | parent | prev | next [-]

Seems like the role of the human operator in the age of AI is to be the entity they can throw in jail if the machine fails (e.g. driver, pilot)

taneq 11 hours ago | parent [-]

I’ve said for years that pragmatically, our definition of a “person” is an entity that can accept liability and take blame.

wcfrobert 10 hours ago | parent | next [-]

LLCs can't go to jail though

amitav1 10 hours ago | parent [-]

Because LLCs aren't people

sroussey 10 hours ago | parent | prev [-]

Not to be confused with “human” thanks to SCOTUS.

stubish 9 hours ago | parent | prev | next [-]

Risk gets passed along until someone accepts it, usually an insurance company or the operator. If the risk was accepted and paid for by Tesla, then the cost would simply be passed down to consumers. All consumers, including those that want to accept the risk themselves. In particular, if you have a fleet of cars it can be cheaper to accept the risk and only pay for mandatory insurance, because not all of your cars are going to crash at the same time, and even if they did, not all in the worst way possible. This is how insurance works, by amortizing lots of risk to make it highly improbable to make a loss in the long run.

seanmcdirmid 10 hours ago | parent | prev | next [-]

I think there is an even bigger insurance problem to worry about: if autonomous vehicles become common and are a lot safer than manual driven vehicles, insurance rates for human driven cars could wind up exploding as the risk pool becomes much smaller and statistically riskier. We could go from paying $200/month to $2000/month if robo taxis start dominating cities.

AnthonyMouse 9 hours ago | parent | next [-]

> if autonomous vehicles become common and are a lot safer than manual driven vehicles, insurance rates for human driven cars could wind up exploding as the risk pool becomes much smaller and statistically riskier.

The assumption there is that the remaining human drivers would be the higher risk ones, but why would that be the case?

One of the primary movers of high risk driving is that someone goes to the bar, has too many drinks, then needs both themselves and their car to get home. Autonomous vehicles can obviously improve this by getting them home in their car without them driving it, but if they do, the risk profile of the remaining human drivers improves. At worst they're less likely to be hit by a drunk driver, at best the drunk drivers are the early adopters of autonomous vehicles and opt themselves out of the human drivers pool.

seanmcdirmid 9 hours ago | parent [-]

Drunk driving isn't the primary mover of high risk driving. Rather you have:

1. People who can't afford self driving cars (now the insurance industry has a good proxy for income that they couldn't tap into before)

2. Enthusiasts who like driving their cars (cruisers, racers, Helcat revving, people who like doing donuts, etc...)

3. Older people who don't trust technology.

None of those are good risk pools to be in. Also, if self driving cars go mainstream, they are bound to include the safest drivers overnight, so whatever accidents/crashes happen afterwards are covered by a much smaller and "active" risk pool. Oh, and those self driving cars are expensive:

* If you hit one and are at fault, you might pay out 1-200k, most states only require 25k-50k of coverage...so you need more coverage or expect to pay more for incident.

* Self driving cars have a lot of sensors/recorders. While this could work to your advantage (proving that you aren't at fault), it often isn't (they have evidence that you were at fault). Whereas before fault might have been much more hazy (both at fault, or both no fault).

The biggest factor comes if self driving cars really are much safer than human drivers. They will basically disappear from the insurance market, or somehow be covered by product liability instead of insurance...and the remaining drivers will be in a pool of the remaining accidents that they will have to cover on their own.

AnthonyMouse 9 hours ago | parent [-]

> Drunk driving isn't the primary mover of high risk driving.

It kind of is. They're responsible for something like 30% of traffic fatalities despite being a far smaller percentage of drivers.

> People who can't afford self driving cars (now the insurance industry has a good proxy for income that they couldn't tap into before)

https://pubmed.ncbi.nlm.nih.gov/30172108/

But also, wouldn't they already have this by using the vehicle model and year?

> Enthusiasts who like driving their cars (cruisers, racers, Helcat revving, people who like doing donuts, etc...)

Again something that seems like it would already be accounted for by vehicle model.

> Older people who don't trust technology.

How sure are we that the people who don't trust technology are older? And again, the insurance company already knows your age.

> Also, if self driving cars go mainstream, they are bound to include the safest drivers overnight

Are they? They're more likely to include the people who spend the most time in cars, which is another higher risk pool, because it allows those people to spend the time on a phone/laptop instead of driving the car, which is worth more to people the more time they spend doing it and so justifies the cost of a newer vehicle more easily.

> Oh, and those self driving cars are expensive

Isn't that more of a problem for the self-driving pool? Also, isn't most of the cost that the sensors aren't as common and they'd end up costing less as a result of volume production anyway?

> Self driving cars have a lot of sensors/recorders. While this could work to your advantage (proving that you aren't at fault), it often isn't (they have evidence that you were at fault). Whereas before fault might have been much more hazy (both at fault, or both no fault).

Which is only a problem for the worse drivers who are actually at fault, which makes them more likely to move into the self-driving car pool.

> The biggest factor comes if self driving cars really are much safer than human drivers.

The biggest factor is which drivers switch to self-driving cars. If half of human drivers switched to self-driving cars but they were chosen completely at random then the insurance rates for the remaining drivers would be essentially unaffected. How safe they are is only relevant insofar as it affects your chances of getting into a collision with another vehicle, and if they're safer then it would make that chance go down to have more of them on the road.

seanmcdirmid 8 hours ago | parent [-]

Only .61% of car crashes involve fatalities, so that’s like .2% of car crashes you are referring to. Probably more due to alcohol, but we don’t know the ratio of accidents that involve alcohol, which would be more telling.

> How sure are we that the people who don't trust technology are older? And again, the insurance company already knows your age

Boomers are already the primary anti-EV demographic, with the complaint that real cars have engines. It doesn’t matter if they know your age of state laws keep them from acting on it.

> that more of a problem for the self-driving pool? Also, isn't most of the cost that the sensors aren't as common and they'd end up costing less as a result of volume production anyway?

I think you misunderstood me: If you get into an accident and are found at fault, you are responsible for damage to the other car. Now, if it’s a clunker Toyota, that will be a few thousand dollars, if it’s a roll Royce, it’s a few hundred thousand dollars. The reason insurances are increasing lately is that the average car on the road is more expensive than it was ten years ago, so insurance companies are paying out more. If most cars are $250k Waymo cars, and you hit one…and you are at fault, ouch. And we will know if it is your fault or not since the Waymo is constantly recording.

> If half of human drivers switched to self-driving cars but they were chosen completely at random then the insurance rates for the remaining drivers would be essentially unaffected.

That’s not how the math works out (smaller risk pools are more expensive per person period). And it won’t be people switching at random to self driving cars (the ones not switching will be the ones that are more likely to have accidents).

AnthonyMouse 3 hours ago | parent [-]

> Only .61% of car crashes involve fatalities, so that’s like .2% of car crashes you are referring to. Probably more due to alcohol, but we don’t know the ratio of accidents that involve alcohol, which would be more telling.

Fatalities get more thoroughly investigated so we have better numbers on them, but if you had to guess whether the people who get behind the wheel drunk were similarly disproportionately likely to bang up their cars in a non-fatal way, what would your guess be?

> Boomers are already the primary anti-EV demographic, with the complaint that real cars have engines.

EVs and self-driving are two different things. Fox News tells boomers that EVs are bad because Republicans have the oil companies as a constituency.

> It doesn’t matter if they know your age of state laws keep them from acting on it.

The only states that do that are Hawaii and Massachusetts.[1]

[1] https://www.cnbc.com/select/best-car-insurance-seniors/

> If most cars are $250k Waymo cars, and you hit one…and you are at fault, ouch. And we will know if it is your fault or not since the Waymo is constantly recording.

If X% of cars are Waymos and you hit another car in your normally priced car and you're at fault, there is an X% chance it will be expensive. If the Waymo hits another car and it's at fault, there is a 100% chance it will be expensive because it will damage itself, and an additional X% chance that it will be very expensive because both cars are.

And again, that's assuming the price stays as high as it is when the production volume increases. A $250,000 car can't become the majority of cars because that percentage of people can't afford that.

> That’s not how the math works out (smaller risk pools are more expensive per person period).

Smaller risk pools don't have higher risk, they have higher volatility, and then if they're too small insurers have to charge a volatility premium. But the auto insurance market is very large and for it to get to the size that it would have volatility issues it would have to be a consequence rather than a cause of the large majority of people switching to self-driving cars.

> And it won’t be people switching at random to self driving cars (the ones not switching will be the ones that are more likely to have accidents).

You keep saying that but it's still not obvious that it's what would happen, and in any event the ones more likely to have accidents are already the ones paying higher insurance premiums -- which is precisely a reason they would have the incentive to be the first to switch to self-driving cars.

mavhc 10 hours ago | parent | prev [-]

The fact you think $200 per month is sane is amusing to people in other countries

seanmcdirmid 10 hours ago | parent | next [-]

Haha, yes, today already sucks badly in many US markets. Imagine what will happen when the only people driving cars manually are "enthusiasts".

theodric 10 hours ago | parent | prev | next [-]

Hell, I was paying €180/yr for my New Beetle a decade ago...

wavesquid 10 hours ago | parent | prev [-]

Is that low or high?

djoldman 12 hours ago | parent | prev | next [-]

That's probably the future; Mercedes currently does do this in limited form:

https://www.roadandtrack.com/news/a39481699/what-happens-if-...

jdminhbg 6 hours ago | parent [-]

Not "currently," "used to": https://www.theverge.com/transportation/860935/mercedes-driv...

It was way too limited to be useful to anyone.

ponector 9 hours ago | parent | prev | next [-]

Why ship owner is paying for the insurance while it's a captain making all decisions?

jgbuddy 17 hours ago | parent | prev | next [-]

Because the operator is liable? Tesla as a company isn't driving the car, it's a ML model running on something like HW4 on bare metal in the car itself. Would that make the silicon die legally liable?

jasoncartwright 16 hours ago | parent | next [-]

Sounds like it's neither self-driving, nor autonomous, if I'm on the hook if it goes wrong.

scottbez1 16 hours ago | parent | next [-]

Yeah, Tesla gets to blame the “driver”, and has a history of releasing partial and carefully curated subsets of data from crashes to try to shift as much blame onto the driver as possible.

And the system is designed to set up drivers for failure.

An HCI challenge with mostly autonomous systems is that operators lose their awareness of the system, and when things go wrong you can easily get worse outcomes than if the system was fully manual with an engaged operator.

This is a well known challenge in the nuclear energy sector and airline industry (Air France 447) - how do you keep operators fully engaged even though they almost never need to intervene, because otherwise they’re likely to be missing critical context and make wrong decisions. These days you could probably argue the same is true of software engineers reviewing LLM code that’s often - but not always - correct.

redanddead 13 hours ago | parent [-]

> has a history of releasing partial and carefully curated subsets of data from crashes to try to shift as much blame onto the driver as possible

Really? Thats crazy.

7 hours ago | parent | prev | next [-]
[deleted]
thelastgallon 16 hours ago | parent | prev | next [-]

Its neither self-driving, nor autonomous, eventually not even a car! (as Tesla slowly exits the car business). It will be 'insurance' on Speculation as a service, as Tesla skyrockets to $20T market cap. Tesla will successfully transition from a small revenue to pre-revenue company: https://www.youtube.com/watch?v=SYJdKW-UnFQ

The last few years of Tesla 'growth' show how this transition is unfolding. S and X production is shutdown, just a few more models to shutdown.

rubyfan 15 hours ago | parent | next [-]

I wonder if they will try to sell off the car business once they can hype up something else. It seems odd to just let the car business die.

redanddead 13 hours ago | parent | prev [-]

Wild prediction, would love to hear the rest of it

iwontberude 11 hours ago | parent | prev [-]

Especially since they can push regressions over the air and you could be lulled into a sense of safety and robustness that isn’t there and bam you pay the costs of the regressions, not Tesla.

throw20251220 16 hours ago | parent | prev | next [-]

Who’s the “operator” of an “autonomous” car? If I sit in it and it drives me around, how am I an “operator”?

renewiltord 16 hours ago | parent [-]

If you get on a horse and let go of the reins you are also considered the operator of the horse. Such are the definitions in our society.

kyleee 11 hours ago | parent [-]

Great analogy, lol

close04 14 hours ago | parent | prev [-]

The point is if the liability is always exclusively with the human driver then any system in that car is at best a "driver assist". Claims that "it drives itself" or "it's autonomous" are just varying degrees of lying. I call it a partial lie rather than a partial truth because the result more often than not is that the customer is tricked into thinking the system is more capable than it is, and because that outcome is more dangerous than the opposite.

Any car has varying degrees of autonomy, even the ones with no assists (it will safely self-drive you all the way to the accident site, as they say). But the car is either driven by the human with the system's help, or is driven by the system with or without the human's help.

A car can't have 2 drivers. The only real one is the one the law holds responsible.

AnthonyMouse 10 hours ago | parent | prev | next [-]

> If it autonomous or self-driving then why is the person in the car paying for the insurance? Surely if it's Tesla making the decisions, they need the insurance?

Suppose ACME Corporation produces millions of self-driving cars and then goes out of business because the CEO was embezzling. They no longer exist. But the cars do. They work fine. Who insures them? The person who wants to keep operating them.

Which is the same as it is now. It's your car so you pay to insure it.

I mean think about it. If you buy an autonomous car, would the manufacturer have to keep paying to insure it forever as long as you can keep it on the road? The only real options for making the manufacturer carry the insurance are that the answer is no and then they turn off your car after e.g. 10 years, which is quite objectionable, or that the answer is "yes" but then you have to pay a "subscription fee" to the manufacturer which is really the insurance premium, which is also quite objectionable because then you're then locked into the OEM instead of having a competitive insurance market.

charcircuit 11 hours ago | parent | prev | next [-]

Not all insurance claims are based off of the choices of the driver.

throw20251220 16 hours ago | parent | prev | next [-]

It’s because you bought it. Don’t buy it if you don’t want to insure.

SoftTalker 16 hours ago | parent | next [-]

Yep, you bought it, you own it, you choose to operate it on the public roads. Therefore your liability.

9rx 16 hours ago | parent [-]

If you bought and owned it, you could sell it to another auto manufacturer for some pretty serious amounts of money.

In reality, you acquired a license to use it. Your liability should only go as far as you have agreed to identify the licenser.

recursive 12 hours ago | parent | next [-]

You can actually do that. Except that they could just buy one themselves.

Companies exist that buy cars just to tear them down and publish reports on what they find.

9 hours ago | parent | next [-]
[deleted]
9rx 9 hours ago | parent | prev [-]

> Companies exist that buy cars just to tear them down and publish reports on what they find.

What does it mean to tear down software, exactly? Are you thinking of something like decompilation?

You can do that, but you're probably not going to learn all that much, and you still can't use it in any meaningful sense as you never bought it in the first place. You only licensed use of it as a consumer (and now that it is subscription-only, maybe not even that). If you have to rebuild the whole thing yourself anyway, what have you really gained? Its not exactly a secret how the technology works, only costly to build.

> Except that they could just buy one themselves.

That is unlikely, unless you mean buying Tesla outright? Getting a license to use it as a manufacturer is much more realistic, but still a license.

recursive 7 hours ago | parent [-]

Check out Munro and Associates. I'm not talking about software. The whole car.

9rx 6 hours ago | parent [-]

For what reason?

In case you have forgotten, the discussion is about self-driving technology, and specifically Tesla's at that. The original questioner asked why he is liable when it is Tesla's property that is making the decisions. Of course, the most direct answer is because Tesla disclaims any liability in the license agreement you must agree to in order to use said property.

Which has nothing to do with an independent consulting firm or "the whole car" as far as I can see. The connection you are trying to establish is unclear. Perhaps you pressed the wrong 'reply' button by mistake?

recursive 16 minutes ago | parent [-]

I started responding to this. I interpreted it to be referring to the whole car.

> Yep, you bought it, you own it, you choose to operate it on the public roads. Therefore your liability.

16 hours ago | parent | prev [-]
[deleted]
Rebelgecko 16 hours ago | parent | prev [-]

I don't think Tesla lets you buy FSD

scottyah 14 hours ago | parent [-]

They do, until Feb 14th.

Rebelgecko 11 hours ago | parent [-]

Even now I think it's a revocable license

abtinf 13 hours ago | parent | prev | next [-]

You insure the property, not the person.

redanddead 11 hours ago | parent [-]

well it's the risk, the combination ..

it's why young drivers pay more for insurance

jimt1234 16 hours ago | parent | prev | next [-]

Not an expert here, but I recall reading that certain European countries (Spain???) allow liability to be put on the autonomous driving system, not the person in the car. Does anyone know more about this?

bluGill 13 hours ago | parent [-]

That is the case everywhere. It is common when buying a product for the contract to include who has liability for various things. The price often changes by a lot depending on who has liability.

Cars are traditionally sold as the customer has liability. Nothing stops a car maker (or even an individual dealer) from selling cars today taking all the insurance liability in any country I know of - they don't for what I hope are obvious reasons (bad drivers will be sure to buy those cars since it is a better deal for them an in turn a worse deal for good drivers), but they could.

Self driving is currently sold as customers has liability because that is how it has always been done. I doubt it will change, but it is only because I doubt there will ever be enough advantage as to be worth it for someone else to take on the liability - but I could be wrong.

ck2 12 hours ago | parent | prev | next [-]

The coder and sensor manufacturers need the insurance for wrongful death lawsuits

and Musk for removing lidar so it keeps jumping across high speed traffic at shadows because the visual cameras can't see true depth

99% of the people on this website are coders and know how even one small typo can cause random fails, yet you trust them to make you an alpha/beta tester at high speed?

loeg 13 hours ago | parent | prev [-]

It isn't fully autonomous yet. For any future system sold as level 5 (or level 4?), I agree with your contention -- the manufacturer of the level 5 autonomous system is the one who bears primary liability and therefore should insure. "FSD" isn't even level 3.

(Though, there is still an element of owner/operator maintenance for level 4/5 vehicles -- e.g., if the owner fails to replace tires below 4/32", continues to operate the vehicle, and it causes an injury, that is partially the owner/operator's fault.)

Night_Thastus 12 hours ago | parent [-]

Wouldn't that requirement completely kill any chance of a L5 system being profitable? If company X is making tons of self-driving cars, and now has to pay insurance for every single one, that's a mountain of cash. They'd go broke immediately.

I realize it would suck to be blamed for something the car did when you weren't driving it, but I'm not sure how else it could be financially feasible.

loeg 12 hours ago | parent | next [-]

No? Insurance costs would be passed through to consumers in the form of up-front purchase price. And probably the cost to insure L5 systems for liability will be very low. If it isn't low, the autonomous system isn't very safe.

AlotOfReading 12 hours ago | parent | prev [-]

The way it works in states like California currently is that the permit holder has to post an insurance bond that accidents and judgements are taken out against. It's a fixed overhead.