• Blackmist@feddit.uk
    link
    fedilink
    English
    arrow-up
    44
    arrow-down
    14
    ·
    8 months ago

    The question isn’t “are they safer than the average human driver?”

    The question is “who goes to prison when that self driving car has an oopsie, veers across three lanes of traffic and wipes out a family of four?”

    Because if the answer is “nobody”, they shouldn’t be on the road. There’s zero accountability, and because it’s all wibbly-wobbly AI bullshit, there’s no way to prove that the issues are actually fixed.

      • Blackmist@feddit.uk
        link
        fedilink
        English
        arrow-up
        20
        arrow-down
        11
        ·
        8 months ago

        Accountability is important. If a human driver is dangerous, they get taken off the roads and/or sent to jail. If a self driving car kills somebody, it’s just “oops, oh well, these things happen, but shareholder make a lot of money so never mind”.

        I do not want “these things happen” on my headstone.

        • Tja@programming.dev
          link
          fedilink
          English
          arrow-up
          9
          arrow-down
          4
          ·
          8 months ago

          So you would prefer to have higher chances of dying, just to write “Joe Smith did it” on it?

        • ipkpjersi
          link
          fedilink
          English
          arrow-up
          8
          arrow-down
          5
          ·
          edit-2
          8 months ago

          But if a human driver is dangerous, and gets put in jail or get taken off the roads, there are likely already more dangerous human drivers taking their place. Not to mention, genuine accidents, even horrific ones, do happen with human drivers. If the rate of accidents and rate of fatal accidents with self-driving vehicles is way down versus human drivers, you are actually risking your life more by trusting in human drivers and taking way more risks that way. Having someone be accountable for your death doesn’t matter if you’ve already died because of them.

          Is it any better if you have “Killed by Bill Johnson’s SUV” on your headstone?

    • dream_weasel@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      3
      ·
      8 months ago

      The answer is the person behind the wheel.

      Tesla makes it very clear to the driver they you still have to pay attention and be ready to take over any time. Full self driving engages the in cabin nanny cam to enforce that you pay attention, above and beyond the frequent reminders to apply turning force to the steering wheel.

      Now, once Tesla goes Mercedes and says you don’t have to pay attention, it’s gonna be the company that should step in. I know that’s a big old SHOULD, but right now that’s not the situation anyway.

      • exanime@lemmy.today
        link
        fedilink
        English
        arrow-up
        6
        ·
        8 months ago

        Now, once Tesla goes Mercedes and says you don’t have to pay attention, it’s gonna be the company that should step in

        That doesn’t give me warm and fuzzies either… Imagine a poor dude having to fight Mercedes or Testla because he was crippled by a sleeping driver and bad AI… Not even counting the lobbying that would certainly happen to reduce and then eliminate their liability

        • dream_weasel@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          6
          ·
          8 months ago

          There will be legal battles for sure. I don’t know how you can argue for anything besides the manufacturer taking responsibility. I don’t know how that doesn’t end up with auto pilot fatalities treated as a class where there’s a lookup table of payouts though. This is the intersection of liability and money/power, so it’s functionally uncharted territory at least in the US.

    • ipkpjersi
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      2
      ·
      edit-2
      8 months ago

      The question isn’t “are they safer than the average human driver?”

      How is that not the question? That absolutely is the question. Just because someone is accountable for your death doesn’t mean you aren’t already dead, it doesn’t bring you back to life. If a human driver is actively dangerous and get taken off the road or put in jail, there are very likely already plenty more taking that human drivers place. Plus genuine accidents, even horrific ones, do happen with human drivers. If the death rate for self-driving vehicles is really that much lower, you are risking your life that much more by trusting in human drivers.

      • ShepherdPie@midwest.social
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        2
        ·
        8 months ago

        Yeah that person’s take seems a little unhinged as throwing people in prison after a car accident only happens if they’re intoxicated or driving recklessly. These systems don’t have to be perfect to save lives. They just have to be better than the average driver.

        • Tja@programming.dev
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          8 months ago

          Hell, let’s put the threshold at “better than 99% of drivers”, because every driver I know thinks they are better than average.

      • sugar_in_your_tea@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        8 months ago

        Exactly.

        We should solve the accountability problem, but the metric should be lives and accidents. If the self-driving system proves it causes fewer accidents and kills fewer people, it should be preferred. Full stop.

        Throwing someone in jail may be cathartic, but the goal is fewer issues on the road, not more people in jail.

    • Maddier1993@programming.dev
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      2
      ·
      8 months ago

      I don’t agree with your argument.

      Making a human go to prison for wiping out a family of 4 isn’t going to bring back the family of 4. So you’re just using deterrence to hopefully make drivers more cautious.

      Yet, year after year… humans cause more deaths by negligence than tools can cause by failing.

      The question is definitely “How much safer are they compared to human drivers”

      It’s also much easier to prove that the system has those issues fixed compared to training a human hoping that their critical faculties are intact. Rigorous Software testing and mechanical testing are within legislative reach and can be made strict requirements.

    • kava@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      7
      ·
      8 months ago

      Because if the answer is “nobody”, they shouldn’t be on the road

      Do you understand how absurd this is? Let’s say AI driving results in 50% less deaths. That’s 20,000 people every year that isn’t going to die.

      And you reject that for what? Accountability? You said in another comment that you don’t want “shit happens sometimes” on your headstone.

      You do realize that’s exactly what’s going on the headstones of those 40,000 people that die annually right now? Car accidents happen. We all know they happen and we accept them as a necessary evil. “Shit happens”

      By not changing it, ironically, you’re advocating for exactly what you claim you’re against.

      • exanime@lemmy.today
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        4
        ·
        8 months ago

        Hmmm I get you point but you seem to be taken the cavalier position of one who’d never be affected.

        Let’s proposed this alternative scenario: AI is 50% safer and would reduce death from 40k to 20k a year if adopted. However, the 20k left will include your family and, unfortunately , there is no accountability therefore, nobody will pay to help raise your orphan nephew or help grandma now that your grandpa died ran over by a Tesla… Would you approve AI driving going forward?

        • kava@lemmy.world
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          1
          ·
          8 months ago

          A) you do realize cars have insurance and when someone hits you, that insurance pays out the damages, right? That is how the current system works, AI driver or not.

          Accidents happen. Humans make mistakes and kill people and are not held criminally liable. It happens.

          If some guy killed your nephew and made him an orphan and the justice system determined he was not negligent - then your nephew would still be an orphan and would get a payout by the insurance company.

          Exact same thing that happens in the case of an AI driven car hitting someone

          B) if I had a button to save 100k people but it killed my mother, I wouldn’t do it. What is your point?

          Using your logic, if your entire family was in the 20,000 who would be saved - you would prefer them dead? You’d rather them dead with “accountability” rather than alive?

            • kava@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              ·
              8 months ago

              Your thought experiment doesn’t work. I wouldn’t accept any position where my family members die and beyond that, it’s immaterial to the scope of discussion.

              Let’s examine various different scenarios under which someone dies in a car accident.

              1. human driver was negligent and causes a fatal car accident.

              Human gets criminal charges. Insurance pays out depending on policy.

              1. human driver was not negligent and causes a fatal car accident.

              Human does not get criminal charged. Insurance pays out depending on policy

              1. AI driver causes a fatal accident.

              Nobody gets criminal charges. Insurance pays out depending on policy.


              You claim that you would rather have 20,000 people die every year because of “accountability”.

              Tell me, what is the functional difference for a family member of a fatal car accident victim in those 3 above scenarios? The only difference is under 1) there would be someone receiving criminal charges.

              They recieve the same amount of insurance money. 2) already happens right now. You don’t mention that in the lack of accountability.

              You claim that being able to pin some accidents (remember, some qualify under 2) on an individual is worth 20,000 lives a year.

              Anybody who has ever lost someone in a car accident would rather have their family member back instead.

              • exanime@lemmy.today
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                edit-2
                8 months ago

                Your thought experiment doesn’t work

                The point of a thought experiment is to think about that proposition, not to replace with whatever you think makes sense

                1. AI driver causes a fatal accident.

                Nobody gets criminal charges. Insurance pays out depending on policy.

                Now here is my concern… You are reducing a human life to a dollar amount just like Ford did with the Pinto. If Mercedes (who is apparently liable), decides they are making more money selling their cars than paying out to people injured or killed by their cars, what’s left to force them to recall/change/fix their algorithm?

                PS: I also never claimed I rather have 20000 more people die for accountability… So, I guess you have to argue that with the part of your brain that made it up

                • kava@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  8 months ago

                  PS: I also never claimed I rather have 20000 more people die for accountability…

                  You said it’s not a question of how much safer it is. You said it’s a question of accountability. So even if it were 50% safer, you claimed it was wrong.

                  And here’s the thing man, I understand where you’re coming from ij that you shouldn’t reduce a life to numbers. But how does AI driving fundamentally change the current situation?

                  Car companies already do this. They calculate whether or not fixing a safety problem will cost more or less than the lawsuits from all the dead people. There’s a famous documented case of this. Maybe it’s the Ford / Pinto thing you are referencing.

                  If you think of AI driving as a safety feature - like seatbelts - would you support it? I don’t know what the actual statistics are, but presumably it’s only going to get better over time.

        • sugar_in_your_tea@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          1
          ·
          8 months ago

          Yes, unless you mean I need to literally sacrifice my family. But if my family was randomly part of the 20k, I’d defend self-driving cars if they are proven to be safer.

          I’m very much a statistics-based person, so I’ll defend the statistically better option. In fact, me being part of that 20k gives me a larger than usual platform to discuss it.

          • exanime@lemmy.today
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            3
            ·
            8 months ago

            No, I do mean literally your family. Not because I’m trying to be mean to you, I’m just trying to highlight you’d agree with a contract when you think the price does not apply to you… But in reality the price will apply to someone, whether they agree with the contract and enjoy the benefits or not

            It’s the exact same situation with real life with the plane manufacturers. They lobby the government to allow recalls not to be done immediately but instead on the regular maintenance of the planes. This is to save money but it literally means that some planes are put there with known defects that will not be addressed for months (or years, depending on the maintenance needed)

            Literally, people who’d never have a loved one in one of those flights decided that was acceptable to save money. They agreed, it’s ok to put your life at risk, statistically, because they want more money

            • Tja@programming.dev
              link
              fedilink
              English
              arrow-up
              4
              ·
              8 months ago

              If there are 20k deaths vs 40k, my family is literally twice as safe on the road, why wouldn’t I take that deal?

              • exanime@lemmy.today
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                3
                ·
                8 months ago

                Read the proposition… It’s a thought experiment what we were discussing

                • Tja@programming.dev
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  1
                  ·
                  8 months ago

                  The proposition is stupid. If you told me that ALL future accidents will be prevented if I agree to kill my family, I would still not do it, that’s just a bad faith trolley problem. Let’s alone just recuding it by half.

                  I reduced it to a more realistic experiment, where my family migth be killed, with the same probability as any other.

                  • exanime@lemmy.today
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    arrow-down
                    2
                    ·
                    8 months ago

                    The proposition is stupid.

                    Oh the depth of reasoning in social media

                    If you told me that ALL future accidents will be prevented if I agree to kill my family, I would still not do it

                    That is exactly the point… Anyone would be 100% happy taking any proposition as long as they don’t have to pay the cost. I was just trying to highlight that

                    In this case, it was all about liability… We have not even come close to prove the current driverless tech is actually better than people’s skills… We all know that automated driving should be safer but we have no clue if we are even taking the right steps.to get there

            • sugar_in_your_tea@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              4
              arrow-down
              1
              ·
              edit-2
              8 months ago

              Then it’s not a fair question. You’re not comparing 40k vs 20k, you’re comparing 40k vs literally my family dying (like the hypothetical train diversion thing), that’s fear mongering and not a valid argument.

              The risk does not go up for my family because of self-driving cars. That’s innate to the 40k vs 20k numbers.

              So the proper question is: if your family was killed in an accident, what would be your reaction if it was a human driver vs AI? For me:

              • human driver - incredibly mad because it was probably preventable
              • AI - still mad, but supportive of self-driving improvements because it probably can be patched

              The first would make me bitter and probably anti-driving, whereas the second would make me constructive and want to help people understand the truth of how it works. I’m still mad in both cases, but the second is more constructive.

              Seeing someone go to jail doesn’t fix anything.

              • exanime@lemmy.today
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                2
                ·
                edit-2
                8 months ago

                Yes, it’s a thought experiment… Not a fair question, just trying to put it in perspective

                Anyone who understands stats would agree 40k death is worse than 20k but it also depends on other factors. All things being equal to today, the 20k proposition is only benefit

                But if we look into the nuance and details emerge, the formula changes. For example, here it’s been discussed that there may be nobody liable. If that’s the case, we win by halving death (absolutely a win) but now the remaining 20k may be left with no justice… Worse, it absolutely creates a perverse incentive for these companies, without liability exposure, to do whatever to maximize profit

                So, not trying to be a contrarian here… I just want to avoid the polarization that is now the rule online… Nothing is just black and white

                • sugar_in_your_tea@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  8 months ago

                  left with no justice

                  But they’d get restitution through insurance. Even if nobody is going to jail, there will still be insurance claims.

                  I agree that there is nuance here, and I think it can largely be solved without a huge change to much of anything. We don’t need some exec or software developer to go to jail for justice to be served, provided they are financially responsible. If the benefits truly do outright the risks, this system should work.

                  Tesla isn’t taking that responsibility, but Mercedes seems to be. Drivers involved in an accident where the self-driving feature was engaged have the right to sue the manufacturer for defects. That’s not necessarily the case for class 2 driving, since the driver is responsible for staying alert and needs to be in contact with the steering wheel. With class 3, that goes away, so the driver could legitimately not be touching the wheel at all when the car is in self-driving mode. My understanding is the insurance company can sue on their customer’s behalf.

                  So the path forward is to set legal precedent assigning fault to manufacturers to get monetary compensation, and let the price of cars and insurance work out the details.

                  • exanime@lemmy.today
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    arrow-down
                    1
                    ·
                    8 months ago

                    But they’d get restitution through insurance. Even if nobody is going to jail, there will still be insurance claims.

                    And that’s where I’m aiming at… If Mercedes decides, like Ford did before them, that it’s cheaper to pay out the insurance claims they lose instead of fixing their bugs then innocent people will have to die so Mercedes can keep up their profit margins.

                    That’s exactly the point I’m trying to make

                    You seem to argue that, on the unproven premise that current AI is better than human drivers, we should let corporations test it out in the real world even if they are not criminally liable ever. For me, that’s a bad deal.

                    Now, imagine we go down this rabbit hole… It’s already 10x cheaper to lobby USA politicians to limit Mercedes liability than it would be for them to actually start paying wrongly death claims

                    In Texas, if you doctor shows up drunk for surgery and leaves you quadriplegic or kills you, the biggest liability exposure has been limited to 250k

                    I love tech and I do believe science, knowledge and the tech it can produce could improve our lives in unimaginable ways… But as long as our approach to it continues to be profit over people, socialise the risk - privatize the profit and corporation being citizens in all aspects except liability, we will never get there

    • slumberlust@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 months ago

      The question for me is not what margins the feature is performing on, as they will likely be better than human error raters, but how they market the product irresponsiblely.