• hoshikarakitaridia@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    7 months ago

    You made me think about this for a second.

    In my head, the reason is not specifically to punish the driver, but to make drivers always be aware and ready to take control again. Yes 100 ppl will have 1000 different ways to react to such a software error, but you need ppl to pay attention, and in law the only way is to use punishment. Obviously this needs to be well calculated but either you have multiple lines of defense (the software, the driver, maybe even additional safety features) or you have to remove the autonomous system.

    • treefrog@lemm.ee
      link
      fedilink
      English
      arrow-up
      5
      ·
      7 months ago

      People are naturally going to pay less attention the more cars drive for them. You can’t partially automate steering. Driver assisted steering is as close as it can be before the liability needs to fall on Tesla and other software manufacturers. A car isn’t a plane. The driver needs to be in control when split second decisions happen, like a child running after a ball.

      If I’m paying for an autopilot, I’m not the pilot. I.e., the driver. The car is. And Tesla’s marketing bullshit and lawyers are going to fail here. This does not fall under puffery. It’s false advertising that’s causing consumers to place undue trust in a product. And the insurance industry is quite concerned just where the liability falls in all of this as well. And as they’re the ones currently having to pay out claims when Tesla wins, they have a vested interest seeing that Tesla doesn’t.

    • michaelmrose@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      7 months ago

      It doesn’t matter for practical purposes you can’t make people pay attention as if driving without the actual engagement of driving. There is going to be a delay in taking over and in a lot of cases it wont matter by the time the human is effectively in control.