• marketsnodsbury@lemm.ee
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      3
      ·
      edit-2
      2 months ago

      The car should be programmed to self-destruct or take out the passengers always.

      Why not both?

      • VeganPizza69 Ⓥ@lemmy.worldOP
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        10
        ·
        edit-2
        2 months ago

        Users here don’t understanding the dilemma nor the programmatic aspects.

        The car has to be programmed to solve the dilemma on the spot:

        1. Crush the people outside to save the people inside.
        2. Intentionally crash into a large object or veer off road and risk crashing into a ditch.

        Not talking about it won’t make this go away. It will simply be some decision made by developers and maybe there’s a toggle for the car owner, a kill switch. Either way, it’s lose-lose.

        As we’re in fuck cars, I’m assuming that people understand that fuck cars. Why should this impunity of killing with cars be furthered by encoding it in automatic programming? Let the owners of vehicles face the immediate consequences of owning such vehicles. That’s fair. Don’t want to die in your robocar? Fine, drive very slowly and very rarely.

        • GeneralEmergency@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          ·
          2 months ago
          1. Crush the people outside to save the people inside.
          1. Intentionally crash into a large object or veer off road and risk crashing into a ditch.

          What?

          That’s not what happened here, and I struggle to imagine any situation where that’s the only two options.

          • VeganPizza69 Ⓥ@lemmy.worldOP
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            6
            ·
            edit-2
            2 months ago

            I struggle to imagine any situation where that’s the only two options.

            Alright, I’ll take this in good faith. Here’s how that happens:

            Speeding.

            As we all know here, speeding makes crashes way worse, and it makes the braking function fail proportionally.

            So, imagine:

            The killer road bot is speeding through a street. It’s a bit narrow, there are cars parked illegally on the sides.

            The killer road bot enters an intersection and makes a left turn with speed and a there’s someone on a crosswalk.

            The killer road bot controls at least these aspects of the car: brakes, acceleration, steering. The brakes can be engaged, but the speed makes them useless in preventing running over the person on the crosswalk. The acceleration is not useful. Everything is happening too fast really, and the killer road bot can’t even calculate which direction the person is walking in on the crosswalk.

            The only useful control left is direction by steering. The killer road bot thus has these choices:

            1. Maintain course, run over person on crosswalk
            2. Change course

            Choice 1 leads to the obvious outcome.

            Choice 2 branches out:

            2.1. Turn left

            2.2. Turn right

            If the killer road bot turns left (2.1), it flips the car over and sends it rolling into other cars, thus endangering the passenger(s).

            If the killed road bot turns right (2.2), it hits a large tree.

            These are the only options.

            edit: typo

            • GeneralEmergency@lemmy.world
              link
              fedilink
              English
              arrow-up
              6
              ·
              2 months ago

              Oh I get now. You have a preconceived agenda that makes this discussion entirely pointless. Either that or you value the trolley problem way too much.

              • VeganPizza69 Ⓥ@lemmy.worldOP
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                5
                ·
                2 months ago

                Yeah, my agenda is public health and equality. I don’t like it when a special class of people has impunity for roaming the land harming people, even less so when that’s automatic.

                  • VeganPizza69 Ⓥ@lemmy.worldOP
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    arrow-down
                    3
                    ·
                    2 months ago

                    Self-destruct doesn’t exclusively mean “blow up”. When the AI system chooses to run off the road or into some tree, that’s also self-destruction.

                  • VeganPizza69 Ⓥ@lemmy.worldOP
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    1 month ago

                    then why do you want to get rid of cars?

                    because cars are a means of stratification and denial of rights . Cars can never be universal rights. It’s literally impossible, so they have inequality baked in as a “car system”.

                    Do you actively hate disabled people?

                    Quite the opposite.

            • Freeman@lemmings.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              28 days ago

              I mean a autonomous vehicle should be programmed to not speed and even not drive faster than reasonable in the present condition.

              In switzerland we have a law that you are not allowed to drive faster than the speed with wich you can come to a full stop at the farthest spot on the road that you can see. (So in a curve you have to drive slowly, because there could be something on the street right in front of you.)

              If a autonomous vehicle respects such rules, then it at least has eniugh time to calculate several outcomes and choose one which has the least damage potential.

              The trolleyproblem is not applicable here as its not a theoretical situation but a practical one.