• webghost0101@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    48
    arrow-down
    16
    ·
    edit-2
    6 days ago

    I have mixed feelings about this prosecution of ai deepfakes.

    Like obviously people should have protection against becoming a victim of such and perpetrators should be held accountable.

    But the line “feds are currently testing whether existing laws protecting kids against abuse are enough to shield kids from AI harms” would be a incredibly dangerous precedent because those are mostly designed for actual physical sex crimes.

    As wrong as it is to create and distribute ai generated sex imagery involving non consenting people it is not even remotely as bad as actual rape and distributing real photos.

    • otp@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      6
      ·
      5 days ago

      I’m some jurisdictions, public urination can put you on a sex offender registry.

      It wouldn’t even matter if you’re trying to be discreet and just have to go but there’s no public washrooms around.

    • rottingleaf@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      39
      ·
      6 days ago

      Creating and distributing anything should be legal if no real person suffers during its creation and if it’s not intended at defamation, forgery, such things.

      • AstralPath@lemmy.ca
        link
        fedilink
        English
        arrow-up
        31
        arrow-down
        4
        ·
        6 days ago

        Bruh how is creating and distributing a non-consensual nude-ified picture of a young girl not a cause for suffering for the victim? Please, explain that to the class.

        Did you just not go to school as a kid? If so, that would explain your absolute ineptitude on this topic. Your opinion is some real “your body, my choice” kind of energy.

        • krashmo@lemmy.world
          link
          fedilink
          English
          arrow-up
          9
          arrow-down
          3
          ·
          6 days ago

          There’s a legitimate discussion to be had about harm reduction here. You’re approaching this topic from an all-or-nothing mindset but there’s quite a bit of research indicating that’s not really how it works in practice. Specifically as it relates to child pornography the argument goes that not allowing artificial material to be created leads to an increase in production of actual child pornography which obviously means more real children are being harmed than would be if other forms were not controlled in the same fashion. The same sort of logic could be applied to revenge porn, stolen selfies, or whatever else we’re calling the kind of thing this article is referring to. It may not be an identical scenario but I still think it would be fair to say that an AI generated image is not as damaging as a real one.

          That is not to say that nothing should be done in these situations. I haven’t decided what I think the right move is given the options in front of us but I think there’s quite a bit more nuance here than your comment would indicate.

          • MagicShel@lemmy.zip
            link
            fedilink
            English
            arrow-up
            10
            ·
            6 days ago

            I think this is probably a really good point. I have no issue with AI generated images, although obviously if they are used to do an illegal thing such has harassment or defamation, those things are still illegal.

            I’m of two minds when it comes to AI nudes of minors. The first is that if someone wants that and no actual person is harmed, I really don’t care. Let me caveat that here: I suspect there are people out there who, if inundated with fake CP, will then be driven to ideation about actual child abuse. And I think there is real harm done to that person and potentially the children if they go on to enact those fantasies. However I think it needs more data before I am willing to draw a firm conclusion.

            But the second is that a proliferation of AI CP means it will be very difficult to tell fakes from actual child abuse. And for that reason alone, I think it’s important that any distribution of CP, whether real or just realistic, must be illegal. Because at a minimum it wastes resources that could be used to assist actual children and find their abusers.

            So, absent further information, I think whatever a person whats to generate for themselves in private is just fine, but as soon as it starts to be distributed, I think that it must be illegal.

        • rottingleaf@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          17
          ·
          6 days ago

          Read my comment again.

          Your opinion is some real “your body, my choice” kind of energy.

          My advice to you would be to improve your reading comprehension before judging this way.

          In particular, the word “defamation”.

      • essteeyou@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        arrow-down
        2
        ·
        edit-2
        6 days ago

        Can you share a full-body shot of yourself please? Don’t worry, you won’t suffer while it gets used to create other content that we’ll distribute to your friends, family, classmates, coworkers, etc.

          • ShepherdPie@midwest.social
            link
            fedilink
            English
            arrow-up
            7
            ·
            6 days ago

            “Deepfakes” are edited pictures of real people. I’d be more inclined to agree with you on completely AI generated images but not something specifically intended to deceive others into thinking they’re viewing a real person’s image.

            • rottingleaf@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              3
              ·
              6 days ago

              Deepfakes are, however the top-level comment I was answering was not limited to deepfakes. And as my further discussion with its author shows, they too didn’t mean only deepfakes.

              Their opinion was that any kind of pornography portraying children, even if it’s not shared with others and not based on pics of real people, should be prosecuted just like making real child pornography.

              You know, this thread has once again reinforced me in my opinion that the best system of government is Aspie Reich. Only people with Aspergers should be allowed to make laws and judge and hold public posts. The rest of fucking chimps just don’t have what it takes to override their chimp instincts.

        • FiskFisk33@startrek.website
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 days ago

          intention is litigated every day. Intention is what differentiates murder from manslaughter. Intention is what differentiates free speech from defamation.

        • boatswain
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 days ago

          My understanding is that intention is not uncommonly litigated; I believe the question of “intent to deceive” is central to trademark law, for example. That’s also what the the “degrees” of murder etc are about.

          Disclaimer: I’m not a lawyer. I do read an awful lot of contacts and talk to lawyers.

        • rottingleaf@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          3
          ·
          6 days ago

          This is not a legal text, you little cheat.

          This is a sentence in natural language, want me to start asking such questions about everything you write?

          If you make a deepfake of someone and share it, then it’s defamation. Taking a picture voluntarily shared and editing it is not a crime.

          • EamonnMR@lemmy.sdf.org
            link
            fedilink
            English
            arrow-up
            1
            ·
            5 days ago

            By litigate I mean, if a person is creating something and says they don’t plan to distribute it, do we take their word for it?

            If it ends up getting distributed anyway, should we take their word that it was an accident?

            We consider people’s private data important enough that if you leak it even by mistake you are on the hook for that. You have a responsibility.

            I think that rather than framing this as something harmless unless distributed and therefore intent to distribute matters, we should treat it as something you have a responsibility not to create because it will be harmful when it is inevitably distributed.

      • Alphane Moon@lemmy.worldOP
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        6 days ago

        You would be fine with AI-gen porn images of your teenage daughter being distributed around the internet?

            • rottingleaf@lemmy.world
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              16
              ·
              6 days ago

              Making forged pics of someone else falls under defamation.

              It’s very clearly not rape, sexual abuse, child pornography or non-consensual pornography.

                • rottingleaf@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  3
                  ·
                  6 days ago

                  Even knowing what Internet is, I sometimes try to pretend the other side is arguing in good faith.

                  I mean, it’s as if someone pushed me and I would try to sue them for cutting my hand off. With that hand present.

                  I would understand the “this punishment is not enough, we have to do more” sentiment, but instead of “more” they are trying to alias a different action with an existing action with harsher punishment.