THE SENATE UNANIMOUSLY passed a bipartisan bill to provide recourse to victims of porn deepfakes — or sexually-explicit, non-consensual images created with artificial intelligence.

The legislation, called the Disrupt Explicit Forged Images and Non-Consensual Edits (DEFIANCE) Act — passed in Congress’ upper chamber on Tuesday.  The legislation has been led by Sens. Dick Durbin (D-Ill.) and Lindsey Graham (R-S.C.), as well as Rep. Alexandria Ocasio-Cortez (D-N.Y.) in the House.

The legislation would amend the Violence Against Women Act (VAWA) to allow people to sue those who produce, distribute, or receive the deepfake pornography, if they “knew or recklessly disregarded” the fact that the victim did not consent to those images.

  • NauticalNoodle
    link
    fedilink
    arrow-up
    39
    arrow-down
    10
    ·
    edit-2
    1 month ago

    Am I the only one that still gets uncomfortable every time the government tries to regulate advanced technology?

    It’s not a Libertarian thing to me as much as it’s a ‘politicians don’t understand technology thing.’

    • Todd Bonzalez@lemm.ee
      link
      fedilink
      arrow-up
      47
      arrow-down
      4
      ·
      1 month ago

      Arguing against laws that prohibit sexual exploitation with high tech tools, because of the nature of technology, would be like arguing against laws that prohibit rape because of the nature of human sexuality.

      The “it still is going to happen” argument doesn’t matter, because the point of the law isn’t to eliminate something 100%, it is to create consequences for those who continue to do what the law prohibits.

      It’s not some slippery slope either, it is extremely easy not to make involuntary pornography of other people.

      • Contravariant@lemmy.world
        link
        fedilink
        arrow-up
        12
        arrow-down
        2
        ·
        1 month ago

        The worrying aspect of these laws are always that they focus too much on the method. This law claims to be about preventing a particular new technology, but then goes on to apply to all software.

        And frankly if you need a clause about how someone is making fake pornography of someone then something is off. Something shouldn’t be illegal simply because it is easy.

        Deepfakes shouldn’t be any more or less illegal than photos made of a doppelgänger or an extremely photorealistic painting (and does photorealism even matter? To the victims, I mean.). A good law should explain why those actions are illegal and when and not just restrict itself to applying solely to ‘technology’ and say oh if it only restricts technology then we should be all right.

        • postmateDumbass@lemmy.world
          link
          fedilink
          arrow-up
          13
          arrow-down
          1
          ·
          1 month ago

          I am not ok with making art a crime.

          Do not give away basic human rights because of emotional appeals.

          If sexual blackmail is the problem, prosecute sexual blackmail.

          • Todd Bonzalez@lemm.ee
            link
            fedilink
            arrow-up
            4
            arrow-down
            2
            ·
            1 month ago

            Where do you draw the line at “art”. If nonconsensual pornography is acceptable to you, how about child porn? Where does art end and crime begin?

              • Todd Bonzalez@lemm.ee
                link
                fedilink
                arrow-up
                6
                ·
                edit-2
                1 month ago

                Okay, and we already have laws against nonconsensual pornography. But here’s the thing: those laws could easily be argued to only count for actual photographs or videos, in that many of these laws were written to tackle “revenge porn” and not “fake porn”, so someone making nonconsensual porn with AI could get charges dropped on the grounds that the porn wasn’t legitimate, or didn’t meet the definition of pornography starring the victim as defined by existing laws (because those laws were written before AI made this a mainstream problem).

                Making laws that explicitly list things like photoshop and AI faceswap or “roop” tools that can create artificial nonconsensual pornography lays the groundwork for consequences for people who sexually abuse others in this way.

                And let’s be realistic, if you use AI to make nonconsensual porn for personal purposes, you may be a creep, but you also aren’t going to get caught if you keep it to yourself. You also aren’t going to harm the person you’re making porn of if you’re not distributing the content you’re creating. You still shouldn’t do it, but it’s practically legal if you simply keep your weird shit to yourself and don’t publicly victimize someone with it. The people who will get caught are going to be in the latter category, and I have no problem seeing the book thrown at them.

                • postmateDumbass@lemmy.world
                  link
                  fedilink
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  1 month ago

                  Just as you point out revenge porn laws have a problem because they miss the mark covering fake porn, outlawing the methods amd mediums of art ued today is also going to miss the mark. Aside from the oppression issues, new methods will come along and circumvemt the targeted laws.

                  Just need to write laws that target the crime, here it would seem to be a combination of defamation, privacy, and assault in play for using sexuality or nudity to harm an individual and/or their reputation. Without harm, what crime is there?

        • JovialMicrobial@lemm.ee
          link
          fedilink
          arrow-up
          3
          ·
          1 month ago

          You can tell when a painting is a painting. Even it’s photorealistic. You know a person created it and that it’s fiction. Often these are hung in galleries where people expect to see art.

          Deep fakes exist to fool people into thinking someone did something(like pornography) when they didn’t…usually with the intention of causing harm to their reputation. That’s already illegal due to defamation laws, so really it’s just an extention of those combined with revenge porn laws.

          The reason they have to include the type of tech in the law is because that tech made it possible for unskilled bad actors to get on it…therefore there’ll be more people committing these types of crimes against others. It’s a good thing they’re addressing this issue.

          • Contravariant@lemmy.world
            link
            fedilink
            arrow-up
            3
            arrow-down
            1
            ·
            edit-2
            1 month ago

            The reason they have to include the type of tech in the law is because that tech made it possible for unskilled bad actors to get on it

            Yeah, and that’s the part I don’t like. If you can’t define why it’s bad without taking into account the skill level of the criminal then I’m not convinced it’s bad.

            As you point out defamation is already illegal and deliberately spreading false information about someone with the intent to harm their reputation is obviously wrong and way easier to define.

            And is that not why you consider a painting less ‘bad’? Because it couldn’t be misconstrued as evidence? Note that the act explicitly says a digital forgery should be considered a forgery even when it’s made abundantly clear that it’s not authentic.

            • JovialMicrobial@lemm.ee
              link
              fedilink
              arrow-up
              5
              ·
              1 month ago

              Look, I’m a professional artist. The general rule is you have to change something 15% to 30%(depending on location) for it to not come into violation of copyright laws. That’s why you see satirical depictions of brands in cartoons and such.

              This new law has to take into consideration art laws, defamation laws, revenge porn laws, slander laws, and the right for a person to own their likeness.

              It is absolutely necessary to reign this in before serious harm is done to someone. The point of writing a law to address this specific issue is because for the law actually be effective, it must be written to address the specific problem this technology presents. I listed the other laws to show its consistent with ones we already have. There’s nothing wrong with adding in another to protect people.

              As for the unskilled part, the point of that is a skilled person creating deepfake porn by hand, frame by frame should get in as much trouble as an unskilled person using ai. The AI is just going to make it so more unethical people are making this crap…so more if it will exist. That’s a problem that needs addressing.

              You have a nice day now.

        • ByteJunk@lemmy.world
          link
          fedilink
          arrow-up
          2
          ·
          1 month ago

          There are stark differences between the scenarios you’re presenting, but going to the core of your point, is it even legal to paint a photorealistic nude?

          I don’t know of any court cases about this specific subject, but I remember when Rush painted Tiger Woods (“The masters at Augusta”), he was sued.

          He got away with not having to pay money to Tiger Woods, but partly because it’s a stylized painting and it pushed towards first amendment rights. This wouldn’t work in a photorealistic depiction, so it seems highly unlikely that such a painting would be OK…

      • UnderpantsWeevil@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 month ago

        it is extremely easy not to make involuntary pornography of other people.

        Eh. The term is ill-defined, so I can see some ultra-orthodox right-wing judge trying to argue that - say - jokes about JD Vance fucking a couch constitute violations of the revenge porn law. I can see some baroque interpretation by Scalia used to prohibit all forms of digitally transmitted pornography. I can also see some asshole trying to claim baby pictures on Facebook leave the company or even the parent liable for child pornography. Etc, etc.

        But a lot of this boils down to vindictive and despicable politicians trying to inflict harm on political opponents by any means necessary. The notion that we can’t have any kind of technology regulation because bad politicians and sadistic cops exist leaves us ceding the entire legislative process to the conservatives who we know are going to abuse the law.

        We shouldn’t be afraid to do the right thing now on the grounds that someone else might do the wrong thing tomorrow.

      • Lets_Eat_Grandma@lemm.ee
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        edit-2
        1 month ago

        People have been using magazine clippings glued onto porn actress heads since before I was born. Photoshop for over a decade.

        How do you delineate between a really good photoshop and “ai”?

          • Lets_Eat_Grandma@lemm.ee
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            3
            ·
            1 month ago

            So you’re saying if someone makes a nude that is remotely similar to your likeness you can sue them.

            What do you do about identical twins if one chooses to be a porn star and takes self shots? Wouldn’t it look the same? Is it a crime to sell nudes if you have an identical twin?

            What about anybody who is not related but looks VERY SIMILAR - we’ve probably all heard stories of this happening.

            Finally, how do you know if it’s a US citizen that created the image vs anybody in any other country not bound by US laws?

            What if an AI creates a nude and then a child is born, and 20 years later they grow up to look identical to the ai generated image?

            There’s so many reasons why generated images should be treated like art and protected as free speech imo. It’s one thing if someone you know makes fake nudes of you and then uses them to ruin your image - that’s likely covered under many other laws including something like slander.

            People have been going to 11 trying to do anything preventing machine learning from being used for absolutely anything. It’s completely predictable because everyone wants a cut of whatever wealth may be generated by a new technology but maybe we should adapt to the new tool rather than punishing everybody for using it. AI is quickly turning into a tool that will only be usable by multibillion dollar companies with in house legal teams that can handle all the lawsuits.

            • NιƙƙιDιɱҽʂ@lemmy.world
              link
              fedilink
              arrow-up
              2
              ·
              1 month ago

              So you’re saying if someone makes a nude that is remotely similar to your likeness you can sue them.

              The law specifies that the images are indistinguishable from reality and are presented as you, directly or indirectly.

              What do you do about identical twins if one chooses to be a porn star and takes self shots? Wouldn’t it look the same? Is it a crime to sell nudes if you have an identical twin?

              If the images were of twin A and presented as twin B, I think this law would apply, as it would be fake porn of twin B.

              What about anybody who is not related but looks VERY SIMILAR - we’ve probably all heard stories of this happening.

              Again, if it’s being presented as someone it’s not, that’s an issue.

              Finally, how do you know if it’s a US citizen that created the image vs anybody in any other country not bound by US laws?

              Then the law does not apply. That’s literally how all laws work.

              What if an AI creates a nude and then a child is born, and 20 years later they grow up to look identical to the ai generated image?

              Was it presented as a nude of that person that did not yet exist? Impressive that they knew what their parents would name them ahead of time. Again, it must be presented as this person, directly or indirectly. This scenario couldn’t happen.

              Regarding trying to ruin someone’s image, I imagine that would indeed fall under some form of defamation laws, although not slander as that is specifically spoken words. I do agree we must tread carefully regarding free speech rights, however are we not also expected to have our rights to privacy? Even if someone isn’t trying to defame us, even if an image is fake, if at face value it’s completely indistinguishable from being real does it make a difference? Obviously there’s no simple solution and I think I agree with you that a law such as this probably isn’t it.

              I’m a bit lost on your last bit, however. Are you saying this law will further push AI into the hands of large corporations? I don’t see that, so much as see them being forced to implement stronger filters, while pushing users to the open source community in search to get around them. Horny people gonna horny and this law won’t stop that, it’d only stop public facing models from producing such content and stop individuals from distributing it.

              • Lets_Eat_Grandma@lemm.ee
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                1 month ago

                I’m a bit lost on your last bit, however. Are you saying this law will further push AI into the hands of large corporations? I don’t see that, so much as see them being forced to implement stronger filters, while pushing users to the open source community in search to get around them. Horny people gonna horny and this law won’t stop that, it’d only stop public facing models from producing such content and stop individuals from distributing it.

                Remember how taxi medallions were worth millions and millions of dollars in NYC and Boston before uber? (1.2m was the peak per medallion in nyc. They dropped to like 35k at one point and are now around 140k.) Remember how uber got billions in VC money to fight the taxi industry lobbyists and effectively operates despite violating the systems that were in place for ages that prevented small independent operators from being taxis without having a bunch of seed money? Those billions of dollars let them win the legal fights and continue operating. You or I could never have challenged it.

                That’s the comparison I was making. If you regulate AI into the ground the only innovations and usage will come from big money interests, because they can eat the lawsuits. Individuals can’t eat the lawsuits. The law only applies to the small fries, the big guys cheat and get away with it because there is no transparency and a whole lot of pinky swears.

      • aidan@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        1 month ago

        Arguing against laws that prohibit sexual exploitation with high tech tools

        I think it’s gross but, it’s not sexual exploitation unless people writing homoerotic fanfictions about YouTubers is also sexual exploitation imo

    • BlackPenguins@lemmy.world
      link
      fedilink
      arrow-up
      13
      arrow-down
      1
      ·
      1 month ago

      AOC is young. I think she’s a politician that does understand the technology and if she needs to she can educate the other boomers on the bill.

      • cmhe@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        1 month ago

        I would be careful about assuming knowledge based on age. Young people might use technology without understanding it, and old people might understand it and don’t want to use it.

        Technology needs to be regulated, and I would not trust people with profit incentives to do so.

        IMO, it is always important to investigate if a regulation wants to prevent a real issue or if they just mention some populist reasons for doing whatever they want.

        • BlackPenguins@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          1 month ago

          I do agree that age is not the determining factor but it definitely plays one. But yeah we need experts advising our old government.

    • Tamo240@programming.dev
      link
      fedilink
      arrow-up
      4
      ·
      edit-2
      1 month ago

      Why even specifically mention AI? Are there already laws that cover creating a sexually explicit likeness without consent e.g. with photoshop or just painting one? If so why wouldn’t those laws also cover AI and if not why wouldn’t the law also wish to cover these cases?

      • NauticalNoodle
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        1 month ago

        I didn’t post with an alternative solution in mind as much as I was looking to elicit conversation that provided more perspective and context that would assuage my concerns. -I’ve since got that from some of the others responses.

    • gorgori@lemmy.world
      link
      fedilink
      arrow-up
      4
      arrow-down
      1
      ·
      1 month ago

      You need to start somewhere. Regulation will always lag behind technology. But sooner or later things will get regulated. Once a good number of people are affected by something, rules will be brought in by the people. That’s how democracy works.

      These rules are never perfect. Sometimes, to make rules effective, they have to be multilayered (swiss cheese model). But that makes them too expensive to implement. So eventually things end in a compromise where cost and effectiveness balance.

    • 11111one11111@lemmy.world
      link
      fedilink
      arrow-up
      4
      arrow-down
      2
      ·
      1 month ago

      —It’s not a Libertarian thing to me as much as it’s a ‘politicians don’t understand firearms thing.’

      —It’s not a Libertarian thing to me as much as it’s a ‘politicians don’t understand healthcare thing.’

      —It’s not a Libertarian thing to me as much as it’s a ‘politicians don’t understand medical thing.’

      • Todd Bonzalez@lemm.ee
        link
        fedilink
        arrow-up
        10
        arrow-down
        1
        ·
        1 month ago

        🎯

        This kind of argument is always just a thought-terminator. No actual argument for why the law is bad, just a low-effort jab that lawmakers are too stupid to pass a good law.

        • NauticalNoodle
          link
          fedilink
          arrow-up
          2
          ·
          1 month ago

          It’s okay, I’m pretty sure they are simply showing that Libertarian ideology is based in a mistrust of politicians — and to a degree, they’re right. I don’t typically promote or discuss the ideology. I just think as I currently watch YouTube videos on the nuances of cybersecurity, that it’s so easy for well-informed professional experts to make far reaching mistakes, that it seems absurd to expect uneducated politicians to create regulations that are simultaneously targeted enough to not cause excess issues while still broad enough to effectively reduce the unwanted behavior.

          I still think back to Clinton arguing that regulating the internet was akin to “nailing jello to the wall.”

          • Todd Bonzalez@lemm.ee
            link
            fedilink
            arrow-up
            3
            ·
            1 month ago

            I still think back to Clinton arguing that regulating the internet was akin to “nailing jello to the wall.”

            But the internet is extremely well regulated, thanks to organizations like ICANN & IANA whose entire existence is the result of a U.S. Department of Commerce contract to properly regulate names, addresses, and routing tables. Failure to obey the rules will get your peering agreements terminated, and sever you from the Internet. Failure to cut off rule violators can get you in trouble.

            It’s what makes sure ISPs don’t break the Internet doing arbitrary shit, and keeps shady groups from improperly routing internet traffic through their datacenters without consequences.

            I think just revisiting the point that politicians aren’t going to be exact when talking about highly technical things like this, and that Clinton very likely meant “The Web” and not “The Internet”.

        • 11111one11111@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          1 month ago

          Because I don’t want to take the time going through every legislative framework I just listed to cite the shitpile of mistakes and revisions or because youve never read the pile of redacted and revised subparagraphs found in every single regulatory legislation? The fact you think it’s that simple to compile proof shows you have never read any actual legislative documents. If you want proof, ill list off the top of my head the shit NYS fucked up because Ive read them some what recently:

          SAFE Act:

          -Original framework had no ban on high capacity magazines because they thought magazines were disposable and as long as they banned the sale the existing HCM’s would be phased out.

          -Required AR registration with no system to enforce it, then continued to deny all FOIL requests and judicial requests of the number of registered AR’s because the turnout was so dismal.

          -Issues red flag laws that have no method of enforcement or way of including when a person is a red flag offender on the background check performed at point of sale.

          -Added the requirement for pistol permit applicants to have range training but never amended the part of the law that bans anyone from using a pistol without being approved for a pistol permit.

          -They couldn’t appropriately define imminent danger and had the state supreme court overturn the regulation entirely

          Recreational Marijuana: seriously just fuckin google this one. It’s so embarrassing how bad they botched the entire thing.

          -NYS had millions of dollars of marijuana crops with nowhere to go because they released tons of farming licenses and barely any dispensary licenses.

          honestly I’m not taking the time to list all of these. Google it and you will see resounding incompetence.

      • NauticalNoodle
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        edit-2
        1 month ago

        Politicians understand Healthcare and guns, they just don’t care about the fallout.

    • mindbleach@sh.itjust.works
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      1 month ago

      I’m definitely worried the copyright addicts will try to kill any form of showing images to a computer.

      If they push this as ‘your daughter does Dallas’ then oh well of course everyone’s against it… but it’s a real short step from there, to networks simply not being allowed to know what politicians look like, to networks simply not being allowed to know what Superman looks like, to effectively strangling this cool new thing.

    • neidu2@feddit.nl
      link
      fedilink
      arrow-up
      2
      arrow-down
      2
      ·
      edit-2
      1 month ago

      At least they have now started to try to learn about the tech they’re trying to regulate, as opposed to Ted Stevens who obviously just read a prepared speech someone else had made in his famous “Series of Tubes” speech.

      • postmateDumbass@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        1 month ago

        More like the silicon valley funding machine had someone write up the bill and hand it over to a senator to retype on official letterhead.