A Telegram user who advertises their services on Twitter will create an AI-generated pornographic image of anyone in the world for as little as $10 if users send them pictures of that person. Like many other Telegram communities and users producing nonconsensual AI-generated sexual images, this user creates fake nude images of celebrities, including images of minors in swimsuits, but is particularly notable because it plainly and openly shows one of the most severe harms of generative AI tools: easily creating nonconsensual pornography of ordinary people.

    • @goldteeth@lemmy.dbzer0.com
      link
      fedilink
      English
      78
      edit-2
      7 months ago

      “Djinn”, specifically, being the correct word choice. We’re way past fun-loving blue cartoon Robin Williams genies granting wishes, doing impressions of Jack Nicholson and getting into madcap hijinks. We’re back into fuckin’… shapeshifting cobras woven of fire and dust by the archdevil Iblis, hiding in caves and slithering out into the desert at night to tempt mortal men to sin. That mythologically-accurate shit.

    • @conciselyverbose@sh.itjust.works
      link
      fedilink
      English
      427 months ago

      Doesn’t mean distribution should be legal.

      People are going to do what they’re going to do, and the existence of this isn’t an argument to put spyware on everyone’s computer to catch it or whatever crazy extreme you can take it to.

      But distributing nudes of someone without their consent, real or fake, should be treated as the clear sexual harassment it is, and result in meaningful criminal prosecution.

      • @treadful@lemmy.zip
        link
        fedilink
        English
        167 months ago

        Almost always it makes more sense to ban the action, not the tool. Especially for tools with such generalized use cases.

      • @ITGuyLevi@programming.dev
        link
        fedilink
        English
        67 months ago

        While I agree in spirit, any law surrounding it would need to be very clearly worded, with certain exceptions carved out. Which I’m sure wouldn’t happen.

        I could easily see people thinking something was of them, when in reality it was of someone else.

      • @cley_faye@lemmy.world
        link
        fedilink
        English
        37 months ago

        I’m not familiar with the US laws, but… isn’t it already some form of crime or something to distribute nude of someone without their consent? This should not change whether AI is involved or not.

        • @T156@lemmy.world
          link
          fedilink
          English
          4
          edit-2
          7 months ago

          It might depend on whether fabricating them wholesale would be considered a nude or not. Legally, it could be considered a different person if you’re making it, since the “nude” is someone else, and you’re putting their face on top, or it’s a complete fabrication made by a computer.

          Unclear if it would still count if it was someone else and they were lying about it being the victim, for example, pretending a headless mirror-nude was sent by the victim, when it was sent by someone else.

    • @roscoe@lemmy.dbzer0.com
      link
      fedilink
      English
      117 months ago

      As soon as anyone can do this on their own machine with no third parties involved all laws and other measures being discussed will be moot.

      We can punish nonconsensual sharing but that’s about it.

      • @CeeBee@lemmy.world
        link
        fedilink
        English
        377 months ago

        As soon as anyone can do this on their own machine with no third parties involved

        We’ve been there for a while now

        • @roscoe@lemmy.dbzer0.com
          link
          fedilink
          English
          67 months ago

          Some people can, I wouldn’t even know where to start. And is the photo/video generator completely on home machines without any processing being done remotely already?

          I’m thinking about a future where simple tools are available where anyone could just drop in a photo or two and get anything up to a VR porn video.

          • @CeeBee@lemmy.world
            link
            fedilink
            English
            257 months ago

            And is the photo/video generator completely on home machines without any processing being done remotely already?

            Yes

              • JDPoZ
                link
                fedilink
                English
                18
                edit-2
                7 months ago

                Stable Diffusion has been easily locally installed and runnable on any decent GPU for 2 years at this point.

                Combine that with Civitai.com for easy to download and run models of almost anything you can imagine - IP, celebrity, concepts, etc… and the possibilities have been endless.

                In fact, with completely free apps like Draw Things on iOS, which allows you to run it on YOUR PHONE locally - where you can download models, tweak, customize, hand it images directly from your mobile device’s library… making this stuff is now trivial on the go.

                • @T156@lemmy.world
                  link
                  fedilink
                  English
                  27 months ago

                  Tensor processors/AI accelerators have also been a thing on new hardware for a while. Mobile devices have them, Intel/Apple include them with their processors, and it’s not uncommon to find them on newer graphics cards.

                  That would just make it easier compared to needing quite a powerful computer for that kind of task.

        • @yildolw@lemmy.world
          link
          fedilink
          English
          67 months ago

          You may be sued for damages if you sell those nude paintings of Rihanna at a large enough scale that Rihanna notices