Hugh Nelson, 27, from Bolton, jailed after transforming normal pictures of children into sexual abuse imagery

A man who used AI to create child abuse images using photographs of real children has been sentenced to 18 years in prison.

In the first prosecution of its kind in the UK, Hugh Nelson, 27, from Bolton, was convicted of 16 child sexual abuse offences in August, after an investigation by Greater Manchester police (GMP).

Nelson had used Daz 3D, a computer programme with an AI function, to transform “normal” images of children into sexual abuse imagery, Greater Manchester police said. In some cases, paedophiles had commissioned the images, supplying photographs of children with whom they had contact in real life.

He was also found guilty of encouraging other offenders to commit rape.

  • Dr. Wesker@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    12
    ·
    edit-2
    2 months ago

    The fuck? Nothing about generating and distributing CSAM material is harmless, and especially if images of real children are being used to generate it.

      • otp@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        6
        ·
        2 months ago

        Would it harm you to have identifiable nude photos of you available for download on the internet?

        Would it harm you to have identifiable nude photos of you being used to train AI so that it can create more nude images that are “inspired” by your nude images?

        Would you be happy to upload your children’s nude photos so that people on the internet can share them and masturbate to them? Would you be harmed if your parents had done that with your images?

        • Mango@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          2 months ago

          As a child? No. In fact, I can milk that for pity money. As an adult, I can’t see how it matters. I don’t like it, but it doesn’t hurt me any.

          Also definitely no.

          Again, double no.

          • otp@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            3
            ·
            edit-2
            2 months ago

            To clarify, the second last question about your children was “would you be happy to …”

            If you wouldn’t be happy to, then why not?

            And if you would be happy to do that, then why? Lol

            • Mango@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              ·
              2 months ago

              You got me there. It’s definitely weird and gross and therefore no. That’s harm enough, but that’s more a matter of it being published and real. This dude doing it for himself is hardly different to me from fantasizing in your head or drawing in your sketchbook. That said, what was his AI training material? He’s also doing this for other people and encouraging rape and shit.

              • otp@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                3
                ·
                2 months ago

                What makes it different than imagining it or drawing it is that the AI is using real photos as training material. If the parents are knowingly providing images, that’s questionable. If the AI is discovering CSAM images, that’s horrible. If it’s using non-CSAM images of children without the knowing consent of the parents, that’s pretty bad too.

                • Mango@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  2 months ago

                  How is AI using real photos any different from a person using their real memory?

                  • otp@sh.itjust.works
                    link
                    fedilink
                    English
                    arrow-up
                    2
                    ·
                    2 months ago

                    Because the AI publishes what it creates based on those images. The AI also doesn’t have imagination the way that a person does. It could accidentally create CSAM material with a child that looks exactly like someone’s child. And it can generate images that look like photos. Someone sketching something from memory can’t do that.

        • Cryophilia@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 months ago

          Most AI generated images are not of real, identifiable people. I agree that deepfake porn is bad, whether of a child or adult, but that’s a separate category.

          • otp@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 months ago

            You’re definitely right, and I’m aware. The smaller the sample size, though, the more likely an AI art generator would create something that looks very similar to a given individual.

            As well, some AI art generators accept prompt images to use as a starting point.