doesn’t it follow that AI-generated CSAM can only be generated if the AI has been trained on CSAM?

This article even explicitely says as much.

My question is: why aren’t OpenAI, Google, Microsoft, Anthropic… sued for possession of CSAM? It’s clearly in their training datasets.

  • hendrik@palaver.p3x.de
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 days ago

    Well, it can draw an astronaut on a horse, and I doubt it had seen lots of astronauts on horses…

    • ExtremeDullard@lemmy.sdf.orgOP
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      2 days ago

      Yeah but the article suggests that pedos train their local AI on existing CSAM, which would indicate that it’s somehow needed to generate AI-generated CSAM. Otherwise why would they bother? They’d just feed them images of children in innocent settings and images of ordinary porn to get their local AI to generate CSAM.

      • Deceptichum@quokk.au
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        That’s not exactly how it works.

        It can “understand” different concepts and mix them, without having to see the combination before hand.

        As for the training thing, that would probably be more LORA. They’re like add-ons you can put on your AI to draw certain things better like a character, a pose, etc. not needed for the base model.

      • hendrik@palaver.p3x.de
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        It’s certainly technically possible. I suspect these AI models just aren’t good at it. So the pedophiles need to train them on actual images.

        I can imagine for example AI doesn’t know what puberty is since it has in fact not seen a lot of naked children. It would try to infer from all the internet porn it’s seen, and draw any female with big breasts, disregarding age. And that’s not how children actually look.

        I haven’t tried, since it’s illegal where I live. But that’s my suspicion why pedophiles bother with training models.

          • Rikudou_Sage@lemmings.world
            link
            fedilink
            arrow-up
            0
            ·
            2 days ago

            And again, what’s the source? The great thing with articles about CSAM is that you don’t need sources, everyone just assumes you have them, but obviously cannot share.

            Did at least one pedo try that? Most likely yes. Is it the best way to get good quality fake CSAM? Not at all.

            • ExtremeDullard@lemmy.sdf.orgOP
              link
              fedilink
              arrow-up
              0
              ·
              2 days ago

              I don’t know man. But I assume associations concerned with child abuse are all over that shit and checking it out. I’m not a specialist of CSAM but I assume an article that says old victims show up in previously-unseen images doesn’t lie, because why would it? It’s not like Wired is a pedo outlet…

              Also, it was just a question. I’m not trying to convince you of anything 🙂

              • hendrik@palaver.p3x.de
                link
                fedilink
                English
                arrow-up
                0
                ·
                edit-2
                2 days ago

                I think that aricle lacks nuance. It’s a bit baity and attends to the usual talking points without contextualizing the numbers or what’s actually happening out there, the consequences or the harm. That makes me believe the author just wants to push some point across.

                But I’ve yet to read a good article on this. Most articles are like this one. But yeah, are a few thousand images much in the context of crime that’s happening online? Where are these numbers from and what’s with the claim that there are more actual pictures out there? I seriously doubt that at this point, if it’s so easy to generate images. And what consequences does all of this have? Does it mean an increase or a decrease in abuse? And lots of services have implemented filters… Are the platforms doing their due diligence? Is this a general societal issue or criminals doing crime?

      • GBU_28@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 days ago

        Training an existing model on a specific set of new data is known as “fine tuning”.

        A base model has broad world knowledge and the ability to generate outputs of things it hasn’t specifically seen, but a tuned model will provide “better” (fucking yuck to even write it) results.

        The closer your training data is to your desired result, the better.