Pedos ruin everything…

  • Scrubbles@poptalk.scrubbles.tech
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    7
    ·
    edit-2
    1 year ago

    I thought the same, but the article puts it into perspective - first training anything requires having content on hand, and that means children were exploited to get it. Second, by enabling and allowing it it allows only deeper desires, possibly pushing people even further. Definitely above our paygrades here to determine if that’s good or bad. The letter calls out that by allowing it we as society are normalizing it, saying it’s okay, and we definitely do not want it normalized.

    Unfortunately, this does probably mean the end of the wild west of AI generation. I don’t think this will stop tools from being created, it sounds like (even though Ars somehow demonized the term ‘open source’) tools like SD and training are going to stick around. (Probably because they know they’d just be forked and continued on later). But I think we’re going to see a lot more regulation on sharing models. Right now there are a few obvious sites with things being shared that even I’ve been shocked are allowed, I think we’re going to see a lot more rules on what can and can’t be uploaded.

    Even as the tech gets easier, fine tuning and training models is a much more involved process that most people won’t want to go through, and so stopping the sharing of those models will cut down on the vast majority. (Similar to those who share vs consume illegal material already on the internet)

    Looks like the first step they’re calling for is basically saying even if CSAM was created by AI, it should still be classified legally as CSAM, and so trading it and sharing it means a one way ticket to jail, and I’m okay with that personally.

    • xionzui@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      31
      arrow-down
      1
      ·
      1 year ago

      Actually, the image generation models can most likely generate that kind of material without ever having seen it before. There probably are people out there training them with real material, but it’s not an absolute prerequisite. They can generalize enough to create that combination from legal pornography and normal images of children

      • Stampela@startrek.website
        link
        fedilink
        English
        arrow-up
        15
        arrow-down
        1
        ·
        1 year ago

        Yup, same reason why you can ask for a fox using a crocodile as a mech and get a good result. The model has the concept of all things requested and mixes them (with varying success).

    • Onihikage@beehaw.org
      link
      fedilink
      English
      arrow-up
      8
      ·
      1 year ago

      first training anything requires having content on hand, and that means children were exploited to get it

      Does it actually require that, though? I feel like a model trained on a sufficiently diverse selection of adult humans would be able to render an approximation of CSAM even if no CSAM was actually used to create the model. If not now, then very, very soon.

      I’m not sure what to do about that, but I am sure that rather than something reasonable like what you’ve said (focusing on the distribution of such material), privacy will end up in the crosshairs, as usual. Humans have always used tools as extensions of ourselves, and these generative tools will soon be used as yet another extension to the mind, up there with search engines. I worry that the legislative responses we actually see on this will stray closer to thought-crime than I’m comfortable with.

      • Scrubbles@poptalk.scrubbles.tech
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        1 year ago

        You’re probably right, on both accounts.

        I’m worried about that too. While for me the jury is still out in my mind if using a tool like SD should count as art in terms of copyright (since it was trained on art that artists didn’t give permission for, especially for use in commercial cases), I think it’s a completely different question on if using it should be regulated.

        At the end of the day, these pedos can still pick up a pen and paper and draw what they want, is this not just doing the same thing but with tech? Now sharing that content can be discussed, but regulating what someone can create does feel wrong