• macniel@feddit.de
    link
    fedilink
    English
    arrow-up
    141
    arrow-down
    11
    ·
    4 months ago

    Mhm I have mixed feelings about this. I know that this entire thing is fucked up but isn’t it better to have generated stuff than having actual stuff that involved actual children?

    • pavnilschanda@lemmy.world
      link
      fedilink
      English
      arrow-up
      94
      arrow-down
      4
      ·
      4 months ago

      A problem that I see getting brought up is that generated AI images makes it harder to notice photos of actual victims, making it harder to locate and save them

        • ricecake@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          11
          ·
          4 months ago

          It does learn from real images, but it doesn’t need real images of what it’s generating to produce related content.
          As in, a network trained with no exposure to children is unlikely to be able to easily produce quality depictions of children. Without training on nudity, it’s unlikely to produce good results there as well.
          However, if it knows both concepts it can combine them readily enough, similar to how you know the concept of “bicycle” and that of “Neptune” and can readily enough imagine “Neptune riding an old fashioned bicycle around the sun while flaunting it’s tophat”.

          Under the hood, this type of AI is effectively a very sophisticated “error correction” system. It changes pixels in the image to try to “fix it” to matching the prompt, usually starting from a smear of random colors (static noise).
          That’s how it’s able to combine different concepts from a wide range of images to create things it’s never seen.

        • pavnilschanda@lemmy.world
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          1
          ·
          4 months ago

          True, but by their very nature their generations tend to create anonymous identities, and the sheer amount of them would make it harder for investigators to detect pictures of real, human victims (which can also include indicators of crime location.

      • TranscendentalEmpire@lemm.ee
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        4
        ·
        4 months ago

        Well that, and the idea of cathartic relief is increasingly being dispelled. Behaviour once thought to act as a pressure relief for harmful impulsive behaviour is more than likely just a pattern of escalation.

          • TranscendentalEmpire@lemm.ee
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            5
            ·
            4 months ago

            Catharsis theory predicts that venting anger should get rid of it and should therefore reduce subsequent aggression. The present findings, as well as previous findings, directly contradict catharsis theory (e.g., Bushman et al., 1999; Geen & Quanty, 1977). For reduc- ing anger and aggression, the worst possible advice to give people is to tell them to imagine their provocateur’s face on a pillow or punching bag as they wallop it, yet this is precisely what many pop psychologists advise people to do. If followed, such advice will only make people angrier and more aggressive.

            Source

            But there’s a lot more studies who have essentially said the same thing. The cathartic hypothesis is mainly a byproduct of the Freudian era of psychology, where hypothesis mainly just sounded good to someone on too much cocaine.

            Do you have a source of studies showing the opposite?

            • 9bananas@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              ·
              4 months ago

              your source is exclusively about aggressive behavior…

              it uses the term “arousal”, which is not referring to sexual arousal, but rather a state of heightened agitation.

              provide an actual source in support of your claim, or stop spreading misinformation.

            • BlanketsWithSmallpox@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              edit-2
              4 months ago

              Yes, but I’m too lazy to sauce everything again. If it’s not in my saved comments someone else will have to.

              E: couldn’t find it on my reddit either. I have too many saved comments lol.

    • retrospectology@lemmy.world
      link
      fedilink
      English
      arrow-up
      86
      arrow-down
      16
      ·
      4 months ago

      The arrest is only a positive. Allowing pedophiles to create AI CP is not a victimless crime. As others point out it muddies the water for CP of real children, but it also potentially would allow pedophiles easier ways to network in the open (if the images are legal they can easily be platformed and advertised), and networking between abusers absolutely emboldens them and results in more abuse.

      As a society we should never allow the normalization of sexualizing children.

      • nexguy@lemmy.world
        link
        fedilink
        English
        arrow-up
        33
        arrow-down
        3
        ·
        4 months ago

        Interesting. What do you think about drawn images? Is there a limit to how will the artist can be at drawing/painting? Stick figures vs life like paintings. Interesting line to consider.

        • retrospectology@lemmy.world
          link
          fedilink
          English
          arrow-up
          29
          arrow-down
          12
          ·
          4 months ago

          If it was photoreal and difficult to distinguish from real photos? Yes, it’s exactly the same.

          And even if it’s not photo real, communities that form around drawn child porn are toxic and dangerous as well. Sexualizing children is something I am 100% against.

        • acockworkorange@mander.xyz
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          4
          ·
          4 months ago

          I’m in favor of specific legislation criminalizing drawn CSAM. It’s definitely less severe than photographic CSAM, and it’s definitely harmful.

      • NewNewAccount@lemmy.world
        link
        fedilink
        English
        arrow-up
        24
        arrow-down
        2
        ·
        4 months ago

        networking between abusers absolutely emboldens them and results in more abuse.

        Is this proven or a common sense claim you’re making?

        • moitoi@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          The far right in France normalized its discourses and they are now at the top of the votes.

          Also in France, people talked about pedophilia at the TV in the 70s, 80s and at the beginning of the 90s. It was not just once in a while. It was frequent and open without any trouble. Writers would casually speak about sexual relationships with minors.

          The normalization will blur the limits between AI and reality for the worse. It will also make it more popular.

          The other point is also that people will always ends with the original. Again, politic is a good example. Conservatives try to mimic the far right to gain votes but at the end people vote for the far right…

          And, someone has a daughter. A pedophile takes a picture of her without asking and ask an AI to produce CP based on her. I don’t want to see things like this.

        • bassomitron@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          5
          ·
          edit-2
          4 months ago

          I wouldn’t be surprised if it’s a mixture of the two. It’s kind of like if you surround yourself with criminals regularly, you’re more likely to become one yourself. Not to say it’s a 100% given, just more probable.

          • Zorque@kbin.social
            link
            fedilink
            arrow-up
            17
            arrow-down
            2
            ·
            4 months ago

            So… its just a claim they’re making and you’re hoping it has actual backing.

            • bassomitron@lemmy.world
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              3
              ·
              edit-2
              4 months ago

              I’m not hoping anything, haha wtf? The comment above me asked if it was a proven statement or common sense and I said I wouldn’t be surprised if it’s both. I felt confident that if I googled it, there would more than likely be studies backing up a common sense statement like that, as I’ve read in the past how sending innocent people or people who committed minor misdemeanors to prison has influenced them negatively to commit crimes they might not have otherwise.

              And look at that, there are academic articles that do back it up:

              https://www.waldenu.edu/online-bachelors-programs/bs-in-criminal-justice/resource/what-influences-criminal-behavior

              Negative Social Environment

              Who we’re around can influence who we are. Just being in a high-crime neighborhood can increase our chances of turning to crime ourselves.4 But being in the presence of criminals is not the only way our environment can affect our behaviors. Research reveals that simply living in poverty increases our likelihood of being incarcerated. When we’re having trouble making ends meet, we’re under intense stress and more likely to resort to crime.

              https://www.law.ac.uk/resources/blog/is-prison-effective/

              Time in prison can actually make someone more likely to commit crime — by further exposing them to all sorts of criminal elements.

              Etc, etc.

              Turns out that your dominant social group and environment influences your behavior, what a shocking statement.

              • Zorque@kbin.social
                link
                fedilink
                arrow-up
                3
                arrow-down
                2
                ·
                4 months ago

                But you didn’t say you had proof with your comment, you said it was probable. Basically saying its common sense that its proven.

                Why are you getting aggressive about actually having to provide proof about something when saying its obvious?

                Also, that seems to imply that locking up people for AI offenses would then encourage truly reprehensible behavior by linking them with those who already engage in it.

                Almost like lumping people together as one big group, instead of having levels of grey area, means people are more likely to just go all in instead of sticking to something more morally defensible.

                • bassomitron@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  3
                  arrow-down
                  1
                  ·
                  4 months ago

                  Because it’s a casual discussion, I think it’s obnoxious when people constantly demand sources to be cited in online comments section when they could easily look it up themselves. This isn’t some academic or formal setting.

                  And I disagree, only the second source mentioned prisons explicitly. The first source mentions social environments as well. So it’s a damned if you do, damned if you don’t situation. Additionally, even if you consider the second source, that source mentions punishment reforms to prevent that undesirable side effect from occuring.

                  I find it ironic that you criticized me for not citing sources and then didn’t read the sources. But, whatever. Typical social media comments section moment.

      • lily33@lemm.ee
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        2
        ·
        edit-2
        4 months ago

        Actually, that’s not quite as clear.

        The conventional wisdom used to be, (normal) porn makes people more likely to commit sexual abuse (in general). Then scientists decided to look into that. Slowly, over time, they’ve become more and more convinced that (regular) porn availability in fact reduces sexual assault.

        I don’t see an obvious reason why it should be different in case of CP, now that it can be generated.

        • Lowlee Kun@feddit.de
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 months ago

          It should be different because people can not have it. It is disgusting, makes them feel icky and thats just why it has to be bad. Conventional wisdom sometimes really is just convential idiocracy.

    • Catoblepas@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      35
      arrow-down
      7
      ·
      4 months ago

      Did we memory hole the whole ‘known CSAM in training data’ thing that happened a while back? When you’re vacuuming up the internet you’re going to wind up with the nasty stuff, too. Even if it’s not a pixel by pixel match of the photo it was trained on, there’s a non-zero chance that what it’s generating is based off actual CSAM. Which is really just laundering CSAM.

      • Ragdoll X@lemmy.world
        link
        fedilink
        English
        arrow-up
        32
        arrow-down
        2
        ·
        4 months ago

        IIRC it was something like a fraction of a fraction of 1% that was CSAM, with the researchers identifying the images through their hashes but they weren’t actually available in the dataset because they had already been removed from the internet.

        Still, you could make AI CSAM even if you were 100% sure that none of the training images included it since that’s what these models are made for - being able to combine concepts without needing to have seen them before. If you hold the AI’s hand enough with prompt engineering, textual inversion and img2img you can get it to generate pretty much anything. That’s the power and danger of these things.

        • Catoblepas@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          5
          ·
          4 months ago

          What % do you think was used to generate the CSAM, though? Like, if 1% of the images were cups it’s probably drawing on some of that to generate images of cups.

          And yes, you could technically do this with no CSAM training material, but we don’t know if that’s what the AI is doing because the image sources used to train it were mass scraped from the internet. They’re using massive amounts of data without filtering it and are unable to say with certainty whether or not there is CSAM in the training material.

    • PM_Your_Nudes_Please@lemmy.world
      link
      fedilink
      English
      arrow-up
      17
      arrow-down
      2
      ·
      4 months ago

      Yeah, it’s very similar to the “is loli porn unethical” debate. No victim, it could supposedly help reduce actual CSAM consumption, etc… But it’s icky so many people still think it should be illegal.

      There are two big differences between AI and loli though. The first is that AI would supposedly be trained with CSAM to be able to generate it. An artist can create loli porn without actually using CSAM references. The second difference is that AI is much much easier for the layman to create. It doesn’t take years of practice to be able to create passable porn. Anyone with a decent GPU can spin up a local instance, and be generating within a few hours.

      In my mind, the former difference is much more impactful than the latter. AI becoming easier to access is likely inevitable, so combatting it now is likely only delaying the inevitable. But if that AI is trained on CSAM, it is inherently unethical to use.

      Whether that makes the porn generated by it unethical by extension is still difficult to decide though, because if artists hate AI, then CSAM producers likely do too. Artists are worried AI will put them out of business, but then couldn’t the same be said about CSAM producers? If AI has the potential to run CSAM producers out of business, then it would be a net positive in the long term, even if the images being created in the short term are unethical.

    • BrianTheeBiscuiteer@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      3
      ·
      4 months ago

      I have trouble with this because it’s like 90% grey area. Is it a pic of a real child but inpainted to be nude? Was it a real pic but the face was altered as well? Was it completely generated but from a model trained on CSAM? Is the perceived age of the subject near to adulthood? What if the styling makes it only near realistic (like very high quality CG)?

      I agree with what the FBI did here mainly because there could be real pictures among the fake ones. However, I feel like the first successful prosecution of this kind of stuff will be a purely moral judgement of whether or not the material “feels” wrong, and that’s no way to handle criminal misdeeds.

      • Chee_Koala@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        ·
        4 months ago

        If not trained on CSAM or in painted but fully generated, I can’t really think of any other real legal arguments against it except for: “this could be real”. Which has real merit, but in my eyes not enough to prosecute as if it were real. Real CSAM has very different victims and abuse so it needs different sentencing.

      • Zorque@kbin.social
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        4 months ago

        Everything is 99% grey area. If someone tells you something is completely black and white you should be suspicious of their motives.

      • Zorque@kbin.social
        link
        fedilink
        arrow-up
        1
        ·
        4 months ago

        Is everything completely black and white for you?

        The system isn’t perfect, especially where we prioritize punishing people over rehabilitation. Would you rather punish everyone equally, emphasizing that if people are going to risk the legal implications (which, based on legal systems the world over, people are going to do) they might as well just go for the real thing anyways?

        You don’t have to accept it as morally acceptable, but you don’t have to treat them as completely equivalent either.

        There’s gradations of questionable activity. Especially when there’s no real victims involved. Treating everything exactly the same is, frankly speaking, insane. Its like having one punishment for all illegal behavior. Murder someone? Death penalty. Rob them? Straight to the electric chair. Jaywalking? Better believe you’re getting the needle.

        • Mastengwe@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          edit-2
          4 months ago

          Ironically, You ask if everything is completely black and white for someone without accepting that there’s nuance to the very issue you’re calling out. And assuming that “everything”- a very black and white term, is not very nuanced, is it?

          No, not EVERYTHING, but some things. And this is one of those things. Both forms should be illegal. Period. No nuance, no argument, NO grey area.

          This does not mean that nuance doesn’t exist. It just means that some believe that it SHOULDN’T exist within the paradigm of child porn.

    • PhlubbaDubba@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      4 months ago

      I think the point is that child attraction itself is a mental illness and people indulging it even without actual child contact need to be put into serious psychiatric evaluation and treatment.

    • Murvel@lemm.ee
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      11
      ·
      4 months ago

      It feeds and evolves a disorder which in turn increases risks of real life abuse.

      But if AI generated content is to be considered illegal, so should all fictional content.

      • SigHunter@lemmy.kde.social
        link
        fedilink
        English
        arrow-up
        27
        arrow-down
        6
        ·
        edit-2
        4 months ago

        Or, more likely, it feeds and satisfies a disorder which in turn decreases risk of real life abuse.

        Making it illegal so far helped nothing, just like with drugs

        • Murvel@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          4
          ·
          4 months ago

          That’s not how these addictive disorders works… they’re never satisfied and always need more.

      • Norgur@kbin.social
        link
        fedilink
        arrow-up
        3
        arrow-down
        1
        ·
        4 months ago

        Two things:

        1. Do we know if fuels the urge to get real children? Or do we just assume that through repetition like the myth of “gateway drugs”?
        2. Since no child was involved and harmed in the making of these images… On what grounds could it be forbidden to generate them?
        • Thorny_Insight@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          edit-2
          4 months ago

          Alternative perspective is to think that does watching normal porn make heterosexual men more likely to rape women? If not then why would it be different in this case?

          The vast majority of pedophiles never offend. Most people in jail for child abuse are just plain old rapists with no special interest towards minors, they’re just an easy target. Pedophilia just describes what they’re attracted to. It’s not a synonym to child rapist. It usually needs to coinside with psychopathy to create the monster that most people think about when hearing that word.

          • ricecake@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 months ago

            That’s a bit of a difference in comparison.
            A better comparison would be “does watching common heterosexual porn make common heterosexual men more interested in performing common heterosexual sexual acts?” or "does viewing pornography long term satiate a mans sex drive?” or “does consumption of nonconsensual pornography correlate to an increase in nonconsensual sex acts?”

            Comparing “viewing child sexual content might lead it engaging in sexual acts with children” to “viewing sexual activity with women might lead to rape” is disingenuous and apples to oranges.

            https://wchh.onlinelibrary.wiley.com/doi/full/10.1002/tre.791

            a review of 19 studies published between 2013 and 2018 found an association between online porn use and earlier sexual debut, engaging with occasional and/or multiple partners, emulating risky sexual behaviours, assimilating distorted gender roles, dysfunctional body perception, aggression, anxiety, depression, and compulsive porn use.24 Another study has shown that compulsive use of sexually explicit internet material by adolescent boys is more likely in those with lower self-esteem, depressive feeling and excessive sexual interest.1

            some porn use in adult men may have a positive impact by increasing libido and desire for a real-life partner, relieving sexual boredom, and improving sexual satisfaction by providing inspiration for real sex.7

            As for child porn, it’s not a given that there’s no relationship between consumption and abusing children. There are studies that indicate both outcomes, and are made much more complicated by one of both activities being extremely illegal and socially stigmatized making accurate tracking difficult.
            It’s difficult to justify the notion that “most pedophiles never offend” when it can be difficult to identify both pedophiles and abuse.

            https://pubmed.ncbi.nlm.nih.gov/21088873/ for example. It looks at people arrested for possession of child pornography. Within six years, 6% were charged with a child contact crime. Likewise, you can find research with a differing conclusion

            Point being, you can’t just hand wave the potential for a link away on the grounds that porn doesn’t cause rape amongst typical heterosexual men. There’s too many factors making the statistics difficult to gather.

  • helpImTrappedOnline@lemmy.world
    link
    fedilink
    English
    arrow-up
    105
    arrow-down
    1
    ·
    edit-2
    4 months ago

    The headline/title needs to be extended to include the rest of the sentence

    “and then sent them to a minor”

    Yes, this sicko needs to be punished. Any attempt to make him the victim of " the big bad government" is manipulative at best.

    Edit: made the quote bigger for better visibility.

    • Ricky Rigatoni@lemm.ee
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      5
      ·
      4 months ago

      You can get away with a lot of heinous crimes by simply not telling people and sharing the results.

      • quindraco@lemm.ee
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        6
        ·
        4 months ago

        You consider it a heinous crime to draw a picture and keep it to yourself?

    • Frozengyro@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      12
      ·
      4 months ago

      That’s sickening to know there are bastards out there who will get away with it since they are only creating it.

      • NeoNachtwaechter@lemmy.world
        link
        fedilink
        English
        arrow-up
        15
        arrow-down
        5
        ·
        4 months ago

        I’m not sure. Let us assume that you generate it on your own PC at home (not using a public service) and don’t brag about it and never give it to anybody - what harm is done?

        • GBU_28@lemm.ee
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          3
          ·
          edit-2
          4 months ago

          Society is not ok with the idea of someone cranking to CSAM, then just walking around town. It gives people wolf-in-sheep-clothing vibes.

          So the notion of there being “ok” CSAM-style ai content is a non starter for a huge fraction of people because it still suggests appeasing a predator.

          I’m definitely one of those people that simply can’t accept any version of it.

        • Frozengyro@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          13
          ·
          4 months ago

          Even if the AI didn’t train itself on actual CSAM that is something that feels inherently wrong. Your mind is not right to think that’s acceptable IMO.

          • DarkThoughts@fedia.io
            link
            fedilink
            arrow-up
            7
            arrow-down
            2
            ·
            4 months ago

            Laws shouldn’t be about feelings though and we shouldn’t prosecute people for victimless thought crimes. How often did you think something violent when someone really pissed you off? Should you have been persecuted for that thought too?

              • DarkThoughts@fedia.io
                link
                fedilink
                arrow-up
                7
                arrow-down
                2
                ·
                4 months ago

                Who are the victims of someone generating such images privately then? It’s on the same level as all the various fan fiction shit that was created manually over all the past decades.

                And do we apply this to other depictions of criminalized things too? Would we ban the depiction of violence & sexual violence on TV, in books, and in video games too?

  • peanuts4life@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    52
    ·
    4 months ago

    It’s worth mentioning that in this instance the guy did send porn to a minor. This isn’t exactly a cut and dry, “guy used stable diffusion wrong” case. He was distributing it and grooming a kid.

    The major concern to me, is that there isn’t really any guidance from the FBI on what you can and can’t do, which may lead to some big issues.

    For example, websites like novelai make a business out of providing pornographic, anime-style image generation. The models they use deliberately tuned to provide abstract, “artistic” styles, but they can generate semi realistic images.

    Now, let’s say a criminal group uses novelai to produce CSAM of real people via the inpainting tools. Let’s say the FBI cast a wide net and begins surveillance of novelai’s userbase.

    Is every person who goes on there and types, “Loli” or “Anya from spy x family, realistic, NSFW” (that’s an underaged character) going to get a letter in the mail from the FBI? I feel like it’s within the realm of possibility. What about “teen girls gone wild, NSFW?” Or “young man, no facial body hair, naked, NSFW?”

    This is NOT a good scenario, imo. The systems used to produce harmful images being the same systems used to produce benign or borderline images. It’s a dangerous mix, and throws the whole enterprise into question.

    • ricecake@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      4 months ago

      The major concern to me, is that there isn’t really any guidance from the FBI on what you can and can’t do, which may lead to some big issues.

      https://www.ic3.gov/Media/Y2024/PSA240329 https://www.justice.gov/criminal/criminal-ceos/citizens-guide-us-federal-law-child-pornography

      They’ve actually issued warnings and guidance, and the law itself is pretty concise regarding what’s allowed.

      (8) “child pornography” means any visual depiction, including any photograph, film, video, picture, or computer or computer-generated image or picture, whether made or produced by electronic, mechanical, or other means, of sexually explicit conduct, where-

      (A) the production of such visual depiction involves the use of a minor engaging in sexually explicit conduct;

      (B) such visual depiction is a digital image, computer image, or computer-generated image that is, or is indistinguishable from, that of a minor engaging in sexually explicit conduct; or

      © such visual depiction has been created, adapted, or modified to appear that an identifiable minor is engaging in sexually explicit conduct.

      (11) the term “indistinguishable” used with respect to a depiction, means virtually indistinguishable, in that the depiction is such that an ordinary person viewing the depiction would conclude that the depiction is of an actual minor engaged in sexually explicit conduct. This definition does not apply to depictions that are drawings, cartoons, sculptures, or paintings depicting minors or adults.

      https://uscode.house.gov/view.xhtml?hl=false&edition=prelim&req=granuleid%3AUSC-prelim-title18-section2256&f=treesort&num=0

      If you’re going to be doing grey area things you should do more than the five minutes of searching I did to find those honestly.

      It was basically born out of a supreme Court case in the early 2000s regarding an earlier version of the law that went much further and banned anything that “appeared to be” or “was presented as” sexual content involving minors, regardless of context, and could have plausibly been used against young looking adult models, artistically significant paintings, or things like Romeo and Juliet, which are neither explicit nor vulgar but could be presented as involving child sexual activity. (Juliet’s 14 and it’s clearly labeled as a love story).
      After the relevant provisions were struck down, a new law was passed that factored in the justices rationale and commentary about what would be acceptable and gave us our current system of “it has to have some redeeming value, or not involve actual children and plausibly not look like it involves actual children”.

    • PirateJesus@lemmy.today
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      6
      ·
      4 months ago

      The major concern to me, is that there isn’t really any guidance from the FBI on what you can and can’t do, which may lead to some big issues.

      The Protect Act of 2003 means that any artistic depiction of CSAM is illegal. The guidance is pretty clear, FBI is gonna raid your house…eventually. We still haven’t properly funded the anti-CSAM departments.

    • retrieval4558@mander.xyz
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      21
      ·
      4 months ago

      Is every person who goes on there and types, “Loli” or “Anya from spy x family, realistic, NSFW” (that’s an underaged character) going to get a letter in the mail from the FBI?

      I’ll throw that baby out with the bathwater to be honest.

      • Duamerthrax@lemmy.world
        link
        fedilink
        English
        arrow-up
        25
        arrow-down
        4
        ·
        edit-2
        4 months ago

        Simulated crimes aren’t crimes. Would you arrest every couple that finds health ways to simulate rape fetishes? Would you arrest every person who watches Fast and The Furious or The Godfather?

        If no one is being hurt, if no real CSAM is being fed into the model, if no pornographic images are being sent to minors, it shouldn’t be a crime. Just because it makes you uncomfortable, don’t make it immoral.

        • Ookami38@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          18
          ·
          4 months ago

          Or, ya know, everyone who ever wanted to decapitate those stupid fucking Skyrim children. Crime requires damaged parties, and with this (idealized case, not the specific one in the article) there is none.

          • DarkThoughts@fedia.io
            link
            fedilink
            arrow-up
            9
            arrow-down
            1
            ·
            4 months ago

            Those were demon children from hell (with like 2 exceptions maybe). It was a crime by Bethesda to make them invulnerable / protected by default.

        • helpImTrappedOnline@lemmy.world
          link
          fedilink
          English
          arrow-up
          13
          ·
          4 months ago

          Simulated crimes aren’t crimes.

          If they were, any one who’s played games is fucked. I’m confident everyone who has played went on a total ramapage murdering the townfolk, pillaging their houses and blowing everything up…in Minecraft.

        • PirateJesus@lemmy.today
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          7
          ·
          4 months ago

          Simulated crimes aren’t crimes.

          Artistic CSAM is definitely a crime in the United States. PROTECT act of 2003.

          • Duamerthrax@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            ·
            4 months ago

            People have only gotten in trouble for that when they’re already in trouble for real CSAM. I’m not terrible interested in sticking up for actual CSAM scum.

            • PotatoKat@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 months ago

              Real children are in training data regardless of if there is csam in the data or not (which there is a high chance there is considering how they get their training data) so real children are involved

            • Meansalladknifehands@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              3
              ·
              4 months ago

              For now, if you read the article, it states that he shared the pictures to form like minded groups where they got emboldened and could support each other and legitimize/normalize their perverted thoughts. How about no thanks.

              • Lowlee Kun@feddit.de
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                4 months ago

                wrong comment chain. people weren’t talking about the criminal shithead the article is about but about the scenario of someone using (not csam trained) models to create questionable content (thus it is implied that there would be no victim). we all know that there are bad actors out there, just like there are rapists and murderers. still we dont condemn true crime lovers or rape fetishists until they commit a crime. we could do the same with pedos but somehow we believe hating them into the shadows will stop them somehow from doing criminal stuff?

              • Duamerthrax@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                4 months ago

                Maybe you should focus your energy on normalized things that actually effect kids like banning full contact sports that cause CTE.

  • horncorn@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    42
    arrow-down
    6
    ·
    4 months ago

    Article title is a bit misleading. Just glancing through I see he texted at least one minor in regards to this and distributed those generated pics in a few places. Putting it all together, yeah, arrest is kind of a no-brainer. Ethics of generating csam is the same as drawing it pretty much. Not much we can do about it aside from education.

    • ricecake@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      4 months ago

      Legally, a sufficiently detailed image depicting csam is csam, regardless of how it was produced. Sharing it is why he got caught, inevitably, but it’s still illegal even if he never brought a minor into it.

    • retrospectology@lemmy.world
      link
      fedilink
      English
      arrow-up
      14
      arrow-down
      32
      ·
      edit-2
      4 months ago

      Lemmy really needs to stop justifying CP. We can absolutely do more than “eDuCaTiOn”. AI is created by humans, the training data is gathered by humans, it needs regulation like any other industry.

      It’s absolutely insane to me how laissez-fair some people are about AI, it’s like a cult.

      • msage@programming.dev
        link
        fedilink
        English
        arrow-up
        23
        arrow-down
        4
        ·
        4 months ago

        While I agree with your attitude, the whole ‘laissez-fair’ thing is probably a misunderstanding:

        There is nothing we can do to stop the AI.

        Nothing.

        The genie is out of the bottle, the Pandora’s box has been opened, everything is out and it won’t ever return. The world will never be the same, and it’s irrelevant what people think.

        That’s why we need to better understand the post-AI world we created, and figure out what do to now.

        Also, to hell with CP. (feels weird to use the word ‘fuck’ here)

        • retrospectology@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          17
          ·
          edit-2
          4 months ago

          Thats not the question, the question is not “can we stop AI entirely” it’s about regulating its development and yes, we can make efforts to do that.

          This attitude of “it’s inevitable, can’t do anything about it” is eerily similar logic to what is used in climate denial and other right-wing efforts. It’s a really poor attitude to have, especially about something as consequential as AI.

          We have the best opportunity right now to create rules about its uses and development. The answer is not “do nothing” as if it’s some force of nature, as opposed toa tool created by humans.

          • msage@programming.dev
            link
            fedilink
            English
            arrow-up
            7
            arrow-down
            1
            ·
            4 months ago

            I hear you, and I don’t necessarily disagree with you, I just know that’s not how anything works.

            Regulations work for big companies, but there isn’t a big company behind this specific case. And those small-time users have run away and you can’t stop them.

            It’s like trying to regulate cameras to not store specific images. Like, I get the sentiment, but sorry, no. It’s not that I would not like that, it’s just not possible.

            • retrospectology@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              7
              ·
              4 months ago

              This argument could be applied to anything though. A lot of people get away with myrder, we should still try and do what we can to stop it from happening.

              You can’t sit in every car and force people to wear a seatbelt, we still have seatbelt laws and regulations for manufacturers.

              • msage@programming.dev
                link
                fedilink
                English
                arrow-up
                5
                arrow-down
                1
                ·
                4 months ago

                Physical things are much easier to regulate than software, much less serverless.

                We already regulate certain images, and it matters very little.

                The bigger payoff will be from educating the public and accepting that we can’t win every war.

                • retrospectology@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  7
                  ·
                  edit-2
                  4 months ago

                  So accept defeat from the start, that’s really just a non-starter. AI models run on hardware, they are developed by specific people, their contents are distributed by specific individuals, code bases are hosted on hardware and on specific outlets.

                  It really does sound like you’re just trying to make excuses to avoid regulation, not that you genuinely have a good reason to think it’s not possible to try.

          • GBU_28@lemm.ee
            link
            fedilink
            English
            arrow-up
            3
            ·
            edit-2
            4 months ago

            Dude the amount of open source, untrackable, distributed ai models is off the charts. This isn’t just about the models offered by subscription from the big players.

            • retrospectology@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              4 months ago

              This is still one of the weaker arguments. There is a lot of malware out there too, people are still prosecuted when they’re caught developing and distributing it, we don’t just throw up our hands and pretend there’s nothing that can be done.

              Like, yeah, some pedophile who also happens to be tech saavy might build his own AI model to make CP, that’s not some self-evident argument against attempting to stop them.

              • GBU_28@lemm.ee
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                edit-2
                4 months ago

                No, like, the tools to do these things are common and readily available. It’s not malware, it’s generalized ai tools, completely embroiled with non image ai work.

                Pandora’s box is wide open. All of this work can be done trivially, completely offline with a basic PC. Anyone motivated can be offline and up and running in a weekend

                You’re asking to outlaw something like a spreadsheet.

                You download a general purpose image ai model, then train and prompt it completely offline

          • L_Acacia@lemmy.one
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            4 months ago

            The models used are not trained on CP. The models weight are distributed freely and anybody can train a LORA on his computer. Its already too late to ban open weight models.

      • DarkThoughts@fedia.io
        link
        fedilink
        arrow-up
        2
        arrow-down
        1
        ·
        4 months ago

        You don’t need CSAM training data to create CSAM images. If your model knows how children looks like, how naked human bodies look like, then it can create naked children. That’s simply how generative models like this work and has absolutely nothing to do with specifically trained models for CSAM using actual CSAM material.

        So while I disagree with him, in that lack of education is the cause of CSAM or pedophilia… I’d say it could help with the general hysteria about LLMs, like the one’s coming from you, who just let their emotions run wild when those topics arise. You people need to understand that the goal should be the protection of potential victims, not the punishment of victimless thought crimes.

  • UnpluggedFridge@lemmy.world
    link
    fedilink
    English
    arrow-up
    35
    ·
    4 months ago

    These cases are interesting tests of our first amendment rights. “Real” CP requires abuse of a minor, and I think we can all agree that it should be illegal. But it gets pretty messy when we are talking about depictions of abuse.

    Currently, we do not outlaw written depictions nor drawings of child sexual abuse. In my opinion, we do not ban these things partly because they are obvious fictions. But also I think we recognize that we should not be in the business of criminalizing expression, regardless of how disgusting it is. I can imagine instances where these fictional depictions could be used in a way that is criminal, such as using them to blackmail someone. In the absence of any harm, it is difficult to justify criminalizing fictional depictions of child abuse.

    So how are AI-generated depictions different? First, they are not obvious fictions. Is this enough to cross the line into criminal behavior? I think reasonable minds could disagree. Second, is there harm from these depictions? If the AI models were trained on abusive content, then yes there is harm directly tied to the generation of these images. But what if the training data did not include any abusive content, and these images really are purely depictions of imagination? Then the discussion of harms becomes pretty vague and indirect. Will these images embolden child abusers or increase demand for “real” images of abuse. Is that enough to criminalize them, or should they be treated like other fictional depictions?

    We will have some very interesting case law around AI generated content and the limits of free speech. One could argue that the AI is not a person and has no right of free speech, so any content generated by AI could be regulated in any manner. But this argument fails to acknowledge that AI is a tool for expression, similar to pen and paper.

    A big problem with AI content is that we have become accustomed to viewing photos and videos as trusted forms of truth. As we re-learn what forms of media can be trusted as “real,” we will likely change our opinions about fringe forms of AI-generated content and where it is appropriate to regulate them.

    • Corkyskog@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      7
      ·
      4 months ago

      It comes back to distribution for me. If they are generating the stuff for themselves, gross, but I don’t see how it can really be illegal. But if your distributing them, how do we know their not real? The amount of investigative resources that would need to be dumped into that, and the impact on those investigators mental health, I don’t know. I really don’t have an answer, I don’t know how they make it illegal, but it really feels like distribution should be.

    • yamanii@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      4 months ago

      partly because they are obvious fictions

      That’s it actually, all sites that allow it like danbooru, gelbooru, pixiv, etc. Have a clause against photo realistic content and they will remove it.

    • nucleative@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 months ago

      Well thought-out and articulated opinion, thanks for sharing.

      If even the most skilled hyper-realistic painters were out there painting depictions of CSAM, we’d probably still label it as free speech because we “know” it to be fiction.

      When a computer rolls the dice against a model and imagines a novel composition of children’s images combined with what it knows about adult material, it does seem more difficult to label it as entirely fictional. That may be partly because the source material may have actually been real, even if the final composition is imagined. I don’t intend to suggest models trained on CSAM either, I’m thinking of models trained to know what both mature and immature body shapes look like, as well as adult content, and letting the algorithm figure out the rest.

      Nevertheless, as you brought up, nobody is harmed in this scenario, even though many people in our culture and society find this behavior and content to be repulsive.

      To a high degree, I think we can still label an individual who consumes this type of AI content to be a pedophile, and although being a pedophile is not in and of itself an illegal adjective to posses, it comes with societal consequences. Additionally, pedophilia is a DSM-5 psychiatric disorder, which could be a pathway to some sort of consequences for those who partake.

    • TheHarpyEagle@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      4 months ago

      It feels incredibly gross to just say “generated CSAM is a-ok, grab your hog and go nuts”, but I can’t really say that it should be illegal if no child was harmed in the training of the model. The idea that it could be a gateway to real abuse comes to mind, but that’s a slippery slope that leads to “video games cause school shootings” type of logic.

      I don’t know, it’s a very tough thing to untangle. I guess I’d just want to know if someone was doing that so I could stay far, far away from them.

      • KillingTimeItself@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        5
        ·
        4 months ago

        for some reason the US seems to hold a weird position on this one. I don’t really understand it.

        It’s written to be illegal, but if you look at prosecution cases, i think there have been only a handful of charged cases. The prominent ones which also include relevant previous offenses, or worse.

        It’s also interesting when you consider that there are almost definitely large image boards hosted in the US that host what could be constituted as “cartoon CSAM” notably e621, i’d have to verify their hosting location, but i believe they’re in the US. And so far i don’t believe they’ve ever had any issues with it. And i’m sure there are other good examples as well.

        I suppose you could argue they’re exempt on the publisher rules. But these sites don’t moderate against these images, generally. And i feel like this would be the rare exception where it wouldnt be applicable.

        The law is fucking weird dude. There is a massive disconnect between what we should be seeing, and what we are seeing. I assume because the authorities who moderate this shit almost exclusively go after real CSAM, on account of it actually being a literal offense, as opposed to drawn CSAM, being a proxy offense.

  • PirateJesus@lemmy.today
    link
    fedilink
    English
    arrow-up
    38
    arrow-down
    5
    ·
    4 months ago

    OMG. Every other post is saying their disgusted about the images part but it’s a grey area, but he’s definitely in trouble for contacting a minor.

    Cartoon CSAM is illegal in the United States. AI images of CSAM fall into that category. It was illegal for him to make the images in the first place BEFORE he started sending them to a minor.

    https://www.thefederalcriminalattorneys.com/possession-of-lolicon

    https://en.wikipedia.org/wiki/PROTECT_Act_of_2003

    • Madison420@lemmy.world
      link
      fedilink
      English
      arrow-up
      19
      arrow-down
      4
      ·
      4 months ago

      Yeah that’s toothless. They decided there is no particular way to age a cartoon, they could be from another planet that simply seem younger but are in actuality older.

      It’s bunk, let them draw or generate whatever they want, totally fictional events and people are fair game and quite honestly I’d Rather they stay active doing that then get active actually abusing children.

      Outlaw shibari and I guarantee you’d have multiple serial killers btk-ing some unlucky souls.

        • RGB3x3@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          7
          ·
          edit-2
          4 months ago

          The problem with AI CSAM generation is that the AI has to be trained on something first. It has to somehow know what a naked minor looks like. And to do that, well… You need to feed it CSAM.

          So is it right to be using images of real children to train these AI? You’d be hard-pressed to find someone who thinks that’s okay.

          • Eezyville@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            11
            ·
            4 months ago

            You make the assumption that the person generating the images also trained the AI model. You also make assumptions about how the AI was trained without knowing anything about the model.

            • RGB3x3@lemmy.world
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              2
              ·
              edit-2
              4 months ago

              Are there any guarantees that harmful images weren’t used in these AI models? Based on how image generation works now, it’s very likely that harmful images were used to train the data.

              And if a person is using a model based on harmful training data, they should be held responsible.

              However, the AI owner/trainer has even more responsibility in perpetuating harm to children and should be prosecuted appropriately.

              • Eezyville@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                10
                arrow-down
                3
                ·
                4 months ago

                And if a person is using a model based on harmful training data, they should be held responsible.

                I will have to disagree with you for several reasons.

                • You are still making assumptions about a system you know absolutely nothing about.
                • By your logic anything born from something that caused suffering from others (this example is AI trained on CSAM) the users of that product should be held responsible for the crime committed to create that product.
                  • Does that apply to every product/result created from human suffering or just the things you don’t like?
                  • Will you apply that logic to the prosperity of Western Nations built on the suffering of indigenous and enslaved people? Should everyone who benefit from western prosperity be held responsible for the crimes committed against those people?
                  • What about medicine? Two examples are The Tuskegee Syphilis Study and the cancer cells of Henrietta Lacks. Medicine benefited greatly from these two examples but crimes were committed against the people involved. Should every patient from a cancer program that benefited from Ms. Lacks’ cancer cells also be subject to pay compensation to her family? The doctors that used her cells without permission didn’t.
                  • Should we also talk about the advances in medicine found by Nazis who experimented on Jews and others during WW2? We used that data in our manned space program paving the way to all the benefits we get from space technology.
                • PotatoKat@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  4 months ago

                  The difference between the things you’re listing and SAM is that those other things have actual utility outside of getting off. Were our phones made with human suffering? Probably but phones have many more uses than making someone cum. Are all those things wrong? Yea, but at least good came out of it outside of just giving people sexual gratification directly from the harm of others.

                • aceshigh@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  7
                  ·
                  4 months ago

                  The topic that you’re choosing to focus on really interesting. what are your values?

              • aesthelete@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                ·
                edit-2
                4 months ago

                Are there any guarantees that harmful images weren’t used in these AI models?

                Lol, highly doubt it. These AI assholes pretend that all the training data randomly fell into the model (off the back of a truck) and that they cannot possibly be held responsible for that or know anything about it because they were too busy innovating.

                There’s no guarantee that most regular porn sites don’t contain csam or other exploitative imagery and video (sex trafficking victims). There’s absolutely zero chance that there’s any kind of guarantee.

            • PotatoKat@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 months ago

              The images were created using photos of real children even if said photos weren’t CSAM (which can’t be guaranteed they weren’t). So the victims were are the children used to generate CSAM

              • sugar_in_your_tea@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                0
                ·
                edit-2
                4 months ago

                Let’s do a thought experiment, and I’d look to to tell me at what point a victim was introduced:

                1. I legally acquire pictures of a child, fully clothed and everything
                2. I draw a picture based on those legal pictures, but the subject is nude or doing sexually explicit things
                3. I keep the picture for my own personal use and don’t distribute it

                Or with AI:

                1. I legally acquire pictures of children, fully clothed and everything
                2. I legally acquire pictures of nude adults, some doing sexually explicit things
                3. I train an AI on a mix of 1&2
                4. I generate images of nude children, some of them doing sexually explicit things
                5. I keep the pictures for my own personal use and don’t distribute any of them
                6. I distribute my model, using the right to distribute from the legal acquisition of those images

                At what point did my actions victimize someone?

                If I distributed those images and those images resemble a real person, then that real person is potentially a victim.

                I will say someone who does this creepy and I don’t want them anywhere near children (especially mine, and yes, I have kids), but I don’t think it should be illegal, provided the source material is legal. But as soon as I distribute it, there absolutely could be a victim. Being creepy shouldn’t be a crime.

                • PotatoKat@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  4 months ago

                  I think it should be illegal to make porn of a person without their permission regardless of if it was shared or not. Imagine the person it is based off of finds out someone is doing that. That causes mental strain on the person. Just like how revenge porn doesn’t actively harm a person but causes mental strafe (both the initial upload and continued use of it). For scenario 1 it would be at step 2 when the porn is made of the person. For scenario 2 it would be a mix between step 3 and 4.

          • deathbird@mander.xyz
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            2
            ·
            4 months ago

            the AI has to be trained on something first. It has to somehow know what a naked minor looks like. And to do that, well… You need to feed it CSAM.

            First of all, not every image of a naked child is CSAM. This is actually been kind of a problem with automated CSAM detection systems triggering false positives on non-sexual images, and getting innocent people into trouble.

            But also, AI systems can blend multiple elements together. They don’t need CSAM training material to create CSAM, just the individual elements crafted into a prompt sufficient to create the image while avoiding any safeguards.

            • PotatoKat@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 months ago

              You ignored the second part of their post. Even if it didn’t use any csam is it right to use pictures of real children to generate csam? I really don’t think it is.

      • Mike@lemmy.ml
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        3
        ·
        4 months ago

        I think the challenge with Generative AI CSAM is the question of where did training data originate? There has to be some questionable data there.

        • scoobford@lemmy.zip
          link
          fedilink
          English
          arrow-up
          10
          arrow-down
          1
          ·
          4 months ago

          That would mean you need to enforce the law for whoever built the model. If the original creator has 100TB of cheese pizza, then they should be the one who gets arrested.

          Otherwise you’re busting random customers at a pizza shop for possession of the meth the cook smoked before his shift.

        • erwan@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          There is also the issue of determining if a given image is real or AI. If AI were legal, that means prosecution would need to prove images are real and not AI with the risk of letting go real offenders.

          The need to ban AI CSAM is even clearer than cartoon CSAM.

      • zbyte64@awful.systems
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        12
        ·
        4 months ago

        Big brain PDF tells the judge it is okay because the person in the picture is now an adult.

        • arefx@lemmy.ml
          link
          fedilink
          English
          arrow-up
          14
          arrow-down
          1
          ·
          edit-2
          4 months ago

          You can say pedophile… that “pdf file” stuff is so corny and childish. Hey guys lets talk about a serious topic by calling it things like “pdf files” and “Graping”. Jfc

          • RGB3x3@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            4 months ago

            Why do people say “graping?” I’ve never heard that.

            Please tell me it doesn’t have to do with “The Grapist” video that came out on early YouTube.

        • surewhynotlem@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          4 months ago

          That’s the issue though. As far as I know it hasn’t been tested in court and it’s quite possible the law is useless and has no teeth.

          With AI porn you can point to real victims whose unconsented pictures were used to train the models, and say that’s abuse. But when it’s just a drawing, who is the victim? Is it just a thought crime? Can we prosecute those?

    • Clbull@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      4 months ago

      I thought cartoons/illustrations of that nature were only illegal in the UK (Coroners and Justices Act 2008) and Switzerland. TIL about the PROTECT Act.

  • SeattleRain@lemmy.world
    link
    fedilink
    English
    arrow-up
    33
    ·
    edit-2
    4 months ago

    America has some of the most militant anti pedophilic culture in the world but they far and away have the highest rates of child sexual assault.

    I think AI is going to revel is how deeply hypocritical Americans are on this issue. You have gigantic institutions like churches committing industrial scale victimization yet you won’t find a 1/10th of the righteous indignation against other organized religions where there is just as much evidence it is happening as you will regarding one person producing images that don’t actually hurt anyone.

    It’s pretty clear by how staggering a rate of child abuse that occurs in the states that Americans are just using child victims as weaponized politicalization (it’s next to impossible to convincingly fight off pedo accusations if you’re being mobbed) and aren’t actually interested in fighting pedophilia.

    • kandoh@reddthat.com
      link
      fedilink
      English
      arrow-up
      19
      ·
      4 months ago

      Most states will let grown men marry children as young as 14. There is a special carve out for Christian pedophiles.

      • ricecake@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        4 months ago

        Fortunately most instances are in the category of a 17 year old to an 18 year old, and require parental consent and some manner of judicial approval, but the rates of “not that” are still much higher than one would want.
        ~300k in a 20 year window total, 74% of the older partner being 20 or younger, and 95% of the younger partner being 16 or 17, with only 14% accounting for both partners being under 18.

        There’s still no reason for it in any case, and I’m glad to live in one of the states that said "nah, never needed .

  • Greg Clarke@lemmy.ca
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    2
    ·
    4 months ago

    This is tough, the goal should be to reduce child abuse. It’s unknown if AI generated CP will increase or reduce child abuse. It will likely encourage some individuals to abuse actual children while for others it may satisfy their urges so they don’t abuse children. Like everything else AI, we won’t know the real impact for many years.

        • LadyAutumn@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          12
          ·
          4 months ago

          It should be illegal either way, to be clear. But you think theyre not training models on CSAM? Youre trusting in the morality/ethics of people creating AI generated child pornography?

          • Greg Clarke@lemmy.ca
            link
            fedilink
            English
            arrow-up
            7
            ·
            4 months ago

            The use of CSAM in training generative AI models is an issue no matter how these models are being used.

            • L_Acacia@lemmy.one
              link
              fedilink
              English
              arrow-up
              4
              arrow-down
              1
              ·
              4 months ago

              The training doesn’t use csam, 0% chance big tech would use that in their dataset. The models are somewhat able to link concept like red and car, even if it had never seen a red car before.

              • AdrianTheFrog@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                ·
                4 months ago

                Well, with models like SD at least, the datasets are large enough and the employees are few enough that it is impossible to have a human filter every image. They scrape them from the web and try to filter with AI, but there is still a chance of bad images getting through. This is why most companies install filters after the model as well as in the training process.

                • DarkThoughts@fedia.io
                  link
                  fedilink
                  arrow-up
                  4
                  ·
                  4 months ago

                  You make it sound like it is so easy to even find such content on the www. The point is, they do not need to be trained on such material. They are trained on regular kids, so they know their sizes, faces, etc. They’re trained on nude bodies, so they also know how hairless genitals or flat chests look like. You don’t need to specifically train a model on nude children to generate nude children.

      • DarkThoughts@fedia.io
        link
        fedilink
        arrow-up
        7
        ·
        edit-2
        4 months ago

        I suggest you actually download stable diffusion and try for yourself because it’s clear that you don’t have any clue what you’re talking about. You can already make tiny people, shaved, genitals, flat chests, child like faces, etc. etc. It’s all already there. Literally no need for any LoRAs or very specifically trained models.

  • Nora@lemmy.ml
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    1
    ·
    edit-2
    4 months ago

    I had an idea when these first AI image generators started gaining traction. Flood the CSAM market with AI generated images( good enough that you can’t tell them apart.) In theory this would put the actual creators of CSAM out of business, thus saving a lot of children from the trauma.

    Most people down vote the idea on their gut reaction tho.

    Looks like they might do it on their own.

    • Itwasthegoat@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      4 months ago

      My concern is why would it put them out of business? If we just look at legal porn there is already beyond huge amounts already created, and the market is still there for new content to be created constantly. AI porn hasn’t noticeably decreased the amount produced.

      Really flooding the market with CSAM makes it easier to consume and may end up INCREASING the amount of people trying to get CSAM. That could end up encouraging more to be produced.

      • Nora@lemmy.ml
        link
        fedilink
        English
        arrow-up
        4
        ·
        4 months ago

        The market is slightly different tho. Most CSAM is images, with Porn theres a lot of video and images.

    • DarkThoughts@fedia.io
      link
      fedilink
      arrow-up
      5
      ·
      4 months ago

      It’s such an emotional topic that people lose all rationale. I remember the Reddit arguments in the comment sections about pedos, already equalizing the term with actual child rapists, while others would argue to differentiate because the former didn’t do anything wrong and shouldn’t be stigmatized for what’s going on in their heads but rather offered help to cope with it. The replies are typically accusations of those people making excuses for actual sexual abusers.

      I always had the standpoint that I do not really care about people’s fictional content. Be it lolis, torture, gore, or whatever other weird shit. If people are busy & getting their kicks from fictional stuff then I see that as better than using actual real life material, or even getting some hands on experiences, which all would involve actual real victims.

      And I think that should be generally the goal here, no? Be it pedos, sadists, sociopaths, whatever. In the end it should be not about them, but saving potential victims. But people rather throw around accusations and become all hysterical to paint themselves sitting on their moral high horse (ironically typically also calling for things like executions or castrations).

    • jaschen@lemm.ee
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      2
      ·
      4 months ago

      It’s also a victimless crime. Just like flooding the market with fake rhino horns and dropping the market price to a point that it isn’t worth it.

    • PirateJesus@lemmy.today
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      3
      ·
      4 months ago

      It would be illegal in the United States. Artistic depictions of CSAM are illegal under the PROTECT act 2003.

      • TheGrandNagus@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        4 months ago

        And yet it’s out there in droves on mainstream sites, completely without issue. Drawings and animations are pretty unpoliced.

  • badbytes@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    4
    ·
    4 months ago

    Breaking news: Paint made illegal, cause some moron painted something stupid.

    • cley_faye@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      4 months ago

      I’d usually agree with you, but it seems he sent them to an actual minor for “reasons”.

    • catloaf@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      2
      ·
      4 months ago

      Some places do lock up spray paint due to its use in graffiti, so that’s not without precedent.

      • Soggy@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        4 months ago

        They lock it up because it’s frequently stolen. (Because of its use in graffiti, but still.)

    • PirateJesus@lemmy.today
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      3
      ·
      4 months ago

      Asked whether more funding will be provided for the anti-paint enforcement divisions: it’s such a big backlog, we’ll rather just wait for somebody to piss of a politician to focus our resources.

  • Ibaudia@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    arrow-down
    1
    ·
    4 months ago

    Isn’t there evidence that as artificial CSAM is made more available, the actual amount of abuse is reduced? I would research this but I’m at work.

    • deathbird@mander.xyz
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      2
      ·
      4 months ago

      It would not need to be trained on CP. It would just need to know what human bodies can look like and what sex is.

      AIs usually try not to allow certain content to be produced, but it seems people are always finding ways to work around those safeguards.

    • SeattleRain@lemmy.world
      link
      fedilink
      English
      arrow-up
      21
      arrow-down
      1
      ·
      4 months ago

      Well yeah. Just because something makes you really uncomfortable doesn’t make it a crime. A crime has a victim.

      Also, the vast majority of children are victimized because of the US’ culture of authoritarianism and religious fundamentalism. That’s why far and away children are victimized by either a relative or in a church. But y’all ain’t ready to have that conversation.

      • sugartits@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        11
        ·
        4 months ago

        That thing over there being wrong doesn’t mean we can’t discuss this thing over here also being wrong.

        So perhaps pipe down with your dumb whataboutism.

        • SeattleRain@lemmy.world
          link
          fedilink
          English
          arrow-up
          12
          arrow-down
          1
          ·
          4 months ago

          It’s not whataboutism, he’s being persecuted because of the idea that he’s hurting children all the while law enforcement refuses to truly persecute actual institutions victimizing children and are often colluding with traffickers. For instance LE throughout the country were well aware of the scale of the Catholic church’s crimes for generations.

          How is this whataboutism.

          • DarkThoughts@fedia.io
            link
            fedilink
            arrow-up
            4
            ·
            4 months ago

            Just to be clear here, he’s not actually persecuted for generating such imagery like the headline implies.

          • sugartits@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            5
            ·
            4 months ago

            Because it’s two different things.

            We should absolutely go after the Catholic church for the crimes committed.

            But here we are talking about the creation of child porn.

            If you cannot understand this very simple premise, then we have nothing else to discuss.

            • SeattleRain@lemmy.world
              link
              fedilink
              English
              arrow-up
              8
              arrow-down
              1
              ·
              4 months ago

              They’re not two different things. They’re both supposedly acts of pedophilia except one would take actual courage to prosecute (churches) and the other which doesn’t have any actual victims is easy and is a PR get because certain people find it really icky.