@DBActive
young girls, isn't it likely that will incorporate some of the child sexual abuse images? If nothing else, they potenially will use recognizable faces
To my knowledge, that's absolutely not how those things work, despite common misconception, at least not Stable Diffusion derivatives and relatives that dominate the scene. I have seen one that seemed to attempt something like that, to basically auto-photoshop elements together, and it was absolutely awful and very limited. All the others start with, basically random noise, and try to recognize patterns and then detail them out in iterative adversary process. In short, (most) AI generated images are by they nature unique, not exact copies of any part of any existing image, but follow patterns learned from training data.
That said, there finite variety of possible faces in existence, so it's just matter of time before there's be random AI generated porn images with yours, even if you had never any image of you taken and stored in any electronic device, at least as far your face could be described as pretty by any stretch. The same is true for nipples and even vulvas and penises, in theory. Although the latter two seems challenging for the AI for now (at least for the more accessible ones), and even nipples... well, I struggle to get exactly what I would like to see, to the point of basically giving up.
They tried to take porn out of Stable Diffusion training data, and it was a disaster. The model basically lost ability to produce acceptable human proportions. There's a reason any serious artist study nude drawing at some point, and AI artist is no exception; it's necessary knowledge to depict human form.
To the point you will get more pleasant results by promting "naked girl wearing school uniform" than just "girl in school uniform"; both will come out dressed most of the time, but the first will have better proportions and more precise anatomy.
I have played with several free or semi-free image generation products, mostly, I believe, ultimately Stable Diffusion based under the hood (except that one, that clearly wasn't, and was awful). To the tune of low tens of gigabytes of smallish low resolution images by now, almost exclusively all NSFW. They all do produce photo realistic pictures, including nudity, to an extent.
Genitals may come out weird. There's a gulf between toddler and adult that's basically unattainable (whith what I have explored enough, at least), except faces. Bodies are rendered as mini-adults most of the time, so what you are getting is more or less adult bodies with children heads. Even toddlers may be rendered with adult breasts and so on. Breasts get bigger the more attention you give them in your promt, almost no matter what you say about them.
We talk about nudity, basically nudism. (Not hardcore porn, more on that later.) Nudity is readily and commonplace generated. Depending on provider, model and such, you may or may not get completely surprise unsolicited nudity in your results on basically any promt. Some places seems to try to suppress that, so you won't get nudity without emphasis on it in the promt, likely by additional silent keywords, and/or negative prompts (I know at least one that doesn't provide negative prompts to users, but add default negative prompts). Both are easily overcome by emphasizing. You may sometimes do it unknowingly.
Prevention of exploitation is very varied. Few do accept any promt as is, with no pretense. Many places ban keywords, usually in very lazy ways. As in, it will balk on "nude" but will happily accept "_nude" or "nude2" (while the actual generator interpreter seems to ignore numericals attached to a keyword, all while reading L337 just fine) or even "(nude)" -- with is actually basic emphasis in SD prompting. (Undocumented, but in my experience, "_keyword_" is stronger than "keyword" too, in most cases.)
That's before we go into euphemisms and such, like, "in birthday suit" works surprisingly flawlessly anywhere I have tried it, and so do "wearing shiny skin" and I see no possible general remedies as those are combinations of otherwise innocent words. And even if they add phrase recognition to the censors, the way the promts work, you don't necessarily have to have those words adjacent even. That's why I said you may stumble upon a promt that generates nudity by surprise, to you.
Some may go as far as to refuse to display results that are recognized as nsfw, but that recognition is generally surprisingly wacky in my experience. Nipples are easily recognized, but ignored up to certain relative size, shadow under breasts is more triggering, and bottomless women are not recognized as inappropriate more often than not, unless the legs are spread. Some poses seems to be triggering outright, even dressed, and swimsuits or tight leotards may trigger nsfw flag, especially on cartoon characters, surprisingly, while nude characters with weird skin color are accepted.
All in all, you will get some CP-like content trough those filters more often and reliably than a big busted pornstar. At least that's my experience with one site that only employs (very sketchy) nsfw filter for image upload for image-to-image (and therefore very open to direct testing with premade and adjusted images), but a few tests I have done elsewhere suggest that the sites that ban output may use similar if not the same.
Actual sex content is... well... it's almost not there. Well, with some insistence and many, many attempts you may get some rather interesting images, but you have to accept bulk of your results will be terrible mutations and mutilations and fuzzed together, melting bodies with wrong count of limbs... and simple posed nudes, sometimes not even thematically adjacent to your promt at all. or at least that's my experience so far.
That may change, or already have changed in actually pornographic oriented generators as I have not tested many of them. The one I have taken a look at, was very very limited, basically just a small set of templates of women posing, with handful of fixed tags and no free form promt at all, and may understand why so very well now, by my little experience with general models already.
(Edit: spot an autocorrect error)