A new report reveals some disturbing news from the world of AI image generation: A Stanford-based watchdog group has discovered thousands of images of child sexual abuse in a popular open-source image ...
LAION, the German research org that created the data used to train Stable Diffusion, among other generative AI models, has released a new dataset that it claims has been "thoroughly cleaned of known ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results