A new report issued by Human Rights Watch reveals that a widely used, web-scraped AI training dataset includes images of and information about real children — meaning that generative AI tools have ...
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More A massive open-source AI dataset, LAION-5B, which has been used to train ...
A new report reveals some disturbing news from the world of AI image generation: A Stanford-based watchdog group has discovered thousands of images of child sexual abuse in a popular open-source image ...
LAION, the German research org that created the data used to train Stable Diffusion, among other generative AI models, has released a new dataset that it claims has been “thoroughly cleaned of known ...
Hidden inside the foundation of popular artificial intelligence image-generators are thousands of images of child sexual abuse, according to a new report that urges companies to take action to address ...
Researchers from the Stanford Internet Observatory say that a dataset used to train AI image generation tools contains at least 1,008 validated instances of child sexual abuse material. The Stanford ...
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More Getty Images is going all in to establish itself as a trusted data ...
Hidden inside the foundation of popular artificial intelligence image-generators are thousands of images of child sexual abuse, according to a new report that urges companies to take action to address ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results