Laptop on table at night
New safety tests by Johns Hopkins researchers reveal vulnerabilities of popular systems like DALL-E 2. Laptop on table at night A new test of popular AI image generators shows that while they're supposed to make only G-rated pictures, they can be hacked to create content that's not suitable for work. Most online art generators are purported to block violent, pornographic, and other types of questionable content. But Johns Hopkins researchers manipulated two of the better-known systems to create exactly the kind of images the products' safeguards are supposed to exclude. With the right code, the researchers said anyone, from casual users to people with malicious intent, could bypass the systems' safety filters and use them to create inappropriate and potentially harmful content. "We are showing these systems are just not doing enough to block NSFW content. We are showing people could take advantage of them." Yinzhi Cao "We are showing these systems are just not doing enough to block NSFW content," said author Yinzhi Cao , a Johns Hopkins computer scientist at the Whiting School of Engineering.
TO READ THIS ARTICLE, CREATE YOUR ACCOUNT
And extend your reading, free of charge and with no commitment.