Explicit Deepfake Images of Taylor Swift Elude Safeguards and Swamp Social Media
Fake, sexually explicit images of Taylor Swift likely generated by artificial intelligence spread rapidly across social media platforms this week, disturbing fans who saw them and reigniting calls from lawmakers to protect women and crack down on the platforms and technology that spread such images.
One image shared by a user on X was viewed 47 million times before the account was suspended on Thursday. X suspended several accounts that posted the faked images of Ms. Swift, but the images were shared on other social media platforms and continued to spread despite those companies’ efforts to remove them.
While X said it was working to remove the images, fans of the pop superstar flooded the platform in protest. They posted related keywords, along with the sentence “Protect Taylor Swift,” in an effort to drown out the explicit images and make them more difficult to find.
Reality Defender, a cybersecurity company focused on detecting A.I., determined with 90 percent confidence that the images were created using a diffusion model, an A.I.-driven technology accessible through more than 100,000 apps and publicly available models, said Ben Colman, the company’s co-founder and chief executive.
As the A.I. industry has boomed, companies have raced to release tools that enable users to create images, videos, text and audio recordings with simple prompts. The A.I. tools are wildly popular but have made it easier and cheaper than ever to create so-called deepfakes, which portray people doing or saying things they have never done.
Researchers now fear that deepfakes are becoming a powerful disinformation force, enabling everyday internet users to create nonconsensual nude images or embarrassing portrayals of political candidates. Artificial intelligence was used to create fake robocalls of President Biden during the New Hampshire primary, and Ms. Swift was featured this month in deepfake ads hawking cookware.