The rise and democratization of AI-generated content presents one of the biggest security and risk threats since the dawn of the internet. With over 3 billion images and 720,000 hours of video produced daily it’s a literal avalanche of content that even big tech firms such as Tik Tok and Facebook can’t adequately tackle.
For those new to the space, AI-generated content refers to media, such as text, images, audio, or video, created using artificial intelligence algorithms. These algorithms analyze and mimic patterns from existing data to produce new content, often indistinguishable from real content. So, that video you are watching of the Israel-Hamas war? It could be a fake created purely for propaganda purposes. Or those scandalous pictures of the FT500 CEO in a compromising situation? Also fabrications. What about that phone call you received from your CFO telling you to wire cash? Also not him.
Deep fakes, a subset of AI-generated content, involve manipulated videos or images that use AI techniques to overlay someone's likeness onto another person's body. Shallow fakes cost virtually nothing to make, and a deep fake can be made for as little as US$500. Last year, it was estimated that around 500,000 deep fakes were distributed. Of these, only 95,820 were identified.
Detecting AI-generated content and deep fakes is challenging and reliably discerning authentic from manipulated images is proving to be increasingly difficult. Indeed, the recent surge of explicit deep fakes of Taylor Swift on X illustrate this challenge. While X blocked searches for Swift’s name, the content continued to spread.
Of greater concern is the rise of AI-generated content on unmoderated social sites where there is little to no content moderation to begin with. These platforms prioritize user engagement over content authenticity, creating an environment where AI-generated content can spread rapidly without adequate oversight.
Unmoderated Social is a breeding ground for AI-generated content
AI-generated content frequently originates on unmoderated social, where fact-checking and content moderation is scarce. Of greater concern is that unmoderated platforms are beginning to create and train their own AI models with data directly sourced from their platforms, where divisive and hateful content are allowed to remain.
Indeed, 8kun, the home of QAnon, has announced they are launching an AI engine and Gab’s AI imagine generation model “Gabby” is already available to all members. Gab CEO Andrew Torba has critiqued existing AI models, claiming “every single one is skewed with a liberal/globalist/talmudic/satanic worldview. What if Gab AI Inc builds a Gab.ai… that is based, has no hate speech filters and doesn’t obfuscate and distort historical and Biblical Truth?”
In contrast, moderated platforms like Youtube, Facebook, TikTok and X, implement various content moderation policies, community guidelines, and automated systems to monitor and regulate the content shared on their platforms. In the context of AI, Tiktok and YouTube require users to disclose and label AI-generated content to prevent misleading viewers, Facebook’s community standards prohibit content that has been significantly altered to mislead viewers, including deep fake videos, and X’s policy explicitly states that the company may label misleading media posts and prohibits users from sharing “synthetic, manipulated, or out-of-context media that may deceive or confuse people and lead to harm (‘misleading media’).”
Monitoring unmoderated social would be a step in the right direction, but given the raison d'etre for these sites, this seems unlikely unless legislators were to make operations untenable without it. As we anticipate this will not take place in the U.S. in the near term, we expect the trend of deep fake creation and distribution on these sites to worsen as unmoderated social platforms continue to introduce more AI engines while moderated social steps up efforts to address AI-generated content on their platforms. Recent manifestations of this problem include:
Individuals and companies can face reputational and financial harm
Even without AI, the digital landscape presents substantial risks for individuals and companies. Over 43 million people in the U.S. have been reported to have been doxxed, and 95% of FTSE 100 companies were reported to have been frequently mentioned by non credible publications in the first half of 2023.
Malicious actors often exploit unmoderated social’s vulnerabilities to manipulate public perception and damage reputations. In 2023, disinformation campaigns against Target and BudLight for their LGBTQ+ initiatives lead to brand boycotts, and Etsy became the latest victim of a QAnon conspiracy theory that alleges certain online retailers are involved in trafficking children, resulting in financial losses and physical threats against Etsy’s executives.
The rise of AI and deep fake technology escalates threats like these to unprecedented levels of sophistication and scale, posing significant challenges in addressing disinformation.
Some examples of harm AI-generated content has caused individuals or companies are as follows:
As AI algorithms become increasingly adept at replicating and manipulating visual content, individuals and businesses face challenges in protecting their identity and mitigating the spread of disinformation. It’s not enough to expect mainstream platforms to remove content in a timely fashion. Security, risk and communications teams need to be armed with AI-enriched tools to ensure they are proactively searching for potentially malicious content across moderated and unmoderated sites to ensure they identify disinformation before it goes viral.
Pyrra’s goal is to make the internet and the world a safer place by identifying and combating hate speech, violent threats, reputation risk and brand damage across the unmoderated and alternative corners of the internet. Should your team need support in monitoring or identifying AI-generated threats or trends online, please contact us at sales@pyrratech.com.
1A shallow fake is a type of manipulated media that has been altered or edited using basic techniques, such as simple editing software or low-tech methods like slowing down or speeding up videos. Unlike deep fakes, which use advanced AI algorithms to create highly convincing fake content, shallow fakes are relatively easy to detect and often involve less sophisticated alterations.
2 Id.
3 David Gilbert. “White Supremacist Networks Gab and 8Kun Are Training Their Own AI Now.” Vice. February, 2023. https://www.vice.com/en/article/epzjpn/ai-chatbot-white-supremacist-gab
4 AI-Generated Content Label. TikTok. https://www.tiktok.com/creators/creator-portal/en-us/community-guidelines-and-safety/ai-generated-content-label/
5 Our approach to responsible AI innovation. Youtube. Nov, 2023. https://blog.youtube/inside-youtube/our-approach-to-responsible-ai-innovation/
6 Facebook Community Standards: Manipulated Media. Meta. https://transparency.fb.com/policies/community-standards/manipulated-media/?source=https%3A%2F%2Fwww.facebook.com%2Fcommunitystandards%2Fmanipulated_media
7 Our synthetic and manipulated media policy. Twitter. April, 2023. https://help.twitter.com/en/rules-and-policies/manipulated-media
8 Ashley Belanger. “Toxic Telegram group produced X’s X-rated fake AI Taylor Swift images, report says.” ARS Technica. Jan, 2024. “https://arstechnica.com/tech-policy/2024/01/fake-ai-taylor-swift-images-flood-x-amid-calls-to-criminalize-deepfake-porn/
9 X banned searches for Swift and queries relating to the photos, instead displaying an error message. And Instagram and Threads, while allowing searches for Swift, display a warning message searching for the images.; Tiffany Hsu. “Fake and Explicit Images of Taylor Swift Started on 4chan, Study Says.” NYT. Feb, 2024. https://www.nytimes.com/2024/02/05/business/media/taylor-swift-ai-fake-images.html
10 Khatsenkova, Sophia. “Israel-Hamas War: This viral image of a baby trapped under rubble turned out to be fake.” Euronews. 24, Oct. 2023. https://www.euronews.com/my-europe/2023/10/24/israel-hamas-war-this-viral-image-of-a-baby-trapped-under-rubble-turned-out-to-be-fake
11 Hsu, Tiffany and Stuart A. Thompson. “A.I. Muddies Israel-Hamas War in Unexpected Way.” New York Times. 30, Oct 2023. https://www.nytimes.com/2023/10/28/business/media/ai-muddies-israel-hamas-war-in-unexpected-way.html
12Biden Down. telegram. January, 2024. https://t.me/Biden_Down/258
13 Case Study: Target’s Celebration of Pride Month Has Put Them at The Center of Far-Right, Anti-LGBTQ+ Rhetoric. Pyrra. October, 2023.
14Case Study: Alt-Social Media and the Bud Light Backlash. Pyrra. April, 2023. https://www.pyrratech.com/articles/case-study-alt-social-media-and-the-bud-light-backlash
15 Etsy Under Fire: The Latest Victim of a QAnon Conspiracy. Pyrra. January, 2024. https://www.pyrratech.com/articles/etsy-under-fire-the-latest-victim-of-a-qanon-conspiracy
16 Cindy Gordon. “Use Of AI In DeepFakes Accelerating Risks To Companies.”Forbes. Dec, 2023. https://www.forbes.com/sites/cindygordon/2023/12/26/use-of-ai-in-deepfakes-accelerating-risks-to-companies/?sh=4df5bcfc7284
17 Eric Hal Schwartz. “Deepfake Drake and the Weeknd Creator Doubles Down With Rihanna and Bad Bunny Track.” Voicebot. April, 2023. https://voicebot.ai/2023/04/26/deepfake-drake-and-the-weeknd-creator-doubles-down-with-rihanna-and-bad-bunny-track/
18 https://coingape.com/google-makes-controversial-decision-on-ripple-ceo-deepfake-scam/
19 “Elon Musk Used in Fake AI Videos.” RMIT. Aug, 2023. https://www.rmit.edu.au/news/factlab-meta/elon-musk-used-in-fake-ai-videos-to-promote-financial-scam
Stay ahead of what's trending on the fringe, with our free weekly newsletter