Experts are cautioning individuals to be wary of AI-generated YouTube content, as a recent uptick with the videos spreading malicious content has been noticed. With the advancements in artificial intelligence technology, it’s becoming increasingly easier to produce highly realistic videos that appear to be created by genuine individuals. However, these deep fakes could be employed for malicious purposes, such as spreading malware, propaganda, or impersonating actual people to deceive viewers. Some cases have seen legitimate YouTube pages hacked and used to spread these very real-looking, but ultimately fake videos around to various audiences. Some of the more well-known malware strains that are being spread include Raccoon, RedLine, and Vidar. To make this threat harder to detect, threat actors are using SEO poisoning to make these videos appear first on YouTube and are adding fake comments and likes to the videos to make them seem more legitimate.
To combat the dissemination of disinformation, experts advise individuals to exercise caution when watching videos that seem too good to be true and verify information from multiple sources. They also urge social media platforms to proactively identify and remove fraudulent content. Employing multi-factor authentication as well as avoiding clicking suspicious links from unknown sources is advised.