Protecting AI against foreign influence operations designed to infect LLMs
NewsGuard’s Foreign Adversary Influence in LLMs Safety Service (FAILSafe) helps AI companies detect and defend against foreign influence operations aimed at tainting AI responses with state-sponsored disinformation narratives and propaganda.
Created in response to a groundbreaking NewsGuard audit that found Russian disinformation networks had infected top AI tools, leading those tools to repeat propaganda narratives 33% of the time, FAILSafe provides AI companies with real-time data, verified by disinformation researchers with an expertise in foreign malign influence, on narratives and sources involved in adverse influence operations run by the Russian, Chinese, and Iranian governments.