OpenAI Cracks Down On State AI Misuse
OpenAI Confronts State-Sponsored AI Abuse
OpenAI, the creators of ChatGPT, announced they have identified and stopped several state-sponsored operations that were misusing their AI technology for malicious activities such as developing malware and conducting espionage.
In their June security report, OpenAI detailed the disruption of multiple attacks. These primarily originated from China and Russia, where actors were reportedly leveraging ChatGPT to generate malicious code or automate social engineering tactics through social media posts and emails.
The OpenAI team stated in their report, "AI investigations are an evolving discipline." They added, "Every operation we disrupt gives us a better understanding of how threat actors are trying to abuse our models, and enables us to refine our defenses."
The report highlighted several case studies demonstrating how threat actors have utilized ChatGPT. Out of ten selected cases, seven focused on social engineering, and two involved generating code for malware.
North Korean IT Worker Impersonation Scheme Disrupted
The report also detailed a previously identified North Korean IT worker scheme. In this operation, threat actors from North Korea used ChatGPT to impersonate IT contractors from different regions.
These actors employed ChatGPT to create fraudulent job applications and resumes. Furthermore, the North Korean operatives used the AI for communicating with third-party collaborators involved in the scam.
The OpenAI team explained, "The core operators used ChatGPT as a research tool to help inform remote-work setups." They continued, "They also engaged our models to generate text concerning the recruitment of real people in the US to take delivery of company laptops, which would then be remotely accessed by the core threat actors or their contractors."
China and Russia Exploit AI for Social Media Manipulation
The majority of the identified operators were linked to China and Russia. Out of the ten attacks detailed, four originated from China and three from Russia. Other cases involved actors from Iran, North Korea, and the Philippines.
One notable operation, codenamed "Uncle Spam," involved Chinese threat actors using ChatGPT to generate social media content arguing both for and against U.S. tariffs, a contentious policy from the Trump administration.
These actors also used ChatGPT to create logos and images for fake groups on social media, aiming to further incite division among the American public.
OpenAI's Countermeasures and AI-Powered Defense
In response to each incident, OpenAI banned the offending accounts, thereby halting these malicious operations.
OpenAI highlighted that just as attackers are using AI, defenders are also leveraging it. The company revealed its own use of ChatGPT and its underlying models to help identify and monitor these threat actors.
The company stated, "By using AI as a force multiplier for our expert investigative teams, in the three months since our last report we’ve been able to detect, disrupt and expose abusive activity including social engineering, cyber espionage, deceptive employment schemes, covert influence operations and scams."