OpenAI, Microsoft shut down malicious accounts linked to China, others
Testing on staging11
(NewsNation) — OpenAI and Microsoft Threat Intelligence shut down accounts linked to five state-affiliated actors, some tied to China and Russia, trying to use AI for malicious reasons, the companies announced Wednesday.
“We terminated accounts associated with state-affiliated threat actors,” OpenAI, the creator of ChatGPT, said in an official statement. “Our findings show our models offer only limited, incremental capabilities for malicious cybersecurity tasks.”
The accounts terminated included China-affiliated Charcoal Typhoon and Salmon Typhoon, Iran-affiliated Crimson Sandstorm, North Korea-affiliated Emerald Sleet and Russia-affiliated Forest Blizzard, according to OpenAI’s statement.
Microsoft Threat Intelligence tracks more than 300 unique threat actors, including 160 nation-state actors, and 50 ransomware groups.
Online actors’ motivations vary, but their efforts may involve similar practices such as learning about potential victims’ industries, locations and relationships; improving software scripts and malware development; and assistance with learning and using native languages, according to Microsoft.
The entities at the center of Wednesday’s announcement used OpenAI services to research companies and cybersecurity tools, debug code and likely create phishing campaigns, among other activities.
“Microsoft and OpenAI have not yet observed particularly novel or unique AI-enabled attack or abuse techniques resulting from threat actors’ usage of AI,” Microsoft wrote in a blog post published on Wednesday. “Microsoft and our partners continue to study this landscape closely.”
OpenAI and the FBI declined to provide further comment.
NewsNation has reached out to Microsoft for further comment but has not yet received a response. The Department of Homeland Security hasn’t immediately returned an email seeking information.
In June, seven major tech companies, including Microsoft and OpenAI, agreed to follow a set of White House AI safety guidelines.
Those voluntary commitments include conducting external security testing of AI systems before they’re released and sharing information about managing AI risks industry-wide as well as with governments, academia and the general public.
The companies also vowed to report vulnerabilities in their products and invest in cybersecurity insider threat safeguards.