OpenAI bans Chinese, North Korean hacker accounts using ChatGPT to launch surveillance

Ai tech, businessman show virtual graphic Global Internet connect Chatgpt Chat with AI, Artificial Intelligence.
(Image credit: Shutterstock/SomYuZu)

  • OpenAI bans accounts linked to China, North Korea for malicious AI-assisted surveillance and phishing
  • Chinese actors used ChatGPT to draft proposals for monitoring tools and behavioral profiling systems
  • North Korean actors tested phishing, credential theft, and macOS malware development using rephrased prompts

OpenAI has banned Chinese, North Korean, and other accounts which were reportedly using ChatGPT to launch surveillance campaigns, develop phishing techniques and malware, and engage in other malicious practices.

In a new report, OpenAI said it observed individuals reportedly affiliated with Chinese government entities, or state-linked organizations, using its Large Language Model (LLM) to help write proposals for surveillance systems and profiling technologies.

These included tools for monitoring individuals and analyzing behavioral patterns.

Exploring phishing

“Some of the accounts that we banned appeared to be attempting to use ChatGPT to develop tools for large-scale monitoring: analyzing datasets, often gathered from Western or Chinese social media platforms,” the report reads.

“These users typically asked ChatGPT to help design such tools or generate promotional materials about them, but not to implement the monitoring.”

The prompts were framed in a way that avoided triggering safety filters, and were often phrased as academic or technical inquiries.

While the returned content did not directly enable surveillance, its outputs were used to refine documentation and planning for such systems, it was said.

The North Koreans, on the other hand, used ChatGPT to explore phishing techniques, credential theft, and macOS malware development.

OpenAI said it observed these accounts testing prompts related to social engineering, password harvesting, and debugging malicious code, especially targeting Apple systems.

The model refused direct requests for malicious code, OpenAI said, but stressed that the threat actors still tried to bypass safeguards by rephrasing prompts, or asking for general technical help.

Just like any other tool, LLMs are being used by both financially motivated, and state-sponsored threat actors, for all sorts of malicious activity.

This AI misuse is evolving, with threat actors increasingly integrating AI into existing workflows to improve their efficiency.

While developers such as OpenAI work hard on minimizing risk and making sure their products cannot be used like this, there are many prompts that fall between legitimate and malicious use. This gray zone activity, the report hints, requires nuanced detection strategies.

Via The Register


Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button!

And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.

You might also like

Sead is a seasoned freelance journalist based in Sarajevo, Bosnia and Herzegovina. He writes about IT (cloud, IoT, 5G, VPN) and cybersecurity (ransomware, data breaches, laws and regulations). In his career, spanning more than a decade, he’s written for numerous media outlets, including Al Jazeera Balkans. He’s also held several modules on content writing for Represent Communications.

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.