OpenAI Reports Rise in Chinese Groups Using ChatGPT for Malicious Activities
OpenAI disclosed in a report released Thursday that it has detected an increasing number of Chinese-linked groups leveraging its AI technology, including ChatGPT, for covert and malicious operations. Although the activities have expanded in scope and tactics, OpenAI noted the operations remain generally small in scale and target limited audiences.
Since its launch in late 2022, ChatGPT and other generative AI tools have raised concerns about misuse, including the rapid creation of human-like text, images, and audio that can be weaponized for misinformation, hacking, or social manipulation. OpenAI regularly monitors and publishes findings on such harmful usage on its platform.
Among the examples cited by OpenAI:
-
Accounts generating politically charged social media posts related to China, including critiques of a Taiwan-centric video game, false claims against a Pakistani activist, and content about the USAID closure. Some posts also criticized U.S. President Donald Trump’s tariffs with messages such as “Tariffs make imported goods outrageously expensive, yet the government splurges on overseas aid. Who’s supposed to keep eating?”
-
Chinese threat actors employing AI to assist in cyber operations, including open-source intelligence gathering, script modification, system troubleshooting, and creating tools for password brute forcing and automating social media actions.
-
Influence campaigns originating from China producing divisive content on U.S. political topics, often supporting opposing sides simultaneously, combined with AI-generated profile images to amplify polarization.
In response, China’s Foreign Ministry dismissed OpenAI’s claims as baseless and stressed its commitment to responsible AI governance and opposition to AI misuse.
OpenAI, valued at around $300 billion after a recent $40 billion funding round, continues to emphasize transparency and vigilance in monitoring misuse of its AI technologies worldwide.



