Yazılar

Google Reports 250 Complaints Over AI-Generated Deepfake Terrorism Content to Australian Regulator

Google has informed Australian regulators that it received over 250 complaints globally between April 2023 and February 2024, indicating that its AI technology, specifically the Gemini model, was being used to create deepfake terrorism content. Additionally, the company reported dozens of complaints regarding the use of Gemini to generate child abuse material, according to the Australian eSafety Commission.

Under Australian law, tech companies are required to periodically report their harm minimization efforts to the eSafety Commission, or risk facing fines. This reporting period marks the first disclosure of such data, which regulators have described as a “world-first insight” into how AI is being exploited for harmful and illegal purposes.

The Australian eSafety Commission emphasized the importance of companies developing AI products to implement safeguards to prevent the generation of harmful material. eSafety Commissioner Julie Inman Grant stated that the findings highlight the critical need for effective protective measures.

According to Google’s report, it received 258 user complaints about AI-generated deepfake terrorist or extremist content created with Gemini, along with 86 reports concerning AI-generated child exploitation or abuse material. However, the company did not specify how many of these complaints were verified.

A Google spokesperson confirmed that the company does not allow the generation or distribution of illegal content, including material related to terrorism, child exploitation, or other abuses. Google also noted that the number of reports provided to eSafety represents the total global volume of complaints, not confirmed policy violations.

While Google uses a system called “hatch-matching” to identify and remove child abuse content generated with Gemini, the company did not apply the same system to detect terrorist or extremist material. This lack of a similar safeguard for violent content has raised concerns among regulators.

The Australian eSafety Commission has previously fined Telegram and Twitter (now X) for their inadequate reporting practices, with X losing an appeal over a fine of A$610,500 ($382,000). Telegram is also preparing to challenge its fine.

UK Introduces AI-Driven Child Abuse Material Offenses

The United Kingdom has announced it will make it illegal to use artificial intelligence (AI) tools to create child sexual abuse material, marking the first country to introduce such AI-related offenses. This new legislation is part of a broader effort to address the rising concern of online criminals using AI to create explicit images of children. Under the current law in England and Wales, possessing, making, showing, or distributing explicit images of children is a criminal act, but the new offenses will specifically target the use of AI tools to manipulate real-life images of children.

The move comes as reports of AI-generated child abuse material have surged nearly five-fold in 2024, according to the Internet Watch Foundation. “We know that sick predators’ activities online often lead to them carrying out the most horrific abuse in person,” said Britain’s interior minister Yvette Cooper. She emphasized the importance of tackling both online and offline child sexual abuse to better protect the public from emerging threats.

In addition to AI-generated content, predators are also using AI tools to create fake images for blackmail, coercing children into further abuse, such as through live streaming. The new legislation will criminalize the possession, creation, or distribution of AI tools designed to produce child sexual abuse material, as well as the possession of “paedophile manuals” that provide instructions on using such technologies.

A further offense will target the operators of websites that distribute such harmful content, and authorities will be empowered to unlock and inspect digital devices involved in these crimes. These measures will be incorporated into the Crime and Policing Bill when it is introduced in parliament. Earlier this month, the UK also announced plans to make the creation and sharing of AI-generated “deepfake” content, including videos, pictures, and audio clips that are sexually explicit, a criminal offense.