Title: Artificial Intelligence and Child Abuse: The Push for Stronger Regulations
Introduction:
Artificial intelligence (AI) tools have the potential to be used for malicious purposes, such as generating child abuse images and terrorist propaganda. In response to this growing concern, Australia’s eSafety Commissioner has announced a world-leading industry standard that requires tech giants like Google and Microsoft to take action against such material on their AI-powered search engines. This article delves into the new industry code, its implications, and the need for greater regulation in the AI realm.
AI-Powered Search Engines and Child Abuse Material:
The new industry code mandates that search engines like Google and Bing must eliminate child abuse material from their search results. It also requires them to prevent generative AI products from being used to create deepfake versions of such material. The eSafety Commissioner, Julie Inman Grant, emphasizes the responsibility of these tech companies to minimize the harms associated with their products.
Expanding Regulatory Scope:
According to Inman Grant, the previous version of the code only covered online material that search engines returned in response to user queries. However, with the advancement of AI technology, it has become necessary to include material that these services can generate, such as deepfake images and videos. The new code will ensure that dangerous content, including child sexual exploitation, pro-terror material, and extreme violence, does not appear in search results.
Researching Technologies to Detect and Identify Deepfakes:
The industry code also obliges search engines to explore technologies that can help users detect and identify deepfake images available on their platforms. This is a proactive measure to tackle the potential misuse of AI tools for illicit purposes. The eSafety Commission considers this framework to be one of the first of its kind globally.
Regulation at the Design and Deployment Phase:
Inman Grant likens the rapid evolution of AI technology to an “arms race” and believes that policymakers and regulators need to rethink their approach. She highlights the importance of implementing regulations at the design and deployment phase of AI tools to stay ahead of potential issues. Drawing a parallel with car safety standards, she underscores the urgency of proactive regulation in the tech industry.
The Need for Collaborative Efforts:
The eSafety Commissioner emphasizes that while regulations can address some of the challenges tied to AI, the tech companies themselves need to be actively involved in developing tools for greater safety. Inman Grant stresses the importance of putting guard rails in place to prevent predators from using AI tools to create synthetic child abuse material. Collaboration between regulators and tech firms is essential to combat the potential misuse of AI.
Conclusion:
The new industry code introduced by Australia’s eSafety Commissioner strengthens regulations around AI-powered search engines to prevent the generation and dissemination of harmful content, specifically child abuse material and deepfakes. This move reflects the need for proactive regulation and collaboration between regulators and technology companies to minimize the risks associated with AI. By addressing these challenges head-on, Australia sets a precedent for other countries to follow in order to harness the potential of AI technology while safeguarding against its misuse.
IntelliPrompt curated this article: Read the full story at the original source by clicking here