ASIC Chair Joe Longo raises concerns about AI regulations

Gaps identified to enhance consumer protection and industry integrity

ASIC Chair Joe Longo raises concerns about AI regulations

Technology

By Roxanne Libatique

Australian Securities and Investments Commission (ASIC) Chair Joe Longo has expressed concerns about the adequacy of current regulations surrounding artificial intelligence in Australia.

In a recent speech at the UTS Human Technology Institute Shaping Our Future Symposium, Longo emphasised that existing laws might fall short in preventing AI-related harms before they occur, necessitating additional efforts to ensure a robust response after such incidents take place.

Quoting from the federal government's interim report on AI regulation, Longo highlighted the evident gap between the current regulatory landscape and the desired ideal.

“[The] theme of ‘bridging the governance gap’ presupposes such a divide. It invites us to consider what AI governance and regulation might look like in the ideal, how great the divide is between that ideal and current circumstances – and, of course, how we might go about bridging that divide,” he said.

Is the existing regulation enough?

Despite the prevailing regulatory framework, Longo questioned whether it is sufficient to address the challenges posed by the accelerated development of AI technologies.

He acknowledged the immense potential benefits of AI, estimating a substantial contribution to Australia's GDP by 2030. However, the rapid pace of AI advancements raises crucial questions about transparency, explainability, and the capacity of existing regulations to adapt promptly.

“It isn't fanciful to imagine that credit providers using AI systems to identify ‘better’ credit risks could (potentially) unfairly discriminate against those vulnerable consumers. And with ‘opaque’ AI systems, the mechanisms by which that discrimination occurs could be difficult to detect. Even if the current laws are sufficient to punish bad action, their ability to prevent the harm might not be,” Longo said.

Recommendations

Longo emphasised the need for transparency and oversight to prevent unintended or unfair practices in various sectors, including insurance. He also suggested potential solutions, such as red-teaming and “AI constitutions,” while acknowledging their vulnerabilities.

Longo raised the possibility of mandatory “AI risk assessments,” akin to the EU approach, emphasising the need to ensure their effectiveness in preventing harm.

“These questions of transparency, explainability, and rapidity deserve careful attention. They can't be answered quickly or off-hand. But they must be addressed if we're to ensure the advancement of AI means an advancement for all. And as far as financial markets and services are concerned, it's clear there's a way to go in answering them,” he said.

In other news, ASIC has started monitoring instances of scammers exploiting its identity on social media and messaging platforms.

Related Stories

Keep up with the latest news and events

Join our mailing list, it’s free!