Australia: Government releases interim response to AI consultation

The Australian Government's interim response to the "Safe and responsible AI in Australia" discussion paper flags a risk-based approach to AI governance in Australia including a mix of voluntary AI safety standards, voluntary labelling and watermarking for generative AI and the development of mandatory guardrails with a particular focus on high-risk and frontier AI applications.

In brief

On 17 January 2024, the Australian Government released its interim response to submissions received in relation to its 2023 Safe and Responsible AI in Australia discussion paper, which sought views on whether Australia has the right regulatory and governance arrangements in place to support the safe and responsible use and development of artificial intelligence (AI) technologies.

The Government in its interim response agrees with several key points expressed in submissions, including concerns that high-risk and frontier AI applications are not currently subject to sufficient regulation and that a risk-based regulatory approach to AI is appropriate, targeting regulatory requirements for AI applications with a higher risk of harm. The interim response proposes that additional guardrails (including mandatory guardrails for high-risk AI) should be put in place to address potential harms associated with high-risk AI applications, subject to further consultation.


Contents

The interim response also indicates the Australian government's next steps for AI governance will focus on:

  • Preventing harm through testing, transparency and accountability
  • Clarifying and strengthening laws
  • International work to support safe development and deployment of AI
  • Maximizing the benefits of AI

Key takeaways

  • The Government's immediate focus is to consider what mandatory AI safeguards are appropriate and how best to implement them as part of a risk-based regulatory approach to AI, informed by developments in other countries. This includes further consultation with industry to formulate an appropriate definition of "high-risk" AI in the Australian context.
  • Once the concept of "high-risk" AI is appropriately defined, a key safety measure the Government will focus on is developing regulatory guardrails around testing, transparency and accountability for such "high-risk" AI applications.
  • The use of low risk AI tools and applications is likely to be allowed to continue under existing legal frameworks without the need for bespoke AI regulation, to preserve the benefit and utility of those technologies.
  • While considering mandatory guardrails for AI development and use, the Government indicates it is already working with industry to develop a voluntary AI Safety Standard and develop options for a voluntary labelling and watermarking scheme for AI-generated materials.
  • The Government is also establishing a temporary expert advisory group to support the development of further guardrails for AI.

In more detail

Proposed actions

The Government will consider and consult on the introduction of new mandatory guardrails for the development and deployment of AI in high-risk settings. These guardrails are expected to focus on:

  • Testing – the Government will consider imposing requirements relating to testing AI systems before and after release, measures of best practice for safety, auditing and performance monitoring, and security thresholds.
  • Transparency – this would involve improving transparent communication to users when AI is used in systems or to generate content, and publicly reporting limitations of AI and the data which is used to train and test an AI model.
  • Accountability – the Government suggests that developers or deployers of AI products should be trained in certain settings, and that there should be specific roles responsible for AI safety. 

The Government also intends to define what qualifies as "high-risk" AI in an Australian context, which will be important to enable businesses to understand their regulatory obligations once mandatory guardrails are in place.

In parallel with its consideration of mandatory guardrails for "high-risk" AI development and use, the Government indicates it is already working with industry to develop a voluntary AI Safety Standard implementing (non-binding) risk-based guardrails, which may help businesses to manage risks associated with AI deployments, whether or not in a high-risk setting.

In the immediate future, the Government also states that it will:

  • Set up a temporary expert advisory group which will coordinate the development of AI guardrails.
  • Task the National AI Centre with collaborating with industry to "produce a best-practice and up-to-date voluntary AI risk-based safety framework” for Australian businesses.
  • Assess, in consultation with industry, the potential for a voluntary watermarking and labelling scheme for businesses to disclose the use of AI generated content in high-risk settings.

Other planned actions include:

  • Considering options for strengthening existing laws to address risks and harms from AI, building on recent and already-proposed reforms on matters relevant to AI (e.g., privacy, online safety, misinformation) – potential reforms to Australia’s privacy laws include requiring non-government entities to conduct a privacy impact assessment for activities with high privacy risks to identify and manage, minimize or eliminate risks (which is already a requirement for government entities) and other amendments with a focus on increasing transparency and integrity of automated decision-making which uses personal information.
  • Taking forward commitments made in the Bletchley Declaration, including supporting the development of a State of the Science report.
  • Continuing to engage internationally, aiming to shape global AI governance opportunities, support Australian involvement in developing technical AI standards, and understand other jurisdictions' responses to AI.
  • Considering opportunities to support the development and adoption of automation technologies such as AI and robots, building on existing investments in this area.

A principled approach leveraging existing requirements

The interim response indicates the Government will not follow the prescriptive approach set to be adopted by the EU that seeks to manage the lifecycle of AI via a comprehensive legislative approach under the AI Act, but will rather adopt a principles-based approach similar to that implemented in the UK and when defining the mandatory guardrails for high-risk AI applications, will look to leverage existing regulatory requirements wherever appropriate.

Implications

While the interim response does not include firm commitments to a legislative path forward for AI regulation, it does openly acknowledge that the current laws and regulations in Australia are not adequate to address the rapid moving nature of AI and its associated risks. The interim response does not provide firm timelines for the implementation of any proposed changes but indicates that there are a mix of short- and long-term developments on the horizon. The Government indicates that it will consult closely with industry, academia and the community on its proposed actions for AI. For this reason, all businesses with an interest in AI (whether developers, providers or users of AI solutions) should carefully consider the interim response and look for opportunities to have their say.

* * * * *

With thanks to Nicola McGran (Summer Clerk) and Liz Grimwood-Taylor (Senior Knowledge Lawyer) for their assistance with this alert.


Copyright © 2024 Baker & McKenzie. All rights reserved. Ownership: This documentation and content (Content) is a proprietary resource owned exclusively by Baker McKenzie (meaning Baker & McKenzie International and its member firms). The Content is protected under international copyright conventions. Use of this Content does not of itself create a contractual relationship, nor any attorney/client relationship, between Baker McKenzie and any person. Non-reliance and exclusion: All Content is for informational purposes only and may not reflect the most current legal and regulatory developments. All summaries of the laws, regulations and practice are subject to change. The Content is not offered as legal or professional advice for any specific matter. It is not intended to be a substitute for reference to (and compliance with) the detailed provisions of applicable laws, rules, regulations or forms. Legal advice should always be sought before taking any action or refraining from taking any action based on any Content. Baker McKenzie and the editors and the contributing authors do not guarantee the accuracy of the Content and expressly disclaim any and all liability to any person in respect of the consequences of anything done or permitted to be done or omitted to be done wholly or partly in reliance upon the whole or any part of the Content. The Content may contain links to external websites and external websites may link to the Content. Baker McKenzie is not responsible for the content or operation of any such external sites and disclaims all liability, howsoever occurring, in respect of the content or operation of any such external websites. Attorney Advertising: This Content may qualify as “Attorney Advertising” requiring notice in some jurisdictions. To the extent that this Content may qualify as Attorney Advertising, PRIOR RESULTS DO NOT GUARANTEE A SIMILAR OUTCOME. Reproduction: Reproduction of reasonable portions of the Content is permitted provided that (i) such reproductions are made available free of charge and for non-commercial purposes, (ii) such reproductions are properly attributed to Baker McKenzie, (iii) the portion of the Content being reproduced is not altered or made available in a manner that modifies the Content or presents the Content being reproduced in a false light and (iv) notice is made to the disclaimers included on the Content. The permission to re-copy does not allow for incorporation of any substantial portion of the Content in any work or publication, whether in hard copy, electronic or any other form or for commercial purposes.