The Federal Government has this week released a Voluntary AI Safety Standard and a consultation paper on ‘Safe and responsible AI in Australia’ for introducing mandatory guardrails for AI in high-risk settings.
The paper seeks submissions on:
Submissions are due on 4 October 2024.
The proposed risk-based approach is focused on preventing harms before people interact with, or are subject to, an AI system and will predominantly apply to AI developers and deployers. This approach will bring Australia in line with the approaches being adopted in the European Union, Canada and the United Kingdom, and is consistent with the approach adopted in the multilateral Bletchley Declaration and Soeul Declaration to which Australia is a signatory.
The voluntary AI Safety Standard is intended to support and promote best practice, and mitigate the potential adverse impacts of AI developers and deployers adopting inconsistent approaches while the government considers its options on mandatory guardrails. It clearly sets expectations for what future legislation may look like.
The proposed approach is to regulate two categories of “high-risk” AI.
The first category is where the proposed use or application of the AI system or general purpose AI (GPAI model) are known or foreseeable. For this category, whether or not the AI system or GPAI is high-risk AI will be determined through the application of a set of principles requiring regard to be given to:
The second proposed category of high-risk AI applies to GPAI models which are advanced and highly capable where all possible risks and applications cannot be predicted.
The mandatory guardrails outlined below would apply to all high-risk AI.
Following the approach in Canada’s proposed Artificial Intelligence Data Act (AIDA), the paper proposes defining GPAI as ‘An AI model that is capable of being used, or capable of being adapted for use, for a variety of purposes, both for direct use as well as for integration in other systems’.
The paper also refers to the approach adopted in the EU AI Act which explicitly prohibits certain uses of AI which:
and deems other uses to be high risk:
The paper seeks feedback on types of AI use that could present an unacceptable level of risk in Australia and should be banned.
The mandatory guardrails are intended to be interoperable with those in other comparable jurisdictions (and, in particular, the EU’s AI Act and Canada’s proposed AIDA), and to align with national and international standards such as ISO/IEC 42001:2023 Artificial Intelligence Management System, which has been recognised as a standard that will support AI governance in Australia and internationally.
The proposed mandatory guardrails largely replicate those in the Voluntary AI standard introduced in August 2024 and require organisations developing or deploying high-risk AI systems to:
The only deviation between the mandatory and voluntary guardrails is the 10th guardrail. In the voluntary guardrails, the 10th is ‘engage your stakeholders and evaluate their needs and circumstances, with a focus on safety’, which emphasises the importance of ongoing engagement with stakeholders to evaluate their needs and circumstances.
The paper considers (and seeks submissions on) three regulatory options available to mandate the proposed mandatory guardrails:
For any questions about the Government’s new Voluntary AI Safety Standard and consultation paper on ‘Safe and responsible AI in Australia’, please contact our Technology team.