In a busy period for Government on technology and data law reforms, in conjunction with its ongoing reviews of AI technology and systems, and complementing the Voluntary AI Safety Standard unveiled by Minister Ed Husic, the Federal Government recently released its ‘Proposals Paper for introducing mandatory guardrails for AI in high-risk settings’ (Paper) as an initial step towards Australia’s regulation of AI.
The Paper follows on from the Government’s interim response to the ‘Safe and Responsible AI in Australia discussion paper’, wherein they committed to developing a regulatory environment ‘with community trust and AI promotion in mind’.
A principles risk-based approach is proposed in the Paper (with a focus on ex ante (preventative) measures balanced by ex post (remedial) measures). Broadly, the Paper proposes a high-level definition of ‘high-risk’ AI by reference to a set of ‘proposed principles’, and the application of these principles to ‘general purpose AI’ (GPAI) models. A short consultation period only is proposed and interested submissions are requested from industry on or before 5.00pm AEST on Friday 4 October 2024.
The Paper is informed by the workings of the AI Expert Group (a 12-person expert panel appointed by the Federal Government earlier this year to guide the introduction of high-risk AI ‘guardrails’), with an eye on how international governments and regulators around the world are tackling the harms arising from AI technology. A specific set of ten mandatory guardrails have been proposed to mitigate harm arising out of the use and development of AI systems, including anticipated protections on accountability, governance, transparency, implementation of risk management processes and an acknowledgment that human ‘in the loop’ interventions are key.
What is curious is that the word ‘ethics’ does not expressly appear in these ten listed guardrails, although one might argue that the focus on transparency and the requirement to build processes to challenge outcomes may be a nod in an ethical direction. That said, it is somewhat disappointing that ethical considerations were not expressly called out as one of the guardrails, particularly noting the focus on high-risk AI.
The Consultation seeks feedback on the best way to transpose these principles and guardrails into mandatory legislation, with three options offered:
Option 1. | Implement through changes to existing regulations (e.g. ACL, Privacy Act, ASIC Act etc); |
Option 2. | Introduce a new principles-based framework, which ‘trickles down’ and is enforced under existing laws; |
Option 3. | Take a whole-of-economy approach and introduce a new specific stand-alone legislation (e.g. an ‘AI Act’). |
Current indications are that Option 2 may be the preferred route for many (already heavily) regulated sectors, but we wait to see the outcomes.
Proposed definition of ‘high risk AI’
A principles-based approach is being taken. The Paper is seeking feedback on a set of ‘Proposed Principles’ for determining high-risk AI settings, and the potential application of these Proposed Principles to ‘general purpose AI’ (GPAI) models.
The ‘Proposed Principles’ include that, in designating an AI system as high-risk due to its use, regard must be given to the risks, as well as the severity and extent, of adverse impacts.
Adverse impacts to the below should be considered: |
---|
An individual’s human rights (as recognised in Australian human rights law without justification and international human rights law obligations); |
An individual’s physical or mental health or safety; |
Adverse legal effects, defamation or similarly significant effects on an individual; |
Adverse impacts to groups of individuals or collective rights of cultural groups; and |
To the broader Australian economy, society, environment and rule of law. |
To reduce the likelihood of harms occurring from the development and use of high-risk AI systems, the following 10 ‘mandatory guardrails’ are proposed, in summary:
Guardrail 1 |
Establish, implement and publish an accountability process including governance, internal capability and a strategy for regulatory compliance. |
Guardrail 2 |
Establish and implement a risk management process to identify and mitigate risks. |
Guardrail 3 |
Protect AI systems, and implement data governance measures to manage data quality and provenance. |
Guardrail 4 |
Test AI models and systems to evaluate model performance and monitor the system once deployed. |
Guardrail 5 |
Enable human control or intervention in an AI system to achieve meaningful human oversight. |
Guardrail 6 |
Inform end-users regarding AI-enabled decisions, interactions with AI and AI-generated content. |
Guardrail 7 |
Establish processes for people impacted by AI systems to challenge use or outcomes. |
Guardrail 8 |
Be transparent with other organisations across the AI supply chain about data, models and systems to help them effectively address risks. |
Guardrail 9 |
Keep and maintain records to allow third parties to assess compliance with guardrails. |
Guardrail 10 |
Undertake conformity assessments to demonstrate and certify compliance with the guardrails. |
Option 1 | Option 2 | Option 3 |
---|---|---|
Adopt within our existing regulatory frameworks (and enforced by current applicable regulators). | Introduce new framework legislation (top-level) which would ‘trickle down into’ existing laws, with associated amendments to existing legislation. | Introduce a whole economy approach by introducing a cross-economy AI specific law (such as an Australian AI Act), to be regulated and governed by a new/specific AI regulator. |
Worth mentioning alongside these proposed changes in the regulation of high-risk AI, the Privacy and Other Legislation Amendment Bill 2024 (Bill) was introduced at the First Reading Stage to the House of Representatives as of 12 September 2024. In addition to other changes to the Privacy Act, the Bill proposes detailed reforms on automated decision-making (ADM) using personal information, such that, if enacted as proposed, entities that are using AI technology (encompassing ADM) – not only high-risk AI systems – would need to ensure transparency to individuals about:
Further information on ADM and these proposed new ADM requirements are set out in our separate Privacy Act Reforms Insight here.
The Paper includes various consultation questions on each of the proposals. This Consultation is open for four (4) weeks, closing at 5pm AEST on Friday 4 October 2024. Examples of questions include:
For more information on the Proposals Paper, click here. If you would like to make a submission, or contribute to a submission being made, or for any additional information on any of these proposals, please do reach out and contact any member of our team.
If you found this insight article useful and you would like to subscribe to Gadens’ updates, click here.
Authored by:
Sinead Lynch, Partner
Lucy Hardyman, Lawyer
Wen Wong, Lawyer