Google AI
The Times Australia
The Times Australia
.

Most Australian government agencies aren’t transparent about how they use AI

  • Written by José-Miguel Bello y Villarino, Senior Research Fellow, Sydney Law School, University of Sydney

A year ago, the Commonwealth government established a policy requiring most federal agencies to publish “AI transparency statements” on their websites by February 2025[1]. These statements were meant to explain how agencies use artificial intelligence (AI), in what domains and with what safeguards.

The stated goal[2] was to build public trust in government use of AI – without resorting to legislation. Six months after the deadline, early results from our research (to be published in full later this year) suggest this policy is not working.

We looked at 224 agencies and found only 29 had easily identifiable AI transparency statements. A deeper search found 101 links to statements.

That adds up to a compliance rate of around 45%, although for some agencies (such as defence, intelligence and corporate agencies) publishing a statement is recommended rather than required, and it is possible some agencies could share the same statement. Still, these tentative early findings raise serious questions about the effectiveness of Australia’s “soft-touch” approach to AI governance in the public sector.

Why AI transparency matters

Public trust in AI in Australia is already low[3]. The Commonwealth’s reluctance to legislate rules and safeguards for the use of automated decision making in the public sector – identified as a shortcoming by the Robodebt royal commission[4] – makes transparency all the more critical.

The public expects[5] government to be an exemplar of responsible AI use. Yet the very policy designed to ensure transparency seems to be ignored by many agencies.

With the government also signalling a reluctance[6] to pass economy-wide AI rules, good practice in government could also encourage action from a disoriented private sector. A recent study[7] found 78% of corporations are “aware” of responsible AI practices, but only 29% have actually “implemented” them.

Transparency statements

The transparency statement requirement is the key binding obligation under the Digital Transformation Agency’s policy[8] for the responsible use of AI in government.

Agencies must also appoint an “accountable [AI] official” who is meant to be responsible for AI use. The transparency statements are supposed to be clear, consistent, and easy to find – ideally linked from the agency’s homepage.

In our research, conducted in collaboration with the Office of the Australian Information Commissioner[9], we sought to identify these statements, using a combination of automated combing through websites, targeted Google searches, and manual inspection of the list of federal entities facilitated by the information commissioner. This included both agencies and departments strictly bound by the policy and those invited to comply voluntarily.

But we found only a few statements were accessible from the agency’s landing page. Many were buried deep in subdomains or required complex manual searching. Among agencies for which publishing a statement was recommended, rather than required, we struggled to find any.

More concerningly, there were many for which we could not find the statement even where it was required. This may just be a technical failure, but given the effort we put in, it suggests a policy failure.

A toothless requirement

The transparency statement requirement[10] is binding in theory but toothless in practice. There are no penalties for agencies that fail to comply. There is also no open central register to track who has or has not published a statement.

The result is a fragmented, inconsistent landscape that undermines the very trust the policy was meant to build. And the public has no way to understand – or challenge – how AI is being used in decisions that affect their lives.

How other countries do it

In the United Kingdom, the government established a mandatory AI register. But as the Guardian[11] reported in late 2024, many departments failed to list their AI use, despite the legal requirement to do so.

The situation seems to have slightly improved this year, but still many high-risk AI systems identified by UK civil society groups[12] are still not published on the UK government’s own register.

The United States has taken a firmer stance. Despite anti-regulation rhetoric from the White House, the government has so far maintained its binding commitments[13] to AI transparency and mitigation of risk.

Federal agencies are required to assess and publicly register their AI systems. If they fail to do so, the rules say they must stop using them.

Towards responsible use of AI

In the next phase of our research, we will analyse the content of the transparency statements we did find.

Are they meaningful? Do they disclose risks, safeguards and governance structures? Or are they vague and perfunctory? Early indications suggest wide variation in quality.

If governments are serious about responsible AI, they must enforce their own policies. If determined university researchers cannot easily find the statements – even assuming they are somewhere deep on the website – that cannot be called transparency.

The authors wish to thank Shuxuan (Annie) Luo for her contribution to this research.

References

  1. ^ by February 2025 (www.digital.gov.au)
  2. ^ stated goal (www.digital.gov.au)
  3. ^ is already low (mbs.edu)
  4. ^ shortcoming by the Robodebt royal commission (robodebt.royalcommission.gov.au)
  5. ^ expects (www.digital.gov.au)
  6. ^ reluctance (ministers.treasury.gov.au)
  7. ^ recent study (www.fifthquadrant.com.au)
  8. ^ policy (www.digital.gov.au)
  9. ^ Office of the Australian Information Commissioner (www.oaic.gov.au)
  10. ^ transparency statement requirement (www.digital.gov.au)
  11. ^ the Guardian (www.theguardian.com)
  12. ^ identified by UK civil society groups (publiclawproject.org.uk)
  13. ^ binding commitments (www.whitehouse.gov)

Read more https://theconversation.com/most-australian-government-agencies-arent-transparent-about-how-they-use-ai-266768

Global tensions rise as war risk pushes oil higher — what it means for Australians

A surge in geopolitical tensions is reshaping global markets and sending a clear signal to households and busine...

Times Magazine

How Decentralised Applications Are Reshaping Enterprise Software in Australia

Australian businesses are experiencing a quiet revolution in how they manage data, execute agreeme...

Bambu Lab P2S 3D Printer Review: High-End Performance Meets Everyday Usability

After a full month of hands-on testing, the Bambu Lab P2S 3D printer has proven itself to be one...

Nearly Half of Disadvantaged Australian Schools Run Libraries on Less Than $1000 a Year

A new national snapshot from Dymocks Children’s Charities reveals outdated books, no librarians ...

Growing EV popularity is leading to queues at fast chargers. Could a kerbside charger network help?

The war on Iran has made crystal clear how shaky our reliance on fossil fuels is. It’s no surpri...

TRUCKIES UNDER THE PUMP AS FUEL PRICES BECOME TWO THIRDS OF OPERATING COSTS FOR SOME BUSINESS OWNERS

As Australia’s fuel crisis continues, truck drivers across the nation are being hit hard despite t...

iPhone: What are the latest features in iOS 26.5 Beta 1?

Apple has quietly released the first developer beta of iOS 26.5, and while it may not be the hea...

The Times Features

The Decentralized DJ: How Play House is Rewriting the M…

The traditional music industry model is currently facing its most significant challenge since the ...

What Australians Use YouTube For

In Australia, YouTube is no longer just a video platform—it is infrastructure. It entertains, e...

Independent MPs warn NDIS funding cuts risk leaving vul…

Federal Independent MPs have called on the Albanese Government to provide greater transparency...

While Fuel Has Our Attention, There Are Many More Issue…

Australia is once again fixated on fuel. Petrol prices rise, headlines follow, political pressu...

Recent outbreaks highlight the risks of bacterial menin…

Outbreaks of bacterial meningococcal disease in England[1] and recent cases in students in New Z...

Nationals leader Matt Canavan promotes work from home t…

Nationals leader Matt Canavan has urged the embrace of work-from-home opportunities as a way to ...

Nearly Half of Disadvantaged Australian Schools Run Lib…

A new national snapshot from Dymocks Children’s Charities reveals outdated books, no librarians ...

Why a Skin Check Should Be Part of Your Gather Round Pl…

There’s a certain rhythm to AFL Gather Round - long days outdoors, packed stands, and a city that ...

Kinder Joy Hosts a Free Night in the Museum Dinosaur Ad…

This April, Kinder Joy invites families to step into a thrilling after-hours dinosaur adventure ...