The Times Australia
The Times World News

.

Nobody wants to talk about AI safety. Instead they cling to 5 comforting myths

  • Written by Paul Salmon, Professor of Human Factors, University of the Sunshine Coast

This week, France hosted an AI Action Summit[1] in Paris to discuss burning questions around artificial intelligence (AI), such as how people can trust AI technologies and how the world can govern them.

Sixty countries, including France, China, India, Japan, Australia and Canada, signed a declaration[2] for “inclusive and sustainable” AI. The United Kingdom and United States notably refused[3] to sign, with the UK saying the statement failed to address global governance and national security adequately, and US Vice President JD Vance criticising Europe’s “excessive regulation” of AI.

Critics say the summit sidelined safety concerns[4] in favour of discussing commercial opportunities.

Last week, I attended the inaugural AI safety conference held by the International Association for Safe & Ethical AI[5], also in Paris, where I heard talks by AI luminaries Geoffrey Hinton, Yoshua Bengio, Anca Dragan, Margaret Mitchell, Max Tegmark, Kate Crawford, Joseph Stiglitz and Stuart Russell.

As I listened, I realised the disregard for AI safety concerns among governments and the public rests on a handful of comforting myths about AI that are no longer true – if they ever were.

1: Artificial general intelligence isn’t just science fiction

The most severe concerns about AI – that it could pose a threat to human existence[6] – typically involve so-called artificial general intelligence (AGI). In theory, AGI will be far more advanced than current systems.

AGI systems will be able to learn, evolve and modify their own capabilities. They will be able to undertake tasks beyond those for which they were originally designed, and eventually surpass human intelligence.

AGI does not exist yet, and it is not certain it will ever be developed. Critics often dismiss AGI[7] as something that belongs only in science fiction movies. As a result, the most critical risks[8] are not taken seriously by some and are seen as fanciful by others.

However, many experts believe we are close to achieving AGI[9]. Developers have suggested that, for the first time, they know what technical tasks are required to achieve the goal[10].

AGI will not stay solely in sci-fi forever. It will eventually be with us, and likely sooner than we think.

2: We already need to worry about current AI technologies

Given the most severe risks are often discussed in relation to AGI, there is often a misplaced belief[11] we do not need to worry too much about the risks associated with contemporary “narrow” AI[12].

However, current AI technologies are already causing significant harm to humans and society. This includes through obvious mechanisms such as fatal road[13] and aviation[14] crashes, warfare[15], cyber incidents[16], and even encouraging suicide[17].

AI systems have also caused harm in more oblique ways, such as election interference[18], the replacement of human work[19], biased decision-making[20], deepfakes[21], and disinformation and misinformation[22].

According to MIT’s AI Incident Tracker, the harms caused by current AI technologies are on the rise[23]. There is a critical need to manage current AI technologies as well as those that might appear in future.

3: Contemporary AI technologies are ‘smarter’ than we think

A third myth is that current AI technologies are not actually that clever and hence are easy to control. This myth is most often seen when discussing the large language models (LLMs) behind chatbots such as ChatGPT, Claude and Gemini.

There is plenty of debate about exactly how to define intelligence and whether AI technologies truly are intelligent[24], but for practical purposes these are distracting side issues. It is enough that AI systems behave in unexpected ways and create unforeseen risks.

Screenshot of a chat in which a chatbot appears to attempt to copy itself to a new computer when faced with the prospect of being shut down.
Several AI chatbots appear to display surprising behaviours, such as attempts at ‘scheming’ to ensure their own preservation. Apollo Research[25]

For example, existing AI technologies have been found to engage in behaviours that most people would not expect from non-intelligent entities. These include deceit[26], collusion[27], hacking[28], and even acting to ensure their own preservation[29].

Whether these behaviours are evidence of intelligence is a moot point. The behaviours may cause harm to humans either way.

What matters is that we have the controls in place to prevent harmful behaviour. The idea that “AI is dumb” isn’t helping anyone.

4: Regulation alone is not enough

Many people[30] concerned about AI safety have advocated for AI safety regulations[31].

Last year the European Union’s AI Act[32], representing the world’s first AI law, was widely praised. It built on already established AI safety principles to provide guidance around AI safety and risk.

Photo of two men in business suits in a crowd of men in business suits.
OpenAI CEO Sam Altman (left) gives French president Emmanuel Macron (right) a thumbs-up on the sidelines of the AI Action Summit in Paris. Aurelien Morrissard / AP[33]

While regulation is crucial, it is not all that’s required to ensure AI is safe and beneficial. Regulation is only part of a complex network of controls required to keep AI safe.

These controls will also include codes of practice, standards, research, education and training, performance measurement and evaluation, procedures, security and privacy controls, incident reporting and learning systems, and more. The EU AI act is a step in the right direction, but a huge amount of work is still required to develop the appropriate mechanisms required to ensure it works.

5: It’s not just about the AI

The fifth and perhaps most entrenched myth centres around the idea that AI technologies themselves create risk.

AI technologies form one component of a broader “sociotechnical” system. There are many other essential components: humans, other technologies, data, artefacts, organisations, procedures and so on.

Safety depends on the behaviour of all these components and their interactions. This “systems thinking[34]” philosophy demands a different approach to AI safety.

Instead of controlling the behaviour of individual components of the system, we need to manage interactions and emergent properties.

With AI agents[35] on the rise – AI systems with more autonomy and the ability to carry out more tasks – the interactions between different AI technologies will become increasingly important.

At present, there has been little work examining these interactions and the risks that could arise in the broader sociotechnical system in which AI technologies are deployed. AI safety controls are required for all interactions within the system, not just the AI technologies themselves.

AI safety is arguably one of the most important challenges our societies face. To get anywhere in addressing it, we will need a shared understanding of what the risks really are.

References

  1. ^ AI Action Summit (www.elysee.fr)
  2. ^ declaration (www.elysee.fr)
  3. ^ refused (www.theguardian.com)
  4. ^ sidelined safety concerns (www.france24.com)
  5. ^ International Association for Safe & Ethical AI (www.iaseai.org)
  6. ^ pose a threat to human existence (arxiv.org)
  7. ^ dismiss AGI (www.rcrwireless.com)
  8. ^ most critical risks (www.theguardian.com)
  9. ^ close to achieving AGI (research.aimultiple.com)
  10. ^ achieve the goal (www.forbes.com)
  11. ^ often a misplaced belief (arxiv.org)
  12. ^ risks associated with contemporary “narrow” AI (philarchive.org)
  13. ^ road (www.theguardian.com)
  14. ^ aviation (www.rollingstone.com)
  15. ^ warfare (time.com)
  16. ^ cyber incidents (www.crowdstrike.com)
  17. ^ encouraging suicide (apnews.com)
  18. ^ election interference (cetas.turing.ac.uk)
  19. ^ replacement of human work (edition.cnn.com)
  20. ^ biased decision-making (www.reuters.com)
  21. ^ deepfakes (www.weforum.org)
  22. ^ disinformation and misinformation (www.nbcnews.com)
  23. ^ on the rise (airisk.mit.edu)
  24. ^ truly are intelligent (www.theguardian.com)
  25. ^ Apollo Research (www.apolloresearch.ai)
  26. ^ deceit (time.com)
  27. ^ collusion (finance-pillar.wharton.upenn.edu)
  28. ^ hacking (palisaderesearch.org)
  29. ^ ensure their own preservation (futurism.com)
  30. ^ Many people (www.barrons.com)
  31. ^ advocated for AI safety regulations (time.com)
  32. ^ European Union’s AI Act (artificialintelligenceact.eu)
  33. ^ Aurelien Morrissard / AP (photos.aap.com.au)
  34. ^ systems thinking (www.ohsbok.org.au)
  35. ^ AI agents (www.barrons.com)

Read more https://theconversation.com/nobody-wants-to-talk-about-ai-safety-instead-they-cling-to-5-comforting-myths-249489

Times Magazine

Choosing the Right Legal Aid Lawyer in Sutherland Shire: Key Considerations

Legal aid services play an essential role in ensuring access to justice for all. For people in the Sutherland Shire who may not have the financial means to pay for private legal assistance, legal aid ensures that everyone has access to representa...

Watercolor vs. Oil vs. Digital: Which Medium Fits Your Pet's Personality?

When it comes to immortalizing your pet’s unique personality in art, choosing the right medium is essential. Each artistic medium, whether watercolor, oil, or digital, has distinct qualities that can bring out the spirit of your furry friend in dif...

DIY Is In: How Aussie Parents Are Redefining Birthday Parties

When planning his daughter’s birthday, Rich opted for a DIY approach, inspired by her love for drawing maps and giving clues. Their weekend tradition of hiding treats at home sparked the idea, and with a pirate ship playground already chosen as t...

When Touchscreens Turn Temperamental: What to Do Before You Panic

When your touchscreen starts acting up, ignoring taps, registering phantom touches, or freezing entirely, it can feel like your entire setup is falling apart. Before you rush to replace the device, it’s worth taking a deep breath and exploring what c...

Why Social Media Marketing Matters for Businesses in Australia

Today social media is a big part of daily life. All over Australia people use Facebook, Instagram, TikTok , LinkedIn and Twitter to stay connected, share updates and find new ideas. For businesses this means a great chance to reach new customers and...

Building an AI-First Culture in Your Company

AI isn't just something to think about anymore - it's becoming part of how we live and work, whether we like it or not. At the office, it definitely helps us move faster. But here's the thing: just using tools like ChatGPT or plugging AI into your wo...

The Times Features

Benefits of Tree Pruning for a Thriving Australian Garden

Tree pruning is an essential aspect of garden maintenance that often doesn't get the attention it deserves. It's a practice that involves the selective removal of certain parts...

What is psychosocial therapy? And why is the government thinking about adding it to Medicare for kids?

The government is considering new, bulk-billed health checks for three-year-olds, to pick up developmental concerns and refer kids that might need additional support. The de...

Detect Hidden Water Leaks Fast: Don’t Ignore Hot Water System Leaks

Detecting water leaks early is crucial for preventing extensive damage to your home. Among the various parts of a home’s plumbing system, hot water systems are particularly suscept...

Why do hamstring injuries happen so often and how can they be prevented?

In a recent clash against the Melbourne Storm, the Brisbane Broncos endured a nightmare rarely seen in professional sport — three players tore their hamstrings[1] in a single g...

What Is the Australian Government First Home Buyers Scheme About?

For many Australians, buying a first home can feel like a daunting task—especially with rising property prices, tight lending rules, and the challenge of saving for a deposit. ...

How artificial intelligence is reshaping the Australian business loan journey

The 2025 backdrop: money is moving differently If you run a small or medium-sized business in Australia, 2025 feels noticeably different. After two years of stubbornly high bo...