Artificial intelligence in Australia needs to get ethical, so we have a plan


By Emma Schleiger, CSIRO and Stefan Hajkowicz, Data61*
Thursday, 09 May, 2019


Artificial intelligence in Australia needs to get ethical, so we have a plan

The question of whether technology is good or bad depends on how it’s developed and used. Nowhere is that more topical than in technolgies using artificial intelligence.

When developed and used appropriately, artificial intelligence (AI) has the potential to transform the way we live, work, communicate and travel.

New AI-enabled medical technologies are being developed to improve patient care. There are persuasive indications that autonomous vehicles will improve safety and reduce the road toll. Machine learning and automation are streamlining workflows and allowing us to work smarter.

Around the world, AI-enabled technology is increasingly being adopted by individuals, governments, organisations and institutions. But along with the vast potential to improve our quality of life, comes a risk to our basic human rights and freedoms.

Appropriate oversight, guidance and understanding of the way AI is used and developed in Australia must be prioritised.

AI gone wild may conjure images of The Terminator and Ex Machina movies, but it is much simpler, fundamental issues that need to be addressed at present, such as:

  • how data is used to develop AI
  • whether an AI system is being used fairly
  • in which situations should we continue to rely on human decision-making?

We have an AI ethics plan

That’s why, in partnership with government and industry, we’ve developed an ethics framework for AI in Australia. The aim is to catalyse the discussion around how AI should be used and developed in Australia.

The ethical framework looks at various case studies from around the world to discuss how AI has been used in the past and the impacts that it has had. The case studies help us understand where things went wrong and how to avoid repeating past mistakes.

We also looked at what was being done around the world to address ethical concerns about AI development and use.

Based on the core issues and impacts of AI, eight principles were identified to support the ethical use and development of AI in Australia.

  1. Generates net benefits: The AI system must generate benefits for people that are greater than the costs.

  2. Do no harm: Civilian AI systems must not be designed to harm or deceive people and should be implemented in ways that minimise any negative outcomes.

  3. Regulatory and legal compliance: The AI system must comply with all relevant international, Australian local, state/territory and federal government obligations, regulations and laws.

  4. Privacy protection: Any system, including AI systems, must ensure people’s private data is protected and kept confidential and prevent data breaches that could cause reputational, psychological, financial, professional or other types of harm.

  5. Fairness: The development or use of the AI system must not result in unfair discrimination against individuals, communities or groups. This requires particular attention to ensure the “training data” is free from bias or characteristics which may cause the algorithm to behave unfairly.

  6. Transparency and explainability: People must be informed when an algorithm is being used that impacts them and they should be provided with information about what information the algorithm uses to make decisions.

  7. Contestability: When an algorithm impacts a person there must be an efficient process to allow that person to challenge the use or output of the algorithm.

  8. Accountability: People and organisations responsible for the creation and implementation of AI algorithms should be identifiable and accountable for the impacts of that algorithm, even if the impacts are unintended.

In addition to the core principles various toolkit items are identified in the framework that could be used to help support these principles. These include impact assessments, ongoing monitoring and public consultation.

A plan, what about action?

But principles and ethical goals can only go so far. At some point we will need to get to work on deciding how we are going to implement and achieve them.

There are various complexities to consider when discussing the ethical use and development of AI. The vast reach of the technology has potential to impact every facet of our lives.

AI applications are already in use across households, businesses and governments, most Australians are already being impacted by them.

There is a pressing need to examine the effects that AI has on the vulnerable and on minority groups, making sure we protect these individuals and communities from bias, discrimination and exploitation. (Remember Tay, the racist chatbot?)

There is also the fact that AI used in Australia will often be developed in other countries, so how do we ensure it adheres to Australian standards and expectations?

Your say

The framework explores these issues and forms some of Australia’s first steps on the journey towards the positive development and use of AI. But true progress needs input from stakeholders across government, business, academia and broader society.

That’s why ethical framework discussion paper is now open to public comment. You have until May 31, 2019, to have your say in Australia’s digital future.

With a proactive approach to the ethical development of AI, Australia can do more than just mitigate against any risks. If we can build AI for a fairer go, we can secure a competitive advantage as well as safeguard the rights of Australians.

The ConversationThis article is republished from The Conversation under a Creative Commons license. Read the original article.

Further reading:
  1. Salmon P, Hancock P, Carden T 2019, 'To protect us from the risks of advanced artificial intelligence, we need to act now', The Conversation, <https://theconversation.com/to-protect-us-from-the-risks-of-advanced-artificial-intelligence-we-need-to-act-now-107615>
  2. Agar N 2019, 'Careful how you treat today's AI: it might take revenge in the future', The Conversation, <https://theconversation.com/careful-how-you-treat-todays-ai-it-might-take-revenge-in-the-future-112611>

*Stefan Hajkowicz is a senior principal scientist, and Emma Schleiger is a research scientist, in strategic foresight in Data61 at CSIRO. They are part of a team of researchers and consultants working on scenario planning, megatrends analysis, risk analysis, decision support and strategy problems.

Image: © stock.adobe.com/au/bizoo_n
Related Articles

Climate-friendly electricity from ammonia

Researchers the Fraunhofer Institute have developed a high-temperature fuel cell stack that can...

Digitalised, sustainable battery cell production

German researchers have developed a flexible winding system for battery cells that is embedded in...

Expired deadline threatens critical infrastructure as compliance lags

The deadline for achieving cybersecurity framework alignment for the SOCI Act expired on 17...


  • All content Copyright © 2024 Westwick-Farrow Pty Ltd