Researchers develop 'vaccine' against attacks on machine learning

CSIRO Head Office

Friday, 21 June, 2019

Researchers develop 'vaccine' against attacks on machine learning

Algorithms 'learn' from the data they are trained on to create a machine learning model that can perform a given task effectively without needing specific instructions, such as making predictions or accurately classifying images and emails. These techniques are already used widely — for example, to identify spam emails, diagnose diseases from X-rays, predict crop yields — and will soon drive our cars.

While the technology holds enormous potential to positively transform our world, artificial intelligence and machine learning are vulnerable to adversarial attacks, a technique employed to fool machine learning models through the input of malicious data causing them to malfunction.

Dr Richard Nock, machine learning group leader at CSIRO's Data61, said that by adding a layer of noise (an adversary) over an image, attackers can deceive machine learning models into misclassifying the image.

"Adversarial attacks have proven capable of tricking a machine learning model into incorrectly labelling a traffic stop sign as a speed sign, which could have disastrous effects in the real world," he said. "Our new techniques prevent adversarial attacks using a process similar to vaccination.

"We implement a weak version of an adversary, such as small modifications or distortion to a collection of images, to create a more difficult training data set. When the algorithm is trained on data exposed to a small dose of distortion, the resulting model is more robust and immune to adversarial attacks."

In a research paper accepted at the 2019 International Conference on Machine Learning (ICML), the researchers also demonstrate that the 'vaccination' techniques are built from the worst possible adversarial examples, and can therefore withstand very strong attacks.

Adrian Turner, CEO at CSIRO's Data61, said this research is a significant contribution to the growing field of adversarial machine learning.

"Artificial intelligence and machine learning can help solve some of the world's greatest social, economic and environmental challenges, but that can't happen without focused research into these technologies," Turner said. "The new techniques against adversarial attacks developed at Data61 will spark a new line of machine learning research and ensure the positive use of transformative AI technologies."

CSIRO recently invested $19 million into an Artificial Intelligence and Machine Learning Future Science Platform, to target AI-driven solutions for areas including food security and quality, health and wellbeing, sustainable energy and resources, resilient and valuable environments, and Australian and regional security.

Data61 also led the development of an AI ethics framework for Australia, released by the Australian Government for public consultation in April 2019.

Image: ©stock.adobe.com/au/peshkova

Related News

Fortescue commissions first battery electric locomotives

Fortescue has announced that it has commenced commissioning of two new battery electric...

Government seeks input on wind and transmission infrastructure

DISR is seeking industry input on the domestic manufacturing of wind turbines and transmission...

Additive manufacturing unlocks power for space and defence missions

An Australian-led additive manufacturing research project is set to transform how long-duration...


  • All content Copyright © 2026 Westwick-Farrow Pty Ltd