News Release

New privacy-preserving robotic cameras obscure images beyond human recognition

The distorted images can still be used by robots to complete their tasks but do not provide a comprehensive visual representation that compromises privacy

Peer-Reviewed Publication

University of Sydney

What a normal camera sees VS what the privacy preserving camera sees

video: 

What a normal camera sees compared with what the researchers' privacy preserving camera sees.

view more 

Credit: University of Sydney and Queensland University of Technology

From robotic vacuum cleaners and smart fridges to baby monitors and delivery drones, the smart devices being increasingly welcomed into our homes and workplaces use vision to take in their surroundings, taking videos and images of our lives in the process. 

In a bid to restore privacy, researchers at the Australian Centre for Robotics at the University of Sydney and the Centre for Robotics (QCR) at Queensland University of Technology have created a new approach to designing cameras that process and scramble visual information before it is digitised so that it becomes obscured to the point of anonymity. 

Known as sighted systems, devices like smart vacuum cleaners form part of the “internet-of-things” – smart systems that connect to the internet. They can be at risk of being hacked by bad actors or lost through human error, their images and videos at risk of being stolen by third parties, sometimes with malicious intent. 

Acting as a "fingerprint,” the distorted images can still be used by robots to complete their tasks but do not provide a comprehensive visual representation that compromises privacy. 

“Smart devices are changing the way we work and live our lives, but they shouldn’t compromise our privacy and become surveillance tools,” said Adam Taras, who completed the research as part of his Honours thesis. 

“When we think of ‘vision’ we think of it like a photograph, whereas many of these devices don’t require the same type of visual access to a scene as humans do. They have a very narrow scope in terms of what they need to measure to complete a task, using other visual signals, such as colour and pattern recognition,” he said. 

The researchers have been able to segment the processing that normally happens inside a computer within  the optics and analogue electronics of the camera, which exists beyond the reach of attackers. 

“This is the key distinguishing point from prior work which obfuscated the images inside the camera’s computer – leaving the images open to attack,” said Dr Don Dansereau, Taras’ supervisor at the Australian Centre for Robotics. ”We go one level beyond to the electronics themselves, enabling a greater level of protection.”  

The researchers tried to hack their approach but were unable to reconstruct the images in any recognisable format. They have opened this task to the research community at large, challenging others to hack their method. 

“If these images were to be accessed by a third party, they would not be able to make much of them, and privacy would be preserved,” said Taras. 

Dr Dansereau said privacy was increasingly becoming a concern as more devices today come with built-in cameras, and with the possible increase in new technologies in the near future like parcel drones, which travel into residential areas to make deliveries.  

“You wouldn’t want images taken inside your home by your robot vacuum cleaner leaked on the dark web, nor would you want a delivery drone to map out your backyard. It is too risky to allow services linked to the web to capture and hold onto this information,” said Dr Dansereau.  

The approach could also be used to make devices that work in places where privacy and security are a concern, such as warehouses, hospitals, factories, schools and airports. 

The researchers hope to next build physical camera prototypes to demonstrate the approach in practice.  

“Current robotic vision technology tends to ignore the legitimate privacy concerns of end-users. This is a short-sighted strategy that slows down or even prevents the adoption of robotics in many applications of societal and economic importance. Our new sensor design takes privacy very seriously, and I hope to see it taken up by industry and used in many applications,” said Professor Niko Suenderhauf, Deputy Director of the QCR, who advised on the project. 

Professor Peter Corke, Distinguished Professor Emeritus and Adjunct Professor at the QCR who also advised on the project said: “Cameras are the robot equivalent of a person's eyes, invaluable for understanding the world, knowing what is what and where it is. What we don’t want is the pictures from those cameras to leave the robot’s body, to inadvertently reveal private or intimate details about people or things in the robot’s environment.” 

INTERVIEWS 
The researchers are available for interview at their robotic imaging lab.  

MEDIA ENQUIRIES 
Luisa Low | luisa.low@sydney.edu.au | +61 438 021 390 

RESEARCH 
The research, Inherently privacy-preserving vision for trustworthy autonomous systems: Needs and solutions, was published by the Journal of Responsible technology and can be found: 
https://doi.org/10.1016/j.jrt.2024.100079  

DECLARATION 

The authors declare no competing interests.


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.