New privacy-preserving robotic cameras obscure images beyond human recognition

5 April 2024
Researchers hope their approach will give privacy to at-home robotic device users
Robotic researchers have developed a new approach to designing cameras, which they say could help protect the images and data collected by smart home devices and internet-of-things technology.

What a normal camera sees compared with what the privacy preserving camera sees. Image: University of Sydney and Queensland University of Technology

From robotic vacuum cleaners and smart fridges to baby monitors and delivery drones, the smart devices being increasingly welcomed into our homes and workplaces use vision to take in their surroundings, taking videos and images of our lives in the process. 

In a bid to restore privacy, researchers at the Australian Centre for Robotics at the University of Sydney and the QUT Centre for Robotics (QCR) at Queensland University of Technology have created a new approach to designing cameras that process and scramble visual information before it is digitised so that it becomes obscured to the point of anonymity. 

Known as sighted systems, devices like autonomous vacuum cleaners form part of the “internet-of-things” – smart systems that connect to the internet. They can be at risk of being hacked by bad actors or lost through human error, their images and videos at risk of being stolen by third parties, sometimes with malicious intent. 

Acting as a "fingerprint,” the distorted images can still be used by robots to complete their tasks but do not provide a comprehensive visual representation that compromises privacy. 

“Smart devices are changing the way we work and live our lives, but they shouldn’t compromise our privacy and become surveillance tools,” said Adam Taras, who completed the research as part of his Honours thesis. 

“When we think of ‘vision’ we think of it like a photograph, whereas many of these devices don’t require the same type of visual access to a scene as humans do. They have a very narrow scope in terms of what they need to measure to complete a task, using other visual signals, such as colour and pattern recognition,” he said. 

What a normal camera sees VS what the privacy preserving camera sees

The researchers have been able to segment the processing that normally happens inside a computer within the optics and analogue electronics of the camera, which exists beyond the reach of attackers. 

“This is the key distinguishing point from prior work which obfuscated the images inside the camera’s computer – leaving the images open to attack,” said Dr Don Dansereau, Taras’ supervisor at the Australian Centre for Robotics and Digital Sciences Initiative. ”We go one level beyond to the electronics themselves, enabling a greater level of protection.”  

The researchers tried to hack their approach but were unable to reconstruct the images in any recognisable format. They have opened this task to the research community at large, challenging others to hack their method. 

“If these images were to be accessed by a third party, they would not be able to make much of them, and privacy would be preserved,” said Taras. 

Dr Dansereau said privacy was increasingly becoming a concern as more devices today come with built-in cameras, and with the possible increase in new technologies in the near future like parcel drones, which travel into residential areas to make deliveries.  

“You wouldn’t want images taken inside your home by your robot vacuum cleaner leaked on the dark web, nor would you want a delivery drone to map out your backyard. It is too risky to allow services linked to the web to capture and hold onto this information,” said Dr Dansereau.  

The approach could also be used to make devices that work in places where privacy and security are a concern, such as warehouses, hospitals, factories, schools and airports. 

The researchers hope to next build physical camera prototypes to demonstrate the approach in practice.  

“Current robotic vision technology tends to ignore the legitimate privacy concerns of end-users. This is a short-sighted strategy that slows down or even prevents the adoption of robotics in many applications of societal and economic importance. Our new sensor design takes privacy very seriously, and I hope to see it taken up by industry and used in many applications,” said Professor Niko Suenderhauf, Deputy Director of the QCR, who advised on the project. 

Professor Peter Corke, Distinguished Professor Emeritus and Adjunct Professor at the QCR who also advised on the project said: “Cameras are the robot equivalent of a person's eyes, invaluable for understanding the world, knowing what is what and where it is. What we don’t want is the pictures from those cameras to leave the robot’s body, to inadvertently reveal private or intimate details about people or things in the robot’s environment.” 


The research, Inherently privacy-preserving vision for trustworthy autonomous systems: Needs and solutions, was published by the Journal of Responsible technology.


The authors declare no competing interests.

Luisa Low

Media and PR Adviser (Engineering & IT)

Related news