Data watchdog warns police forces over use of live facial recognition
The UK data watchdog has warned police forces testing live facial recognition technology that there remains significant privacy and data protection issues which must be addressed.
South Wales Police and the Met Police have been trialling live facial recognition (LFR) technology as a possible way to reduce crime, but the move has been divisive.
Facial recognition technology maps faces in a crowd by measuring the distance between facial features, then compares results with a “watch list” of images, which can include suspects, missing people and persons of interest.
The Information Commissioner’s Office (ICO) said it understood the legitimate aims of the controversial system but told police forces that they need to do more to demonstrate their compliance with data protection law, including in how watch lists are compiled and what images are used.
“We understand the purpose is to catch criminals,” Information Commissioner Elizabeth Denham said.
“But these trials also represent the widespread processing of biometric data of thousands of people as they go about their daily lives.
“And that is a potential threat to privacy that should concern us all.
“I believe that there needs to be demonstrable evidence that the technology is necessary, proportionate and effective considering the invasiveness of LFR.”
The watchdog – which is currently investigating the use of live facial recognition – also raised concern about technological bias, which can see more false positive matches from certain ethnic groups.
In May, a court heard from an activist who claimed that the use of facial recognition technology by the police is a breach of human rights law.
Lawyers representing Ed Bridges claim South Wales Police violated his privacy and data protection rights by processing an image taken of him in public, action which he says caused him “distress”.
South Wales Police argue that use of facial recognition technology does not infringe the privacy or data protection rights of Mr Bridges as it is used in the same way as photographing a person’s activities in public, and it does not retain the data of those not on its watch list.
But it does keep CCTV images from the scanning process for up to 31 days.
The ICO said the resulting judgment will form an important part of its investigation, and it will need to consider it before publishing its findings.
“In recent months we have widened our focus to consider the use of LFR in public spaces by private sector organisations, including where they are partnering with police forces,” Ms Denham continued.
“We’ll consider taking regulatory action where we find non-compliance with the law.”
According to a study by the University of Essex (see https://twitter.com/
Said Javvad Malik, security awareness advocate at KnowBe4:
“Recognition technologies will always have some degree of false positive and false negatives. Facial recognition, especially outside of controlled environments is still very much a developing area of research and therefore it’s not surprising to hear of potentially low accuracy rates.
“Such technologies will need a lot of training and tweaking before they even get to close to an acceptable level where it can automatically be trusted. Until such a time, such technologies should always be used to augment and never to replace the human analyst. It is, therefore, important that organisations such as the police force don’t rely entirely on facial recognition to apprehend criminals.”