[Science] UK court backs police use of face recognition, but fight isn’t over – AI

0
187
[Science] UK court backs police use of face recognition, but fight isn’t over – AI


By Donna Lu South Wales Police have been testing facial recognition technologyMatthew Horwood/Getty A man from Cardiff, UK, says the police breached his human rights when they used facial recognition technology, but today a court ruled that the police’s actions were lawful. That is, however, hardly the end of the matter. South Wales Police has been trialling automated facial recognition (AFR) technology since April 2017. Other forces around the country are trialling similar systems, including London’s Metropolitan Police. Bridges may have been snapped during a pilot called AFR Locate. This involved taking pictures of people’s faces from live CCTV feeds and processing them in real time to extract biometric information. If a face matched a database of people on a police watchlist, they would be flagged. Otherwise, the data was deleted immediately. The force used this technology around 50 times at public events between May 2017 and April 2019, including at football matches and concerts. An estimated 500,000 faces were scanned. Advertisement At the hearing, it emerged that the police could not now check whether images of Bridges had been taken. If they had been, his pictures would have been deleted. The court found that AFR Locate’s collection and processing of data met the conditions set out in the Human Rights Act and in UK data protection legislation. The ruling is a blow for privacy campaigners. But plenty of people still seem to be worried about facial recognition technology. Read more: A UK police force is dropping tricky cases on advice of an algorithm The Ada Lovelace Institute, a technology research firm, today released the results of a survey on public opinion on facial recognition. Completed with YouGov, it found that the majority of people in the UK want government restrictions on police use of AFR. This shows that the court ruling is out of step with public expectations, says the institute’s director Carly Kind. The Information Commissioner’s Office (ICO), the UK data regulation regulator, has also voiced concerns. It recently conducted an investigation into the first police pilots of facial recognition. And in a statement today, it warned that facial recognition has the potential to “undermine rather than enhance confidence in the police”. “We will now consider the court’s findings in finalising our recommendations and guidance to police forces,” said an ICO spokesperson. For his part, Bridges intends to appeal the ruling. “This disappointing judgement does not reflect the very serious threat that facial recognition poses to our rights and freedoms,” said Bridges’ lawyer Megan Goulding. How to live with intelligent machines: Joanna Bryson at New Scientist Live Today’s judgement only applies to the police, but facial recognition technology is used by other organisations and in various ways. “The issue remains that a good many people are adopting this new technology without… ensuring that they comply appropriately with the requirements of the legislation,” says Will Richmond-Coggan, a data protection litigator at Freeths, a law firm. “That is where privacy is being impacted negatively,” he says. The ruling also does not address whether the technology should be used, says Kind. “Should this technology be lawful?” says Kind. “Should police be allowed to use it to monitor people on a day to day basis? Should supermarkets, schools and employers be allowed to use it?” In light of this, Kind says it would be good for the government to engage with the public more broadly on this issue. “We believe the government has a responsibility to reconcile [public expectations and existing regulation], including by undertaking a review of the legal framework governing not only facial recognition technologies but all biometrics technologies,” she says. More on these topics: artificial intelligence crime law

Leave a Reply