why we shouldn't ban the police from using it altogether
- Written by Asress Adimi Gikay, Senior Lecturer in AI, Disruptive Innovation and Law, Brunel University London
The UK police are being accused of breaking ethical standards by using live facial recognition technology to help fight crime. A recent report[1] by the University of Cambridge into trials of the technology by forces in London and south Wales was particularly concerned about the “lack of robust redress” for anyone suffering harm. It spoke of the need to “protect human rights and improve accountability” before facial recognition is used more widely.
The Cambridge team wants a broad ban on police using the technology, and they are not alone. UK civil liberties group Big Brother Watch has been running a “stop facial recognition[2]” campaign as the government mulls how to regulate AI technologies[3]. Meanwhile, 12 NGOs recently called on[4] EU legislators to completely ban it, along with various other forms of biometric identification, in their upcoming AI Act[5].
Simply banning this technology would be a mistake, however. In my view, there’s a good case for a more measured approach.
Growing police use
The police forces in London and south Wales appear to be the only two in the UK currently using live facial recognition, which uses artificial intelligence software[6] to compare an individual’s digital facial image with an existing facial image to estimate similarity. Manchester Police trialled it but were forced to pause[7] by the surveillance camera commissioner[8] in 2018 for not obtaining the necessary approvals.
In 2020 an appellate court also ruled against[9] south Wales’ use of the technology, concluding the force’s legal framework for deployment effectively gave them unlimited discretion to do so. It made no difference to the court that the police had notified the public (known as overt operational deployment).
Despite this ruling, facial recognition can still broadly be used by police, although numerous other forces[10] have said they are not[11] doing so at present.
The London Metropolitan Police increasingly use facial recognition to locate missing persons, suspects, witnesses[13] and victims. They have scanned individuals’ faces in city squares and at public events, using a facial recognition camera[14] typically placed on a police vehicle or street pole. The public are alerted[15] to the deployment through notices as they enter the recognition zone – unless that compromises policing tactics or deployment is urgent.
Between February 2020 and July 2022, the Met deployed the techology in eight locations including Piccadilly Circus[16]. They are estimated[17] to have viewed more than 150,000 faces, leading to nine arrests but also eight occasions where they targeted the wrong person.
The pros and cons
Facial recognition has evolved in recent years, for instance to work in real time, but inaccuracies and errors remain. In New Jersey, 228 wrongful arrests[18] were reportedly made using (non-real time) facial recognition between January 2019 and April 2021. One black American[19] spent 11 days in jail after being wrongly identified. False identifications can also lead to everything from missed flights to distressing police interrogations.
Specific groups are disproportionately affected. A 2019 US study[20] found that women are two-to-five times more likely to be falsely identified, while the risks are ten-to-100 times greater for black and Asian faces than white ones. Given that police already disproportionately stop and search[21] ethnic minorities, this shortcoming in the technology could potentially even be used to sustain such practices.
BradleyStearn[22]Another risk is that police covertly install facial recognition cameras permanently. This could help the state to crack down on public protests, for example. There is already a pending legal challenge against Russia[23] before the European Court of Human Rights over such practices, and fear of state surveillance is one reason why many want this technology banned.
Nonetheless, facial recognition has its benefits. It can help[24] police to find serious criminals, including terrorists, not to mention missing children[25] and people at risk of harming themselves or others.
Like it or not, we also live under colossal corporate surveillance capitalism already. The UK and US[26] have among the most installed CCTV cameras in the world. London residents are filmed 300 times[27] a day on average, and police can usually use the data without a search warrant. As if that wasn’t bad enough, big tech companies know almost everything[28] personal about us. Worrying about live facial recognition is inconsistent with our tolerance of all this surveillance.
A better approach
Instead of an outright ban, even of covert facial recognition, I’m in favour of a statutory law to clarify when this technology can be deployed. For one thing, police in the UK can currently use it to track people on their watchlists, but this can include even those charged with minor crimes. There are also no uniform criteria for deciding who can be listed.
Under the EU’s proposed law[29], facial recognition could only be deployed against those suspected of crimes carrying a maximum sentence of upwards of three years. That would appear to be a reasonable cut-off.
Secondly, a court or similar independent body should always have to authorise deployment, including assessing whether it would be proportionate to the police objective in question. In the Met, authorisation currently has to come from a police officer ranked superintendent or higher[30], and they do[31] have to make a call[32] on proportionality – but this should not be a police decision.
We also need clear, auditable ethical standards for what happens during and after the technology is deployed. Images of wrongly identified people should be deleted immediately, for instance. Unfortunately, Met policy on this is unclear at present. The Met is trying to use the technology responsibly in other respects, but this is not enough in itself.
Last but not least, the potential for discrimination[33] should be tackled by legally requiring developers to train the AI on a diverse enough range of communities to meet a minimum threshold. This sort of framework should allow society to enjoy the benefits of live facial recognition without the harms. Simply banning something that requires a delicate balancing of competing interests is the wrong move entirely.
References
- ^ A recent report (www.mctd.ac.uk)
- ^ stop facial recognition (bigbrotherwatch.org.uk)
- ^ regulate AI technologies (www.gov.uk)
- ^ recently called on (edri.org)
- ^ AI Act (eur-lex.europa.eu)
- ^ artificial intelligence software (ico.org.uk)
- ^ forced to pause (www.express.co.uk)
- ^ surveillance camera commissioner (www.gov.uk)
- ^ ruled against (www.judiciary.uk)
- ^ other forces (www.psni.police.uk)
- ^ they are not (www.gmp.police.uk)
- ^ Trismegist San (www.shutterstock.com)
- ^ witnesses (www.met.police.uk)
- ^ facial recognition camera (www.judiciary.uk)
- ^ public are alerted (www.met.police.uk)
- ^ Piccadilly Circus (www.met.police.uk)
- ^ are estimated (www.met.police.uk)
- ^ 228 wrongful arrests (incidentdatabase.ai)
- ^ black American (edition.cnn.com)
- ^ A 2019 US study (nvlpubs.nist.gov)
- ^ stop and search (www.theguardian.com)
- ^ BradleyStearn (www.shutterstock.com)
- ^ legal challenge against Russia (www.hrw.org)
- ^ It can help (www.securityindustry.org)
- ^ missing children (www.reuters.com)
- ^ UK and US (papltd.co.uk)
- ^ 300 times (www.theguardian.com)
- ^ know almost everything (guardian.ng)
- ^ proposed law (eur-lex.europa.eu)
- ^ superintendent or higher (www.met.police.uk)
- ^ they do (www.met.police.uk)
- ^ make a call (www.met.police.uk)
- ^ potential for discrimination (news.stanford.edu)