Last month, it acknowledged it had shared people’s pictures with the managers of the city’s King’s Cross Estate development.
It had previously denied the alliance.
In a new report, the Met added that it had only shared seven images and did not believe there had been similar arrangements with other private bodies.
It said the pictures were of “persons who had been arrested and charged/cautioned/reprimanded or given a formal warning” and had been provided by Camden Borough Police. The aim, it added, had been to “prevent crime, to protect vulnerable members of the community or to support the safety strategy”.
But it admitted that it had no record of whether the estate manager’s surveillance camera system had ever made facial matches of those involved, nor whether any police action had been taken as a result.
“The findings of this report need to be caveated by noting the limitations of technology which was not designed to be audited in this way, and the limitations of corporate memory,” it explained.
It also confirmed that the facial recognition system in use at the estate was that of the Japanese firm NEC – something the management firm, Argent, had repeatedly declined to divulge itself.
NEC’s systems have also been deployed by the Met as well as South Wales Police in their own trials of live facial recognition.
The Met also confirmed that the image-sharing arrangement had lasted between May 2016 and March 2018, and added that a new agreement had been put in place at the start of this year. However, it said no images had been shared under the new tie-up.
Argent had previously said it intended to launch a more advanced facial recognition system at the property but had yet to do so. It has since ditched the proposal.
The Met has again apologised for misinforming the mayor and members of London’s Assembly about its involvement and blamed the mistake on the agreement having been struck at a “borough level”.
London’s deputy mayor for policing and crime, Sophie Linden, added that she had been informed that the police service had written to all the city’s basic command units to make it “clear that there should be no local level agreements on the use of live facial recognition”.
In a statement given to the BBC she added: “The Mayor and I are committed to holding the Met to account on its use of facial recognition technology and that’s why the [Met’s] commissioner agrees with us that there will be no further deployment anywhere in London until all of the conditions set out in the London Policing Ethics Panel report have been addressed.”
Use of the tech was frozen earlier in the year before details of the King’s Cross partnership emerged.
British Transport Police had previously confirmed it too had shared images with Argent for use in its facial recognition system.
Privacy campaigners have raised concerns about the affair because it had not been apparent to the public that facial recognition scans were in use in what is a popular open-air site, home to shops, offices, education and leisure facilities.
Moreover, any formal tie-up between the police and an independent organisation concerning the use of facial recognition is supposed to be flagged to a surveillance camera commissioner. The watchdog previously blocked another similar arrangement involving police in Manchester and a local shopping centre.
487 total views, 1 views today