Two legal challenges have been launched against police forces in south Wales and London over their use of automated facial recognition (AFR) technology on the grounds the surveillance is unregulated and violates privacy. The claims are backed by the human rights organisations Liberty and Big Brother Watch following complaints about biometric checks at the Notting Hill carnival, on Remembrance Sunday, at demonstrations and in high streets. Liberty is supporting Ed Bridges, a Cardiff resident, who has written to the chief constable of South Wales police alleging he was tracked at a peaceful anti-arms protest and while out shopping. Big Brother Watch is working with the Green party peer Jenny Jones who has written to the home secretary, Sajid Javid, and the Metropolitan police commissioner, Cressida Dick, urging them to halt deployment of the "dangerously authoritarian" technology. If the forces do not stop using AFR systems then legal action will follow in the high court, the letters said.
If you're reading this in the United States, there's a 50 percent chance that a photo of your face is in at least one database used in police facial-recognition systems. Police departments in nearly half of U.S. states can use facial-recognition software to compare surveillance images with databases of ID photos or mugshots. Some departments only use facial-recognition to confirm the identity of a suspect who's been detained; others continuously analyze footage from surveillance cameras to determine exactly who is walking by at any particular moment. Altogether, more than 117 million American adults are subject to face-scanning systems. These findings were published Tuesday in a report from Georgetown Law's Center for Privacy and Technology.
Civil liberties group Big Brother Watch has launched a legal challenge against the use of automatic facial recognition technology by London's Metropolitan Police force. The privacy campaigners described the Met's "China-style" facial recognition system, which uses AI software to match people's faces to a criminal database, as "dangerously authoritarian." "Facial recognition is the latest Orwellian mass surveillance tool to be lawlessly rolled out by the state," Big Brother Watch writes on the campaign website. "These real-time facial recognition cameras are biometric checkpoints, identifying members of the public without their knowledge. Police have begun feeding secret watchlists to the cameras, containing not only criminals but suspects, protesters, football fans and innocent people with mental health problems."
The last day of January 2019 was sunny, yet bitterly cold in Romford, east London. Shoppers scurrying from retailer to retailer wrapped themselves in winter coats, scarves and hats. The temperature never rose above three degrees Celsius. For police officers positioned next to an inconspicuous blue van, just metres from Romford's Overground station, one man stood out among the thin winter crowds. The man, wearing a beige jacket and blue cap, had pulled his jacket over his face as he moved in the direction of the police officers.
OSAKA – Despite advances in facial recognition technology, the police in Osaka still rely on pure skill to find fugitives, with investigators using only their memory to arrest dozens of wanted criminals every year. While other police forces in the world have "super recognizer" units that hunt down fugitives, the so-called miatari (look and hit) technique used in Osaka has contributed to the arrests of over 4,000 criminals in Japan since the Osaka Prefectural Police introduced it as a formal investigative method in November 1978. There has not been a single wrongful arrest. "The best part of this method is being able to detect fugitives who are hard to find in normal investigations," said a senior investigator in Osaka. He says a forensic analysis is an imperative part of criminal investigations, but "we want to pass on the tradition because our job is to make sure no one gets away with a crime."