{"id":1279,"date":"2022-11-15T13:13:38","date_gmt":"2022-11-15T13:13:38","guid":{"rendered":"https:\/\/mpelembe.net\/?p=1279"},"modified":"2022-11-15T15:07:20","modified_gmt":"2022-11-15T15:07:20","slug":"facial-recognition-an-ethical-policing-tool","status":"publish","type":"post","link":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/","title":{"rendered":"Facial recognition: An ethical policing tool?"},"content":{"rendered":"<p>By Samuel Woodhams | Digital rights researcher and journalist<\/p>\n<table>\n<tbody>\n<tr>\n<td>Facial recognition technology made headlines again last month as researchers at the University of Cambridge, UK said that the\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-y\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-y\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw0zc6q3SujV1JxtQidUcR-c\">UK police\u2019s use of the technology was unethical and potentially unlawful<\/a>. The report from the Minderoo Centre for Technology and Democracy urged police to stop using live facial recognition (LFR) in public spaces and said trials by the Metropolitan Police and South Wales Police failed to meet the \u201cminimum legal and ethical standards.\u201d<!--more--><\/p>\n<p>The report highlighted what it called a lack of transparency, accountability, and oversight surrounding the use of the technology, while noting that it \u201cposes threats to human rights, especially for racialised and marginalised communities.\u201d<\/p>\n<p><strong><a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-j\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-j\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw06hDvV7MVQam90HKtEPHCm\">WATCH: Xinjiang to London: Chinese surveillance tech in the UK<\/a><\/strong><\/p>\n<p>The researchers are far from the first to highlight the harms of facial recognition technology. More than\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-t\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-t\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw1yKEb3kU8v9XSh5uoSCRf1\">200 civil society organisations<\/a>, along with the\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-i\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-i\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw308_3kxLWI7MdHebV2Udml\">UN High Commissioner on Human Rights<\/a>\u00a0and the\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-d\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-d\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw3qN6yaQOuXrDzHDfVqsOPW\">European Parliament<\/a>, have called for a ban on its use in public spaces. The risks are also becoming more well-known publicly thanks to the likes of\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-h\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-h\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw0q3gMA3JT92-O3MGngCE8R\">Coded Bias, a Netflix documentary<\/a>\u00a0that covers Joy Buolamwini\u2019s crucial work at the Massachusetts Institute of Technology (MIT) exposing the discriminatory impact and racial biases contained within the technology.<\/p>\n<p>But despite the criticism, facial recognition technology is still being used by police forces across the world, from\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-k\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-k\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw2IWZL2M9k0j1TrLEfcA2_s\">Colombia<\/a>\u00a0to\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-u\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-u\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw2kgmrhSex5YKMj0u-gZxOy\">China<\/a>, while meaningful regulation lags behind.<\/p>\n<p>&nbsp;<\/p>\n<p><img decoding=\"async\" class=\"CToWUd a6T lazyload\" tabindex=\"0\" data-src=\"https:\/\/ci4.googleusercontent.com\/proxy\/_NClRPhcKtJ9GyNCJUMrm4Q9KosMvLWBTQBxcoh_94Asu2tuunJzuu1S5L-u2jGJzNOB07pyPi7bMdGNFBAOpuDDRXvkYd4pPjxNRHPUDLHKEA34SD09dvrvi9g0lDQhVBtwDEMbkd7CbbaCUpisP2XkIqItVhdmwjWl-DNceon7OeL9L3SV-W6DBmHyKposHkeGFoUnv71ZJI1kpWhOkQhX_biDnTY=s0-d-e1-ft#https:\/\/prod.context.news\/dA\/22f91745e2d3fc254ff3b1336ac854dd\/picture\/2022-08-02T084457Z_966813116_RC27OV9TJ4RS_RTRMADP_3_USA-CHINA-TAIWAN.JPG\/640w\" alt=\"A surveillance camera is seen near a Chinese flag in Shanghai, China\" width=\"100%\" height=\"auto\" data-bit=\"iit\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" \/><\/p>\n<p class=\"m_-2244817513273254883imgCaption\">A surveillance camera is seen near a Chinese flag in Shanghai, China August 2, 2022. REUTERS\/Aly Song<\/p>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<table>\n<tbody>\n<tr>\n<td>&nbsp;<\/p>\n<p>In the UK, the legality of the tech has been called into question several times over the past few years. In 2020, the\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-o\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-o\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw11Z0JGAMl_A2rp69T1OES3\">South Wales Police\u2019s use of the technology was found to be unlawful<\/a>\u00a0by the Court of Appeal. The year before, the<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-b\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-b\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw0wH_lPecp5IDtDCDbG5SlO\">\u00a0Automated Facial Recognition Technology Bill<\/a>\u00a0was introduced to ban the use of the technology in public places, though it was never ratified.<\/p>\n<p>Despite this, some UK police forces remain steadfast in their support of the technology. Responding to the recent report from The University of Cambridge, Mark Travis of South Wales Police told the Guardian that \u201c<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-n\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-n\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw2FMwovKwqGtIN8dhmazgJa\">the whole aim of using facial recognition technology is to keep the public safe<\/a>.\u201d<\/p>\n<p>\u201cI believe the public will continue to support our use of all the available methods and technology to keep them safe, providing what we do is legitimate and proportionate,\u201d he said.<\/p>\n<p>But accurately gauging public opinion is difficult as there\u2019s been little meaningful public consultation on the topic. Meanwhile, mounting evidence indicates that it&#8217;s far less effective than some police forces would like to think. In fact, according to the Metropolitan Police\u2019s own figures, just nine arrests were made across five operations between February 2020 and July 2022, during which\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-p\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-p\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw3sLCilDxE27QwFcSpJS-SW\">more than 125,000 peoples\u2019 faces were scanned<\/a>.<\/p>\n<p>There has been some success in limiting police forces\u2019 use of facial recognition internationally. More than 20 cities in the United States have moved to\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-x\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-x\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw3QPXWVCsvyX41HOXAGUwwA\">restrict the use of the technology<\/a>, while constraints have also been introduced in\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-m\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-m\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw0wjO2tkEzVJWmsK6UrdASt\">Belgium<\/a>,\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-c\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-c\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw16r5OnZDmTManx4-gbtowG\">Morocco<\/a>\u00a0and\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-q\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-q\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw3pCUBT1ePCiMaMg1zY2Ljf\">Luxembourg<\/a>.<\/p>\n<p>However, it remains to be seen how long these regulations will stay in place.\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-a\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-a\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw02kqoo6MPOg3QxXJIwav8S\">Several US cities have already undone their restrictions<\/a>. And, even if meaningful regulations on LFR are passed, there are countless other technologies being developed that could come with similar problems.<\/p>\n<p>&nbsp;<\/p>\n<p><img decoding=\"async\" class=\"CToWUd a6T lazyload\" tabindex=\"0\" data-src=\"https:\/\/ci3.googleusercontent.com\/proxy\/YqJvBbDjTCzUOS8ZqRJA1XbuuBhNqCos-v3fx6IjMHIndwrg-D_-r3blVRzz51XN2lx0aA_kEtaLagtza-WBFXDe7yGelCx0B-FH7O4uqSmzuNU3wpqnmWZg-wj9h-rHNbynmeRe1wYbPtzcrtFiNH2g4NIlIUO-w5TI=s0-d-e1-ft#https:\/\/prod.context.news\/dA\/08009059254baa74f7f289227fb2be6c\/picture\/RTS38ZVG_Layout_-_Comp(1).jpg\/640w\" alt=\"A CCTV security surveillance camera overlooks a street as people walk following the spread of the coronavirus disease (COVID-19) in Beijing, China May 11, 2020. REUTERS\/Thomas Peter\" width=\"100%\" height=\"auto\" data-bit=\"iit\" src=\"data:image\/svg+xml;base64,PHN2ZyB3aWR0aD0iMSIgaGVpZ2h0PSIxIiB4bWxucz0iaHR0cDovL3d3dy53My5vcmcvMjAwMC9zdmciPjwvc3ZnPg==\" \/><\/p>\n<p class=\"m_-2244817513273254883imgCaption\">A CCTV security surveillance camera overlooks a street as people walk following the spread of the coronavirus disease (COVID-19) in Beijing, China May 11, 2020. REUTERS\/Thomas Peter<\/p>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<table>\n<tbody>\n<tr>\n<td>\n<h3>Beyond facial recognition<\/h3>\n<p>&nbsp;<\/p>\n<p>While live facial recognition dominates headlines, there are other types of technology with strikingly similar capabilities that often fly under the radar.<\/p>\n<p>Last year, the UK\u2019s Metropolitan Police spent \u00a33 million on retrospective facial recognition technology (RFR). RFR uses facial recognition to scan images already collected by CCTV, rather than scanning people in real-time. In an\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-f\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-f\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw2Xdcj-vBW7iy5Yr6sYdizd\">article I wrote last year<\/a>, experts warned that the technology can be used in much the same way as LFR and has many of the same risks. Despite this, there are almost no limitations on its use in the UK and abroad.<\/p>\n<p>For some police forces, simply detecting faces in a crowd doesn\u2019t go far enough. There are now several products that not only identify people, but also\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-z\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-z\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw21ivFtE2GAA2vA1rZXQxVd\">analyse peoples\u2019 emotions<\/a>. The technology, which is already being used in the\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-v\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-v\/&amp;source=gmail&amp;ust=1668604132401000&amp;usg=AOvVaw0udTc06rm6EqEWMykKmD8S\">heavily monitored region of Xinjiang, China<\/a>, can then supposedly help police predict crime.<\/p>\n<p>Like a lot of new surveillance technology, however, it\u2019s unlikely to live up to the hype and the technology has repeatedly been shown to be inaccurate. Worse yet, it\u2019s been accused of being based on \u201c<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-e\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-e\/&amp;source=gmail&amp;ust=1668604132402000&amp;usg=AOvVaw1iev1jIFxLZkkrnAY2Ufyz\">racist pseudoscience<\/a>\u201d that could lead to higher rates of discriminatory policing.<\/p>\n<p>The UK\u2019s Information Commissioner&#8217;s Office (ICO)\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-s\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-s\/&amp;source=gmail&amp;ust=1668604132402000&amp;usg=AOvVaw2d8MCCXHQongq0yiYU6IsX\">has warned against the use of such technology<\/a>\u00a0in the UK. But as attempts to regulate live facial recognition show, meaningful restrictions on potentially dangerous surveillance technology can take years to establish.<\/p>\n<p>With new guidance from the ICO on biometric technologies expected next spring, it\u2019s crucial that it looks ahead and offers proactive, comprehensive and meaningful guidance on facial recognition and other forms of biometric surveillance that may soon become staple parts of contemporary policing.<\/p>\n<p><strong>Recommended Reading<\/strong><\/p>\n<p><strong>Khari Johnson,\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-g\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-g\/&amp;source=gmail&amp;ust=1668604132402000&amp;usg=AOvVaw06e7GO4DSi4d9OxxJaOJjW\">How Wrongful Arrests Based on AI Derailed 3 Men\u2019s Lives<\/a>, Wired, March 7, 2022.<\/strong><\/p>\n<p>Discussions of facial recognition can often overlook the human impact. This article shows how wrongful arrests caused by facial recognition software damage peoples\u2019 lives and highlights the importance of regulating its use.<\/p>\n<p><strong>Evani Radiya-Dixit,\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-w\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-w\/&amp;source=gmail&amp;ust=1668604132402000&amp;usg=AOvVaw0yBGHCmhjHyIZovVY50RW8\">A Sociotechnical Audit: Assessing Police Use of Facial Recognition<\/a>, Minderoo Centre for Technology and Democracy, University of Cambridge<\/strong><\/p>\n<p>The full report from the University of Cambridge offers detailed information regarding the ethical and legal problems surrounding the police\u2019s use of the technology. It also includes an audit that could be adopted for analysing future trials of the technology in the UK.<\/p>\n<p><strong>Lauren Rhue,\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-yd\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-yd\/&amp;source=gmail&amp;ust=1668604132402000&amp;usg=AOvVaw1_5aY8L65ppHb4ejcT8LUD\">Emotion-Reading Tech Fails the Racial Bias Test<\/a>, January 3 2019.<\/strong><\/p>\n<p>This article from back in 2019 shows that racial bias within facial recognition software is far from new. The study found that emotion-reading technology considers black faces to be angrier than white faces, something that could have awful consequences in a policing\u00a0<span class=\"il\">context<\/span>.<\/p>\n<p><strong>Nicol Turner Lee and Caitlin Chin,\u00a0<a href=\"https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-yh\/\" target=\"_blank\" rel=\"noopener\" data-saferedirecturl=\"https:\/\/www.google.com\/url?q=https:\/\/thomsonreutersfoundation.cmail20.com\/t\/d-l-zstjz-jhkldtildl-yh\/&amp;source=gmail&amp;ust=1668604132402000&amp;usg=AOvVaw30oBsQ_hCdFUBNxl5bwddn\">Police Surveillance and Facial Recognition: Why Data Privacy Is Imperative for Communities of Color<\/a>, Brookings Institution, April 12, 2022.<\/strong><\/p>\n<p>This report for the Brookings Institution makes the case for stronger privacy protections in the United States to help limit the risks of facial recognition technology, particularly as they relate to communities of colour.<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n","protected":false},"excerpt":{"rendered":"<p>By Samuel Woodhams | Digital rights researcher and journalist Facial recognition technology made headlines again last month as researchers at the University of Cambridge,<a class=\"moretag\" href=\"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/\">Read More&#8230;<\/a><\/p>\n","protected":false},"author":1,"featured_media":1280,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"googlesitekit_rrm_CAowu7GVCw:productID":"","_crdt_document":"","activitypub_content_warning":"","activitypub_content_visibility":"","activitypub_max_image_attachments":3,"activitypub_interaction_policy_quote":"anyone","activitypub_status":"","footnotes":""},"categories":[43],"tags":[3233,2609,2502,3240,3235,3244,771,3012,3237,1404,3243,3236,3234,3239,1405,1043,773,3246,3238,3245,1406,472,3242,1401,2664,2588,2471,723,744,3010,3241],"class_list":["post-1279","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-security","tag-automatic-identification-and-data-capture","tag-beijing","tag-belgium","tag-bill","tag-biometrics","tag-caitlin-chin","tag-china","tag-closed-circuit-television","tag-colombia","tag-crime-prevention","tag-evani-radiya-dixit","tag-face-id","tag-facial-recognition-system","tag-lauren-rhue","tag-law-enforcement","tag-learning","tag-london","tag-luxembourg","tag-mark-travis","tag-morocco","tag-national-security","tag-netflix","tag-nicol-turner-lee","tag-prevention","tag-samuel-woodhams-digital","tag-shanghai","tag-surveillance","tag-united-kingdom","tag-united-states","tag-video-surveillance","tag-xinjiang"],"featured_image_src":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png","blog_images":{"medium":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1-300x200.png","large":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png"},"ams_acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Facial recognition: An ethical policing tool? - Mpelembe Network<\/title>\n<meta name=\"description\" content=\"Facing the facts: Facial recognition\u2019s ethical and legal limitations\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Facial recognition: An ethical policing tool? - Mpelembe Network\" \/>\n<meta property=\"og:description\" content=\"Facing the facts: Facial recognition\u2019s ethical and legal limitations\" \/>\n<meta property=\"og:url\" content=\"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/\" \/>\n<meta property=\"og:site_name\" content=\"Mpelembe Network\" \/>\n<meta property=\"article:published_time\" content=\"2022-11-15T13:13:38+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2022-11-15T15:07:20+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png\" \/>\n\t<meta property=\"og:image:width\" content=\"640\" \/>\n\t<meta property=\"og:image:height\" content=\"426\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"admin\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"5 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/\"},\"author\":{\"name\":\"admin\",\"@id\":\"https:\\\/\\\/mpelembe.net\\\/#\\\/schema\\\/person\\\/2421ebbf3150931b1066b10a196d7608\"},\"headline\":\"Facial recognition: An ethical policing tool?\",\"datePublished\":\"2022-11-15T13:13:38+00:00\",\"dateModified\":\"2022-11-15T15:07:20+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/\"},\"wordCount\":1090,\"image\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/mpelembe.net\\\/wp-content\\\/uploads\\\/2022\\\/11\\\/unnamed-1.png\",\"keywords\":[\"Automatic identification and data capture\",\"Beijing\",\"Belgium\",\"Bill\",\"Biometrics\",\"Caitlin Chin\",\"China\",\"Closed-circuit television\",\"Colombia\",\"Crime prevention\",\"Evani Radiya-Dixit\",\"Face ID\",\"Facial recognition system\",\"Lauren Rhue\",\"Law enforcement\",\"Learning\",\"London\",\"Luxembourg\",\"Mark Travis\",\"Morocco\",\"National security\",\"Netflix\",\"Nicol Turner Lee\",\"Prevention\",\"Samuel Woodhams Digital\",\"Shanghai\",\"Surveillance\",\"United Kingdom\",\"United States\",\"Video surveillance\",\"Xinjiang\"],\"articleSection\":[\"Security\"],\"inLanguage\":\"en-US\"},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/\",\"url\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/\",\"name\":\"Facial recognition: An ethical policing tool? - Mpelembe Network\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/mpelembe.net\\\/wp-content\\\/uploads\\\/2022\\\/11\\\/unnamed-1.png\",\"datePublished\":\"2022-11-15T13:13:38+00:00\",\"dateModified\":\"2022-11-15T15:07:20+00:00\",\"author\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/#\\\/schema\\\/person\\\/2421ebbf3150931b1066b10a196d7608\"},\"description\":\"Facing the facts: Facial recognition\u2019s ethical and legal limitations\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/#primaryimage\",\"url\":\"https:\\\/\\\/mpelembe.net\\\/wp-content\\\/uploads\\\/2022\\\/11\\\/unnamed-1.png\",\"contentUrl\":\"https:\\\/\\\/mpelembe.net\\\/wp-content\\\/uploads\\\/2022\\\/11\\\/unnamed-1.png\",\"width\":640,\"height\":426},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/facial-recognition-an-ethical-policing-tool\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/mpelembe.net\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Facial recognition: An ethical policing tool?\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/mpelembe.net\\\/#website\",\"url\":\"https:\\\/\\\/mpelembe.net\\\/\",\"name\":\"Mpelembe Network\",\"description\":\"Collaboration Platform\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/mpelembe.net\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/mpelembe.net\\\/#\\\/schema\\\/person\\\/2421ebbf3150931b1066b10a196d7608\",\"name\":\"admin\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/c66a2765397adfb52418f6f2310640167a0af23ce662da1b68c8a0b8650de556?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/c66a2765397adfb52418f6f2310640167a0af23ce662da1b68c8a0b8650de556?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/c66a2765397adfb52418f6f2310640167a0af23ce662da1b68c8a0b8650de556?s=96&d=mm&r=g\",\"caption\":\"admin\"},\"sameAs\":[\"https:\\\/\\\/mpelembe.net\"],\"url\":\"https:\\\/\\\/mpelembe.net\\\/index.php\\\/author\\\/admin\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Facial recognition: An ethical policing tool? - Mpelembe Network","description":"Facing the facts: Facial recognition\u2019s ethical and legal limitations","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/","og_locale":"en_US","og_type":"article","og_title":"Facial recognition: An ethical policing tool? - Mpelembe Network","og_description":"Facing the facts: Facial recognition\u2019s ethical and legal limitations","og_url":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/","og_site_name":"Mpelembe Network","article_published_time":"2022-11-15T13:13:38+00:00","article_modified_time":"2022-11-15T15:07:20+00:00","og_image":[{"width":640,"height":426,"url":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png","type":"image\/png"}],"author":"admin","twitter_card":"summary_large_image","twitter_misc":{"Written by":"admin","Est. reading time":"5 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/#article","isPartOf":{"@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/"},"author":{"name":"admin","@id":"https:\/\/mpelembe.net\/#\/schema\/person\/2421ebbf3150931b1066b10a196d7608"},"headline":"Facial recognition: An ethical policing tool?","datePublished":"2022-11-15T13:13:38+00:00","dateModified":"2022-11-15T15:07:20+00:00","mainEntityOfPage":{"@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/"},"wordCount":1090,"image":{"@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/#primaryimage"},"thumbnailUrl":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png","keywords":["Automatic identification and data capture","Beijing","Belgium","Bill","Biometrics","Caitlin Chin","China","Closed-circuit television","Colombia","Crime prevention","Evani Radiya-Dixit","Face ID","Facial recognition system","Lauren Rhue","Law enforcement","Learning","London","Luxembourg","Mark Travis","Morocco","National security","Netflix","Nicol Turner Lee","Prevention","Samuel Woodhams Digital","Shanghai","Surveillance","United Kingdom","United States","Video surveillance","Xinjiang"],"articleSection":["Security"],"inLanguage":"en-US"},{"@type":"WebPage","@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/","url":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/","name":"Facial recognition: An ethical policing tool? - Mpelembe Network","isPartOf":{"@id":"https:\/\/mpelembe.net\/#website"},"primaryImageOfPage":{"@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/#primaryimage"},"image":{"@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/#primaryimage"},"thumbnailUrl":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png","datePublished":"2022-11-15T13:13:38+00:00","dateModified":"2022-11-15T15:07:20+00:00","author":{"@id":"https:\/\/mpelembe.net\/#\/schema\/person\/2421ebbf3150931b1066b10a196d7608"},"description":"Facing the facts: Facial recognition\u2019s ethical and legal limitations","breadcrumb":{"@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/#primaryimage","url":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png","contentUrl":"https:\/\/mpelembe.net\/wp-content\/uploads\/2022\/11\/unnamed-1.png","width":640,"height":426},{"@type":"BreadcrumbList","@id":"https:\/\/mpelembe.net\/index.php\/facial-recognition-an-ethical-policing-tool\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/mpelembe.net\/"},{"@type":"ListItem","position":2,"name":"Facial recognition: An ethical policing tool?"}]},{"@type":"WebSite","@id":"https:\/\/mpelembe.net\/#website","url":"https:\/\/mpelembe.net\/","name":"Mpelembe Network","description":"Collaboration Platform","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/mpelembe.net\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/mpelembe.net\/#\/schema\/person\/2421ebbf3150931b1066b10a196d7608","name":"admin","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/c66a2765397adfb52418f6f2310640167a0af23ce662da1b68c8a0b8650de556?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/c66a2765397adfb52418f6f2310640167a0af23ce662da1b68c8a0b8650de556?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/c66a2765397adfb52418f6f2310640167a0af23ce662da1b68c8a0b8650de556?s=96&d=mm&r=g","caption":"admin"},"sameAs":["https:\/\/mpelembe.net"],"url":"https:\/\/mpelembe.net\/index.php\/author\/admin\/"}]}},"_links":{"self":[{"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/posts\/1279","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/comments?post=1279"}],"version-history":[{"count":2,"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/posts\/1279\/revisions"}],"predecessor-version":[{"id":1282,"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/posts\/1279\/revisions\/1282"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/media\/1280"}],"wp:attachment":[{"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/media?parent=1279"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/categories?post=1279"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/mpelembe.net\/index.php\/wp-json\/wp\/v2\/tags?post=1279"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}