Greetings, and welcome to this week’s newsletter.
For me, this is one of the most interesting letters in a while. The WIRED story on the use of facial recognition in Ukraine to identify dead Russian soldiers is a case on point, particularly with its focus on ‘positive’ facial recognition deployments as a means of obscuring other deployments. Some of the other issues not addressed in the article include issues to do with respect for the dead under IHL, and broader considerations arising from previously reported practices of Ukrainian soldiers contacting the deceased’s relatives, or posting to social media, to let them know that they are dead (which must engage the right to private and family life, while also bringing the concept of human dignity into play).
MIT Tech Review also looks at how AI can actually be quite helpful in disaster response. It feels like the benefits of AI are often unaddressed. I wonder how much of this is due to overstating claims as to AI’s effectiveness (leading to appropriate skepticism) but also maybe the lack of an evidence base for specific AI tools (something I’m thinking about a lot in work at the moment).
Turning to socio-economic rights, the Washington Post has a story on the use of algorithms to decide who gets laid off, something that was touched on in a previous newsletter, but this story has quite a lot of useful/interesting detail.
We also have two book announcements, in a brand new section below. If you have any suggestions for upcoming events, books, etc. please do drop us a line.
Thanks as ever to Sarah Zarmsky, and hope you all have a great week.
The dark side of AI: Queen Mary digital culture expert reveals how algorithms are generating hate and discrimination,Queen Mary University of London News
Robot helps students with learning disabilities stay focused, University of Waterloo News
AI is starting to pick who gets laid off, The Washington Post
Who Is Liable When AI Kills?, Scientific American
So How Fair Is Your AI, Exactly?, Towards Data Science
Improving trust in autonomous technology, MIT Technology Review
How OpenAI is trying to make ChatGPT safer and less biased, MIT Technology Review
How AI can actually be helpful in disaster response, MIT Technology Review
How Citizen is trying to remake itself by recruiting elderly Asians, MIT Technology Review
The Right’s New Culture-War Target: “Woke” AI, The Washington Post
How to reduce the risks from AI’s original sin, The Washington Post
The big idea: should robots take over fighting crime?, The Guardian
Why facial recognition could spell the end of anonymity in public, The Sunday Times
*Disclaimer: The selected articles and chapters were not evaluated for their research methods and do not necessarily reflect the views of the AI & Human Rights Blog
The ethical dimensions of Google autocomplete, Rosie Graham, Big Data & Society
The Matrix of Privacy: Data Infrastructure in the AI-Powered Metaverse, Leon Yehuda Anidjar, Nizan Geslevich Packin, and Argyri Panezi, available on SSRN
The right to information or data sovereignty? Sending unsolicited messages to Russians about the war in Ukraine, Yao-Tai Li and Katherine Whitworth, Big Data & Society
Social media advertising for clinical studies: Ethical and data protection implications of online targeting, Rainer Mühlhoff and Theresa Willem, Big Data & Society
Google Search and the creation of ignorance: The case of the climate crisis, Jutta Haider and Malte Rödl, Big Data & Society
Justice, trust, and moral judgements when personnel selection is supported by algorithms, Tina Feldkamp, Markus Langer, Leo Wies & Cornelius J. König, European Journal of Work and Organizational Psychology
Economies of Virtue: The Circulation of ‘Ethics’ in AI, edited by Thao Phan, Jake Goldenfein, Declan Kuch, and Monique Mann