Advertisement
Opinion

Face recognition tech is everywhere but it has prejudices it won’t admit to

As the Met Police face calls to drop face recognition technology, Rage Inside the Machine author Robert Elliot Smith explores the biases of AI systems and says maybe it’s time to get back to the human touch

Five years ago, my father-in-law leaned his finger too hard on the button of my iPhone, and a burst of 10 holiday snaps of my wife and me was uploaded to Google’s photo cloud. Days later, I received a photo of a moment that had never happened. Without the intervention of any person, an algorithm had taken my wife’s smile from one picture in that burst, and convincingly pasted it on to her face in another photo, so that it seemed as if we had smiled in the same moment. At the time, this was a minor curiosity, and my blog piece about it generated a few interviews and articles.

Fast forward to today, when face recognition is being rolled out into national CCTV systems, Chinese police are using AR goggles that can tag potential offenders in their line-of-sight, and deep fake videos are poised to influence the next American election. Bursts of images of all of us are now continuously scanned by algorithms, the people in those images are electronically identified, and automated technology can easily manipulate images and video to create moments that never happened. My provocative blog entitled The AIs are now rewriting history could be coming true, and it may be time to reconsider what history tells us about image analysis and its effects on society.

It may be time to reconsider what history tells us about image analysis and its effects on society.

The first scientific system for identifying criminals from images was invented by Alphonse Bertillon, who (in 1882) not only developed the mug shot, but a system of physical measurements to uniquely identify criminals. This system was used internationally for many years despite its flaws. For example, in 1903, a black man named Will West was incarcerated in Leavenworth in Kansas only to find another black man with the same name and the same measurements imprisoned there as well. Errors of this sort were inevitable since the Bertillon measures reflected the racial biases of the era; simplifying measures, based on how people look, cannot avoid generalisations about the people portrayed.

Due to mistakes like the Will West case, fingerprints (which were developed in 1892 by Francis Galton, Darwin’s first cousin, the inventor of eugenics, and the inspiration for Bertillon’s work) rapidly replaced the Bertillon system. However, generalisations about offenders based on images persisted. Around the same time as Bertillon was developing his system, Italian criminologist Cesare Lombroso created his method of “criminal anthropology” which directly generalised about the relationship between face measurements and an individual’s criminal potential. While the inherent prejudice in this system has been historically deprecated, criminal anthropology lives on in the form of “offender profiling”.

We now live in a world inundated with image data, provided by dense CCTV, the rise of police body cams, and ubiquitous smartphone cameras. Careful processes of image evaluation by people are increasingly impractical, and therefore, scrutiny of images by AI algorithms is rapidly becoming the only way to cope with this glut of visual data.

Many would assume that algorithms solve the problem of unreliable and biased witnesses because algorithms are thought to be inherently objective. However, the hidden reality is that algorithms rely on simplified metrics, not unlike those in the systems of Bertillon and Lombroso. Only now a sea of mathematical and computational techniques hide those simplifications. Because of their vast complexity, no one can query an algorithm as they could an eyewitness. Thus algorithmic biases can easily be overlooked. Due to flaws in design and biases in the data used to create them, algorithms can be inherently racist, as is being demonstrated in many strands of recent research. These biases are not easily overcome.

Advertisement
Advertisement

Moreover, the reliability of the image streams that algorithms provide is due to become increasingly suspect. “AI-enhanced” imaging is becoming an aspect of cameras, and “deep fake” video techniques will soon be a part of the repertoire of autonomous algorithms. Human eyewitnesses’ vulnerability to incentives and persuasion has long been a problem in their reliability, and likewise, what an algorithm recalls as a real image is liable to be influenced by the incentives of the algorithm’s operators. A stretched criminal justice system, increasingly evaluated on efficiency metrics, may unconsciously turn its algorithms towards pictures that are more likely to get a conviction. Likewise, the algorithms involved in individual social media feeds are likely to deliver videos that reinforce a user’s currently held beliefs.

These effects could lead to a worldview that either dangerously enforces a common, centrally dictated perspective, or shatters reality into opposing viewpoints that polarise and erode trust between people. How can we agree on a shared view of the world when the images we see of it may be entirely manipulated, possibly by forces we cannot see, understand, or control?

Ironically, human trust may be the only solution.

Ironically, human trust may be the only solution. Several organisations are now creating technology that can aid users in determining the sources of news content, and these technologies are sure to eventually include ways of determining the sources of images and video. Any source could, of course, manipulate the reality seen in online content, but if people could rely on the regulated standards of institutions (for instance, journalistic standards), it may be possible to create a stable, shared view of the world. Ironically, the dissolution of confidence in the mechanically recorded image may force people to redevelop their faith in one another’s human worldview.

Robert Elliot Smith is an AI engineer and the author of Rage Inside the Machine: The Prejudice of Algorithms, and How to Stop the Internet Making Bigots of Us All (Bloomsbury, £20)

Advertisement

Support the Big Issue

For over 30 years, the Big Issue has been committed to ending poverty in the UK. In 2024, our work is needed more than ever. Find out how you can support the Big Issue today.
Vendor martin Hawes

Recommended for you

Read All
Social care is on its knees. It's no wonder public dissatisfaction is at a record high
social care
Evan John

Social care is on its knees. It's no wonder public dissatisfaction is at a record high

Investment in social housing is an investment in people
John Bird

Investment in social housing is an investment in people

Two-child limit on benefits is cruel and unfair. Politicians must rethink ahead of general election
two child limit/ three kids
Martin O'Neill

Two-child limit on benefits is cruel and unfair. Politicians must rethink ahead of general election

Want to rebuild the UK economy, Jeremy Hunt? Start with our beautiful public libraries
library
Jo Cornish

Want to rebuild the UK economy, Jeremy Hunt? Start with our beautiful public libraries

Most Popular

Read All
Renters pay their landlords' buy-to-let mortgages, so they should get a share of the profits
Renters: A mortgage lender's window advertising buy-to-let products
1.

Renters pay their landlords' buy-to-let mortgages, so they should get a share of the profits

Exclusive: Disabled people are 'set up to fail' by the DWP in target-driven disability benefits system, whistleblowers reveal
Pound coins on a piece of paper with disability living allowancve
2.

Exclusive: Disabled people are 'set up to fail' by the DWP in target-driven disability benefits system, whistleblowers reveal

Here's when UK households to start receiving last cost of living payments
next dwp cost of living payment 2023
3.

Here's when UK households to start receiving last cost of living payments

Citroën Ami: the tiny electric vehicle driving change with The Big Issue
4.

Citroën Ami: the tiny electric vehicle driving change with The Big Issue