Can Facial Recognition Overcome Its Racial Bias?

The dystopian surveillance state of science fiction media is within reachand some privacy activists argue that its already here. Facial recognition advancements have spiked fear and uncertainty over misuse and civil liberties infringements, but with the alarm comes a wave of activists bringing solutions.
What is facial recognition?
Facial recognition is a form of artificial intelligence. Artificial intelligence broadly refers to the development of computers to perform tasks that would normally require human intelligence. If you have an email, you are indebted to AI for directing spam to a separate folder instead of flooding your inboxthe computer learned to recognize the pattern of spam and filter it accordingly. If you have a YouTube account or music streaming service, your personalized recommendations are the product of an AI algorithm.
For facial recognition, algorithms are written to measure the geometry of someones face, compare those unique measurements to a database of faces, and return potential matches with varying degrees of certainty.
Facial recognition can offer convenience, such as the millions of people who use their face to unlock their iPhones. It can also be used as a surveillance tool, like the , a largely Muslim minority the government has been sending to detention camps.
As AI embeds itself into daily technological life, privacy activists and technology enthusiasts agree that the powerful tool is here to stay. But, the implementation of the technology has complex problems, as well as disagreements about which solutions are actually solutions.
Resolving the bias in the technology
The current facial recognition technology available has racial and gender bias. A 2018 study by MIT found that while determining gender using three different facial recognition programs, the .
In 2018, Amazons facial recognition tool, Amazon Rekognition, . While comparing the lawmakers faces against mugshot databases, Rekognition misidentified lawmakers of color at a rate of 39% even though they only made up 20% of the people.
Brian Brackeen, founder of facial recognition technology company Kairos, says that its possible to train the bias out of the technology.
How?
Facial recognition algorithms improve by being supplied large datasets of faces. Existing facial recognition products work well on pale males because the algorithms were supplied datasets of majority White men, reflective of the tech industry itself. An MIT and Stanford University study found that a . Brackeen says this bias toward the people who make the technology is common; similarly, algorithms in Asia tend to perform well on Asian males and not as well on White males.
To solve the bias, the algorithms need to be supplied a more diverse set of faces to learn from, but high-quality datasets of people of color arent always available. For example, cellphone photos of dark-skinned people are unreliable because the light sensors are inconsistent and rarely accurately reflect the persons skin tone, Brackeen says. Additionally, many companies struggle to find enough high-resolution photos that can be consensually added to training datasets because it was full of faces scraped from the web and videos under the Creative Commons license.
Thats why Brackeen supports the use of Generative Adversarial Networks to train facial recognition algorithms. GANs are able to create computer-generated faces made from a composite of photos of real people. The and can be generated to meet specifications such as age, race, gender, and photo quality.
We can create a million Black men, a million Asian men, a million whatever, and just load the AI with those generated people, Brackeen says. Using that method, he believes there will be no bias in facial recognition AI three years from now.
But even if facial recognition AI becomes a completely accurate tool, Brackeen believes the technology should never be used by law enforcement. He sees facial recognition as a way to boost convenience. For example, being used as a way to checkout at the grocery store fasterthink Amazon Go storesor being implemented at Disneyland to speed up the wait times on buying passes or getting on rides. In those instances, the consumer would opt-in to using facial recognition and the consequences of the technology failing would be minor annoyances, such as a longer checkout process or slower lines to get on rides.
Government use of the technology
In the hands of law enforcement, however, the tool could prove dangerous. If someone who is pulled over for speeding is misidentified as being on a dangerous persons database via an officers body cam feed, that situation could turn fatal.
These impacts are too great, Brackeen says.
Its unclear how many law enforcement departments have access to facial recognition technology and how those departments that do have access use the tool. The Georgetown Law Center on Privacy and Technology was the first group to present a . The team submitted public records requests to more than 100 law enforcement departments, finding that there was no regulation on the facial recognition technology that .
We ourselves were surprised by some of the findings we had, mostly in terms of the sheer scope of use of facial recognition at the state and local level across the country, says Clare Garvie, a senior associate with the Georgetown Law Center on Privacy and Technology. Also, [we were surprised] at the complete absence of any legal imposed regulations or even policies implemented by law enforcement to constrain its use.
Garvies work revealed that at least 52 police departments of the 100 queried have access to, use, or have previously used facial recognition technology. Several large police departments, including those in Los Angeles, Chicago, and Dallas, either had access to or were exploring the use of real-time facial recognition that can continuously scan pedestrians faces and compare them against databases such as mugshots or drivers licenses.
Additionally, Georgetowns research found that because of the lack of policy and limitations on how the technology is used, and earnestly using the output.
For example, in 2017, a pixelated image from a surveillance camera of a suspect stealing beer from a CVS returned no matches when run through the NYPDs facial recognition technology. Someone mentioned that the suspect resembled actor Woody Harrelson, so a picture of Harrelson was ran through the program and the detectives picked someone who they believed looked like the suspect from the resulting matches. That celebrity match was sent back to the investigating officers, and an arrest was made.
The stakes are too high in criminal investigations for the police to be using probe photos and doppelg瓣ngers, Garvie says.
Additionally, because most facial recognition programs used by law enforcement search mugshot databases, people of color, particularly young Black men, are overrepresented in the possible matches. Even if the technologys racial bias problem is solved, the use of facial recognition within the existing criminal justice system could just replicate the over-policing of Black and Brown communities.
Its not appropriate to ask individuals to affirmatively fight for our right to privacy against government intrusion, Garvie says. That said, I think the single most important role individuals can play is demanding transparency and accountability from their local and state officials, and pushing their legislators for the legislation and regulation that they deem appropriate in this field.
Garvie advocates for a moratorium, or a temporary halt, on governments use of facial recognition technology while researchers and lawmakers have time to catch up and consider the impacts of this fast-moving technology. San Francisco went a step further last May, approving a .
Immediately after the ordinance was put in place, city employees scrambled to remove the face unlock feature from their government-issued iPhones. The language in the ordinance was then amended to allow non-surveillance facial recognitionlike face unlock on the iPhoneto be used, but Garvie highlights it as an example of why its so difficult to develop legislation that anticipates for the future without being too broad and cutting off AI uses that are more helpful than hurtful.
Its very challenging to think about legislation that regulates exactly what you want it to, that doesnt leave loopholes but isnt also overinclusive, Garvie says. Its a challenge, and its never going to be perfect.
How the ban will affect the San Francisco Police Departments criminal investigations, though, is uncertain. While the SFPD did not respond to YES!’s request for comment, the Georgetown Law report found the and could search a half-million to a million mug shots.
The San Francisco Sheriffs Department has more autonomy under California state law. Nancy Crowley, the director of communications for the sheriff’s department, says the ordinance has not impacted the department’s operations and that the department may still use facial recognition within the county jails.
I dont know to what extent we use it, Crowley says.
Activists such as Evan Greer, director of operations at Fight for the Future, believe the San Francisco ban is a step in the right direction, but say other government efforts to curtail facial recognition use are insufficient.
We dont think there are meaningful limitations you can put on this technology, says Greer.
Greer believes that while developing the language around legislative regulation may be tricky, an outright ban on facial recognition surveillance by government, law enforcement, and private corporations sidesteps the concern for future-proofing and overinclusion.
Fight for the Future, a nonprofit grassroots advocacy group, has built a , making headlines when they got at the festivals. The group has since expanded to organizing on college campuses and supporting movements by acting as an .
Greer and Fight for the Futures focus is on slowing down the rapid adoption of facial recognition and giving lawmakers time to do their jobsput a policy in place that regulates its use. In the meantime, Greers opposition to facial recognition surveillance is not quelled by the removal of bias and increasing accuracy of the technology.
If we have systems in place that make it possible to enforce laws 100% of the time, then there is no space for us to test whether those laws are just, Greer says, highlighting civil rights, LGBTQ rights, and the legalization of marijuana as examples. For me, at a philosophical level, privacy isnt about what you have to hide, its about our ability to evolve as a human society.
Increasing the accessibility of advocacy
Because facial recognition is developed by small groups of people within the tech field, the technology operates within a black box. Yet, its impact is largely on the public, who cant see inside the black box and who dont have the information to fully understand its impact.
The end of the for legislatures, law enforcement, facial recognition companies, and community leaders. While suggestions for law enforcement and lawmakers are pretty standardsuch as officers must have probable cause to run a photo through facial recognitionthe section for community leaders is unique because it provides individuals with the questions they need to ask to get information from their police department. Questions such as: Who is enrolled in the police face recognition database? What legal requirements must be met before officers run a face recognition search? How does the agencys face recognition use policy protect free speech?
These questions reveal the need to involve people at every level. The Georgetown report reflects only 100 of the 18,000 police departments across the country. Garvie believes getting more people prepared to advocate for themselves is critical to uncovering what we still dont know about law enforcements use of facial recognition.
Preparing people to advocate for themselves is also why Brooklyn-based artist, researcher, and technologist Mimi Onuoha teamed up with Mother Cyborg, aka Detroit-based artist, DJ, and educator Diana Nucera, to make an .
The two met at a conference about the future of AI, engaging in conversation surrounding the ethics and potential pitfalls of AI. They bonded over how they were going to share this information with their communities. In Detroit, where Nucera lives, 40% of people dont have access to the internet, so how is she to engage her community in complex AI topics when they are still working on getting them online?
It was clear to me that we were in this room of pretty privileged folks in the field, talking about these things that really my neighbors should be talking about, Nucera says.
The two collaborated to distill the most essential concepts of AI down to a simple format that could be used for self-teaching, or as a tool for a teacher to use in the classroom. The zine not only gives communities the tools to understand and talk about AI, but also to foster conversation around the growing tech that isnt dominated by a fatalist point of view.
There must be a way to talk about this that doesnt end in this strange robots are going to kill us way, Onuoha says. We know that there are alternate conversations we can have.
The 84-page zine engages in that alternate conversation by including exercises and prompts that demystify algorithm building by having the reader identify algorithms in their own lives. The format is digestible, building upon a basis of knowledge to get to the more difficult ideas, like whether algorithms can be used to quantify and judge people in an ethical way.
The booklet sold out of physical copies twice, and Onuoha and Nucera continue to run workshops for non-tech and tech people alike, leading critical discussion on what regulation and data collection should look like and who is being served or harmed by the use of AI and facial recognition.
Ultimately, the zine provides enough information to act as a launchpad to engage in those critical conversations. Both Onuoha and Nucera want to make sure that their neighbors, mail carriers, and anyone else they may see in their community have the language and information they need to participate in the discussion about AI surveillance being implemented in their community.
They are the only ones who can say how its impacting them and whether or not they want it, Nucera says. The rest of it is just assumption. So how do you get people to join those conversations? Theyve got to learn the talk.
![]() |
Isabella Garcia
is a former solutions reporter and former editorial intern for YES! Media. Her work has appeared in The Malheur Enterprise and YES! Magazine. Isabella is based in Portland. She can be reached at isabellagarcia.website.
|