Facial recognition technology is failing to recognise transgender people, new research has shown, raising concerns about discrimination as the use of the software becomes increasingly prevalent.
Researchers at the US University of Colorado, Boulder, tested facial recognition systems from tech giants IBM, Amazon, Microsoft and Clarifai, on photographs of trans men, and found they were misidentified as women 38% of the time.
Cisgender women and men – or those who identify as their birth gender – were correctly identified 98.3% of the time and 97.6% of the time respectively.
The software also failed to recognise people who did not define themselves as male or female – also known as non-binary, agender, or genderqueer – 100% of the time.
The results highlight that even the most up-to-date technology only views gender in two set categories, the report’s lead author, Morgan Klaus Scheuerman, said in a statement.
“While there are many different types of people out there, these systems have an extremely limited view of what gender looks like,” Scheuerman said.
Facial recognition remains highly controversial, but is gaining in use by police and immigration services. The market for the technology is predicted to double in the next 15 years, according to research group, MarketsandMarkets.
Software that excludes trans and non-binary people may prove discriminatory, rendering such persons invisible to a technology that is becoming increasingly incorporated into daily life.
Misidentification can even be actively harmful, such as at airport security, where trans people are often subject to invasive body searches or harassment, if their ID does not match their gender.
The Transportation Security Administration (TSA) is currently rolling out facial recognition at airports across the United States.
A spokesman from LGBT+ group, Stonewall, said:
“It’s concerning to hear that facial recognition software is misgendering trans people. The experience of being deliberately misgendered is deeply hurtful for trans people.
“We would encourage technology developers to bring in and consult with trans communities to make sure their identity is being respected.”
The study also suggested the software relies on outdated gender stereotypes in its facial analysis. Scheuerman, who is male with long hair, was categorised as female half of the time.
“When you walk down the street you might look at someone and presume that you know what their gender is, but that is a really quaint idea from the ’90s, and it is not what the world is like anymore,” said senior author, Jed Brubaker, an assistant professor of Information Science.
“As our vision and our cultural understanding of what gender is has evolved, the algorithms driving our technological future have not. That’s deeply problematic.”
-Molly Millar – Thomson Reuters Foundation