Facial Recognition AI Software Discriminating Against People With Dark Skin
Racist AI software discriminates against people with dark skin according to a new study.
Strange flaws in an AI l recognition software has prompted it to begin discriminating against people with dark skin according to a new study.
According to The Gender Shades project, conducted by the MIT Media Lab, tested facial recognition software from Microsoft, BM, and Face++ on over 1,2000 images if people from 6 European and African countries.
Despite the software being able to guess the gender if a person 99% of the time, it made over 35% more mistakes when studying people with darker skin.
Images of men saw around 8% more errors, while images of woman saw an error rate of over 20%.
RT.com reports: The products made fewer errors when identifying subjects with lighter skin, with error rates ranging between 11 and 20 percent. “We found that all classifiers performed best for lighter individuals and males overall. The classifiers performed worst for darker females,” the report concluded.
The study claims that, while facial recognition software is increasingly used by law enforcement in the US, none of these products have been publicly tested for accuracy. “This kind of technology is being built on machine-learning techniques. And machine-learning techniques are based on data. So if you have biased data in the input and it’s not addressed, you’re going to have biased outcomes,” Joy Buolamwini, the study’s lead author, told NPR.
The MIT report said that these failures require urgent attention if companies are to build fair and transparent facial analysis systems, especially as machine-learning software is now being used to decide loan, job and university applications.
According to a report from Georgetown University’s Center on Privacy and Technology, some 16 states have let the FBI compare driver’s licenses to the faces of suspected criminals. More than 117 million Americans are thought to be on the facial recognition database.
The software is being rolled out by police forces around the world. In August, it was revealed that police in the UK have catalogued more than 20 million facial images – the equivalent of one third of Britain’s population. Earlier this month, Chinese police began tests of its facial recognition glasses to identify ‘fugitives.’ The glasses, which were being used by police in a train station, are connected to a tablet-like device that allows officers to take mugshots and compare them to a police database.