Fixing AI Discrimination

Fixing AI Discrimination

It is hard to believe that machines can discriminate. We hate discrimination. Discrimination is a phase of racism even when done by machines. As you walk on the streets of Lagos, you can easily notice Nivea advertisement sign-posts promising Nigerian women how a cream will make their skins “visibly fairer”. Apparently dark skin is not good enough.  That is a data point for the internet. If Nigerians prefer whiter skin, Google has taken note. That means beauty must be white in Nigeria.

Simply, if you ask Google and machines to judge in a beauty contest in Nigeria, it may call it for girls with fairer skin since it has datasets that many Nigerian women are using creams to turn their dark skins to white ones. AI (artificial intelligence) uses data and that data informs what it thinks is normal. Usually, the largest cohort of the datasets shapes its constructs of normality.

This company makes dark skin to become fairer and many Nigerian girls are believers

If you make a chatbot and feed it 20,000 messages with 19,000 of those messages racist and crude, the likelihood that the bot will see racism as being normal is there. That is why it is very easy to train any Twitter bot to be anything you want: saint, racist, etc. Just feed the data you want and over time, that bot correlates the most data as the new normal. Unless there are breakers in the design, there is nothing you can do about it, if you really want a near-natural bot.

That bots can be stupid does not make it right: developers have to find ways to mitigate that problem. Among many options, one easy way is to find a way to generate “balanced data”. For Africa, if we do not generate contents, allowing Silicon Valley, Paris and London to feed the new species of AI with only data from the Western World, the AIs will see them as the normal data. In other words, if the bot sees out of every 20,000 photos of girls, only 100 are dark-skinned, it may not necessarily capture dark photos as being normal. The system will default to white photos as the normal state. In some extreme cases, it may simply throw away the dark photos as totally non-human. Mitigating that problem will be feeding say 11,000 white and 9,000 dark photos. With that balanced datasets, the AI will have a better equilibrium.

That reminds me of a training I went while in the industry on IP protection. We were told to respond to email conversations via email instead of asking the person to talk things over. For example, if someone writes you capturing a statement like “I saw that Intel used this design and has a patent on it, there is a way I can get around the patent”. You do not tell your subordinate to see you for you to explain. It is better you write “Please if the design is patented, leave it and explore other designs”. The problem with talking it over without documenting is it that if bad things happen and there is litigation, what will make it to court is the written evidence. That is what the AI searching the emails will be fed with.

While that analogy: Africa and the black race will have to generate its own datasets to ensure machines can use same as they build the new data economy. Even if we are complaining of the obvious AI discrimination, without generating data, nothing will change. If you allow one cohort of people to be writing, talking and generating data, Google and the rest will think that the world is simply about those cohorts. That is why Amazon Alexa, a personal assistant AI, struggles with my Nigerian accent: it does not see that version of English a lot, so it is abnormal in its own world. It is not necessarily discriminating against me, it is just using datasets they have fed it to deconstruct my communication. Unfortunately, I am not sure they have any datasets from Nigeria.

Africa needs to create data to balance the game. Complaining on how machines dehumanize us will not fix the problem. It will only get worse unless we are ready to participate as technology creators and not just consumers who merely consume whatever they package for us.

Sure, that does not stop the makers from making sure decency rules in the market with circuit breakers to prevent situations where humans are classified as animals. No excuses on such failures!


---Visit our Store for my books, cases, etc. Now, enjoy our consolidated subscription for all contents (past, present and future).
-- We offer Advisory Services (tech, strategy & Africa).
---Sign-up to my Founders Mentoring (click here).

Share this post

One thought on “Fixing AI Discrimination

  1. Nice to see someone finally sees what I am trying to say to the AI community. Nigeria and blacks in general is underrepresented and its definitely going to affect us in an unforeseen future. I’m an AI researcher and developer. If you’re reading this and you are interested in having our own open-source data for Nigerian or blacks in general. Please do hit me up here https://www.linkedin.com/in/jeremiah-fadugba-5b3a5375/
    Datasets relating for both NLP and computer vision.

    Reply

Post Comment