Twitter and Zoom were over the weekend spotted to have a racial bias in their visual algorithms. It all started when someone noted how Zoom appeared to be removing the head of people with a darker skin pigmentation when they use a virtual background, while it does not make this move on people having a lighter skin pigmentation. In the tweet used to report the Zoom issue, it was ironically spotted that Twitter too appears to have a racial bias when it cropped thumbnails to favour the face of a white person over a black one. Twitter has responded to the outrage that emerged, saying it was clear it had more work to do.
Zoom initially appeared to have a problem with its virtual background algorithms that manifest as a racial bias. Researcher Colin Madland posted a thread on Twitter on Saturday that underlined the issue with the face-detection algorithm that allegedly erases black faces when applying a virtual background on the video conferencing app. Gadgets 360 has reached out to Zoom for getting a clarification on the algorithm and will update this space when the company responds. In the same thread, with Madland posting photos of each user in the chat, Twitter’s image thumbnail cropping algorithm seemed to be favouring Madland over his black colleague.
In response to Madland’s observations, Twitter Chief Design Offer Dantley Davis said “It’s 100 percent our fault. No one should say otherwise. Now the next step is fixing it.”
Soon after, several Twitter users posted photos on the microblogging platform to highlight the apparent bias. An example was cryptographic engineer Tony Arcieri, who on Sunday tweeted the mugshots of former US President Barack Obama and senate majority leader Mitch McConnell to understand whether the platform’s algorithm would highlight the former or latter. Arcieri used different patterns of putting the mugshots in the images, but in all cases, Twitter showed McConnell over Obama.
Trying a horrible experiment…
Which will the Twitter algorithm pick: Mitch McConnell or Barack Obama? pic.twitter.com/bR1GRyCkia
— Tony “Abolish (Pol)ICE” Arcieri 🦀 (@bascule) September 19, 2020
However, once the engineer inverted the colours of the mugshots, Obama’s image showed up on the cropped view. Intertheory producer Kim Sherrell also found that the algorithm tweaks the preference once the image of Obama is changed with a higher contrast smile.
Some users also found that the algorithm appears to give focus to brighter complexions even in case of cartoons and animals.
I wonder if Twitter does this to fictional characters too.
Lenny Carl pic.twitter.com/fmJMWkkYEf
— Jordan Simonovski (@_jsimonovski) September 20, 2020
I wonder if Twitter does this to fictional characters too.
Lenny Carl pic.twitter.com/fmJMWkkYEf
— Jordan Simonovski (@_jsimonovski) September 20, 2020
Twitter spokesperson Liz Kelley responded to the tweets raising racial bias allegations against the platform and said, “We tested for bias before shipping the model and didn’t find evidence of racial or gender bias in our test, but it’s clear that we’ve got more analysis to do.” She added saying, “We’ll open source our work so others can review and replicate.”
Back in 2017, Twitter discontinued face detection for automatically cropping images in users’ timeline and deployed a saliency detection algorithm that was aimed to focus on “salient” image regions. Twitter engineer Zehan Wang tweeted that the team conducted some bias studies before releasing the new algorithm and at that time found that there was “no significant bias between ethnicities (or genders).” However, he added that the company would review the study provided by Twitter users.