Twitter has ‘more analysis to do’ after algorithm shows possible racial bias
Twitter is learning first-hand about the challenges of eliminating racial bias in algorithms. The social network’s Liz Kelley said the company had “more analysis” to do after cryptographic engineer Tony Arcieri conducted an experiment suggesting Twitter’s algorithm was biased in prioritizing photos. When attaching photos of Barack Obama and Mitch McConnell to tweets, Twitter seemed to exclusively highlight McConnell’s face — Obama only popped up when Arcieri inverted the colors, making skin color a non-issue.
Others tried reversing photo and name orders to no avail. A higher-contrast smile did work, Intertheory’s Kim Sherrell found. Scientist Matt Blaze, meanwhile, noticed that the priority seemed to vary depending on the official Twitter app used. Tweetdeck was more neutral, for instance.
Kelley said that Twitter had checked for bias before using the current algorithm, but “didn’t find evidence” at the time. She added that Twitter would open source its algorithm studies to help others “review and replicate.”
There’s no guarantee that Twitter can correct this. However, the experiment does show the very real dangers of algorithmic bias regardless of intent. It could shove people out of the limelight, even if they’re central to a social media post or linked news article. You might have to wait a long while before issues like this are exceptionally rare.
thanks to everyone who raised this. we tested for bias before shipping the model and didn’t find evidence of racial or gender bias in our testing, but it’s clear that we’ve got more analysis to do. we’ll open source our work so others can review and replicate. https://t.co/E6sZV3xboH
— liz kelley (@lizkelley) September 20, 2020
(20)