Fair moderation is hard but fair, scalable moderation is harder
I'm Alice Hunsberger. Trust & Safety Insider is my weekly rundown on the topics, industry trends and workplace strategies that trust and safety professionals need to know about to do their job.
This week, I'm thinking about content moderation at scale, and what role frontline moderators have to play.
Thanks to everyone who reached out about last week's edition– it seems like it really resonated with many of you. Get in touch if you've got ideas about what I should write about next.
Here we go! — Alice
The recent takedown of Kidflix — one of the largest paedophile platforms in the world — revealed a chilling truth: child exploitation online is no outlier - it’s engineered into the infrastructure of the internet.
With CSAM spreading through advanced tech and everyday platforms, safety can’t be an afterthought. Resolver Trust & Safety works at the frontline, helping platforms design systems that detect and disrupt abuse before it scales.
Because the next Kidflix is already being built - and the time to act is now.
What it would take to prioritise quality in content moderation
We often talk about moderation as a numbers game: how fast can we respond, how much can we automate, how many decisions can we review per minute. But for those of us who’ve worked across the system, the real question is more fundamental: how do we know if we’re getting it right?
A recent research paper — The Role of Expertise in Effectively Moderating Harmful Social Media Content— has looked at this question in the context of real-world harm and genocide. The authors investigated social media moderation during times of conflict and genocide, focusing on posts targeting Tigrayans during the 2020-2022 Tigray war. It came to the interesting conclusion that:
“open discussions and deliberation meetings where moderators can exchange contextual information and resolve disagreements to be more effective ways of ensuring that harmful content is appropriately flagged.”
They did this by giving moderators space to think more deeply, take context into account, exercise their judgment, and then spend as much time as they needed to make the decisions themselves:
"We decided on 55 posts per week to restrict the participants’ exposure to harmful content to a maximum of one hour per day, based on feedback from participants and the psychological impacts this content can cause. After each round of annotation, our expert annotators spent a minimum of 60 minutes discussing their disagreements and arrived at a final agreed-upon label for each post after deliberation."
What was notable to me was that, even after this lengthy deliberation step, there were still significant disagreement; as high as 71% and, after several rounds, a still very high 33%. This shows how difficult making moderation decisions is.