7 guidelines for online content moderation
How do we monitor online content without limiting free speech? Alternately, how can we circumscribe the non-sociable aspects of online participation when we want to dig deeper into hot issues in engagement projects – issues that can draw on deeply held, emotional views?
Content moderation is vital to any online engagement. From the automated moderation of inappropriate content to sanctioning “grey areas” that require human intervention. Effective content moderation buoys robust – and reflective – considerations on important public issues and facilitates participants willingness to get online and join in conversations that impact their lives. It also ensures a sense of accessibility and inclusion – uppermost to all online engagement.
While software selection is crucial, integrating online engagement and moderation shapes our experience of digital deliberation. Deliberation is a social process, potentially involving many people, which requires participants to be exposed to information that is both broad and deep. Given this fact, ‘emotional connection’ is a necessity to spark participation. Therefore, the ‘design’ and ‘management’ of the space is critically important.
Unlike monologue or debate, dialogue is what happens when participants start to read and respond to each other’s comments. They ask questions and they build on ideas. They may challenge arguments or assertions, but they do so to better understand the rationale or the underlying belief, or background story. There is mutual respect, and there is a focus on “solutioneering”.
Online content moderation, then, is crucial.
A pragmatic set of guidelines, etiquettes and sanctions provides the best possible experience for participants to engage and explore issues. These work to optimise the issue and listen to nuances within any given community context, allowing everyone to have their say without fear, intimidation or retribution.
Used to describe the act of “rule-keeping”, moderation ensures participant comments are within the site moderation rules (is there any bad language, is the comment sexist, racist, homophobic, does it contain any links to bad content?) and checks for hectoring and intolerant behaviour. Basically, moderators look for any content that might drive participants away from the process out of fear of being attacked, or it stops being a constructive space. Where software analyses comments for bad language and spam, human intervention analyses comments for more nuanced breaches of the rules.
Seven guidelines for online content moderation
1. Acceptable behaviour
You must have a clear set of rules that bound acceptable behaviour for user generated contributions. These may vary from project to project, but include references to:
- posting personal information,
- naming organisational staff, particularly in a negative light,
- defamatory content,
- acceptable language,
- bullying, hectoring and insulting,
- external links,
- advertising, and
- comments on moderation policies and processes.
2. Breaching moderation
You must also have a clear set of sanctions for breaching the moderation rules. For example:
- content removal,
- content editing,
- temporary suspension of access privileges, and
- permanent blocking of access privileges.
You should consider including a set of guidelines for appropriate etiquette in the context of your particular project. These are, in the main, to promote positive behaviours, rather than to control poor behaviours, and may include broader instructions like “be respectful”, and specific education like, “avoid CAPS LOCK”.
4. Post-hoc moderation
Dialogue works best when it is allowed to flow, so you must find a way to use “post-hoc” moderation. That is, moderation, AFTER the comment (or content) has been allowed to go live on the site.
5. Protocols around comments
Depending on the perceived “risk” of user-generated content egregiously breaching the site rules, you will need to tighten or loosen the protocols around the “comment review period”. Very low risk issues and groups may require almost no moderation, whereas highly emotional and politically contested issues may require real-time 24/7 human oversight.
6. Automated and human filters
Your moderation should include BOTH automated filtering AND human systems. Automated filters are good at picking up black-listed words and SPAM, they are incapable of picking up other poor behaviours.
7. ‘Back up’ processes
Your moderation system should also include “back-up” processes, such as “community flagging”, because your moderators may not be familiar with all of the nuances of the issues under consideration, and may not, therefore, pick up all of the issues.
At Granicus, we independently moderate all of our client’s EngagementHQ sites to keep the conversation safe and on topic.