Mention the word ‘forum’ to someone today. Their immediate thought will certainly be of an internet forum, with its members discussing anything from astrophysics to finding cheap offers in supermarkets. It’s a way of using the internet that goes back to the network’s very earliest days: even before the World Wide Web, early adopters talked and argued through services like USENET. And, while much online conversation has moved to social media and other spaces, it’s fair to say that forums laid the groundwork for those platforms, establishing expectations about how online interaction operates.
In short, it’s easy to talk and think about forums without being conscious of the word’s original meaning: in the Roman Empire, a forum was the public space built at the center of every town or habitation. Established principally as a place for people to buy and sell goods, the forum became the heart of Roman social life, as a place to meet, chat, and debate. Often, the modern idea of the Roman forum is one of the political and philosophical discussions – or, often, arguments.
Moderation through the ages
Just as behavior in a Roman forum would have been kept in check by social standards and peer pressure, early online forums relied on oversight from individual users with special privileges volunteering to maintain safety and civility.
Of course, the speed and scale of online conversation soon outstripped individuals’ abilities to keep up, and more formal solutions had to be found. Professional moderation teams are now common, and meeting the challenge also necessitated moves towards automated moderation, first in the form of word filters and, more recently, with AI-based approaches, an always-on solution to an always-on problem.
What does it mean to always be on for online speech? These systems don’t just need to oversee speech; they also need to adapt and learn to keep pace with the way language (especially on the internet) is adapted and evolved. ‘Forum,’ after all, is not the only word that has had its meaning changed by the digital age: one of the remarkable things about the internet is the speed at which it generates new words and new meanings for old words.
The crux of this problem lies in that, while automated digital systems tend to categorize the world into neat boxes, such as ‘acceptable’ and ‘unacceptable’ speech, language is fundamentally ambiguous.
For instance, the dictionary definition of the word ‘dead’ would make it a fairly unambiguously negative piece of speech. Posted as a reply under a joke, it (or a skull emoji) would actually signify an exaggerated way of saying that the user found it hilarious. This kind of ambiguity is rife, too, in online gaming, where ‘gg,’ meaning ‘good game,’ is used as a virtual handshake with one’s opponent. After a particularly one-sided match, saying ‘gg’ might be deeply antagonistic behavior.
While these are relatively light examples, the same pattern can be found in the darkest areas of online speech, where hate groups who are aware of the attempts that content moderators make to keep them out of communities will frequently change how their language is coded to disguise their intent.
The next frontier
All of this is coming in the context of a level of activity that, across the internet, amounts to billions or trillions of daily interactions. In these kinds of edge cases, the posted content can be too new and variable for AI-powered solutions to respond to and too voluminous for human operators to keep up with.
This is something which, as we look to a healthier future for content and content moderation, the industry as a whole will need to work on and take seriously. Historically, businesses have tended towards being over-cautious, preferring to block safe speech than accidentally allowing unsafe speech. However, the tide is turning on this route, and we think the next frontier for content moderation will be to take a consultative approach toward more subtle solutions.
As we talk to customers, we’re always keen to learn more about how their users use language, where it slips through the cracks of the systems we’re building, and what kind of insight they would need to manage user interactions more effectively – and if you’d like to join the conversation, we’d love to hear from you too.
In the meantime, the most effective approaches will draw on the full content moderation toolbox, applying word filters, AI oversight, and human intervention, where their respective strengths are most impactful. As we do so, we are building the insight, expertise, and experience to deliver an approach to content moderation that is fully alive with how language evolves.
The DSA: An executive summary of the new online rules for platform businesses
Here’s a friendly guide, an executive summary if you will, to be compliant with the Digital Services Act for online businesses.
Content filtering vs content moderation: The key to scaling
Content filtering is only part of the content moderation process, but it’s an important gatekeeper that allows platforms to scale. Let’s have a closer look.
Misinformation vs disinformation: What is the difference and how do they interact?
Learn about misinformation and disinformation, how they interact, how false information spreads, and an unusual example from 1835.
Report: The effects of fraud and poor content quality on online marketplaces
The new Besedo report highlights the effects of fraud and poor content quality on online marketplaces. Get easy-to-understand graphs, stats, and insights.
How to Qualify Marketplace Sellers to Improve Conversion and Retention Rates
Master the art of qualifying sellers for better conversions and long-term user loyalty. Dive in to elevate your marketplace game.