True, we’re starting to have some of the necessary tools, like toxicity labeling ML models via API. But behavior norms are shaped by communities and what’s toxic in one context might be fine in another, and vice versa.
-
-
Show this thread
-
Calibrating something like this to your specific community or platform is not trivial, and even once you have calibrated the model, there’s the question of what you do with its output once you’ve got it. Is it automatically actioned? Reviewed by a human?
Show this thread -
Then you have to measure the impact of your policy, a task which is a sprawling nightmare in and of itself. Do you focus on business impact, the effect on your users? How do you evaluate the impact of someone who made an off-color joke vs. someone who’s a continual net negative?
Show this thread -
Obviously, there aren’t a lot of standards or best practices around this yet. We wouldn’t be having controversies like the ones mentioned in this article if we did.
Show this thread -
As people continue to live more of their lives online, strewing UGC in their wake wherever they go, the conversation about moderation standards will be forced. I’ll be curious what kind of winding road it ends up taking
Show this thread
End of conversation
New conversation -
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.

