Parallel Worlds of Moderation: Simulating Online Civility with LLMs
Opening — Why this matters now Every major platform claims to be tackling online toxicity—and every quarter, the internet still burns. Content moderation remains a high-stakes guessing game: opaque algorithms, inconsistent human oversight, and endless accusations of bias. But what if moderation could be tested not in the wild, but in a lab? Enter COSMOS — a Large Language Model (LLM)-powered simulator for online conversations that lets researchers play god without casualties. ...