Editorial note: This blog was first posted on July 9, 2019, and last updated December 2, 2020, to reflect additional changes made to our rules against hateful conduct.
We create our rules to keep people safe on Twitter, and they continuously evolve to reflect the realities of the world we operate within. Our primary focus is on addressing the risks of offline harm, and research* shows that dehumanizing language increases that risk.
As we develop the Twitter Rules in response to changing behaviors and challenges with serving the public conversation, we understand the importance of considering a global perspective and thinking about how policies may impact different communities and cultures. Since 2019, we’ve prioritized feedback from the public, external experts, and our own teams to inform the continued development of our hateful conduct policy.
Expanding our hateful conduct policy
While we encourage people to express themselves freely on Twitter, abuse, harassment and hateful conduct continue to have no place on our service. In July 2019, we expanded our rules against hateful conduct to include language that dehumanizes others on the basis of religion or caste. In March 2020, we expanded the rule to include language that dehumanizes on the basis of age, disability, or disease. Today, we are further expanding our hateful conduct policy to prohibit language that dehumanizes people on the basis of race, ethnicity, or national origin.
We will require Tweets like those below to be removed from Twitter when they are reported to us. We will also continue to surface potentially violative content through proactive detection and automation. If an account repeatedly breaks the Twitter Rules, we may temporarily lock or suspend the account; more on our range of enforcement options here.
The Twitter Rules help set expectations for everyone on the service and are updated to keep up with evolving online behaviors, speech, and experiences we observe. In addition to applying our iterative and research-driven approach to the expansion of the Twitter Rules, we’ve reviewed and incorporated public feedback to ensure we consider a wide range of perspectives.
With each update to this policy, we’ve sought to expand our understanding of cultural nuances and ensure we are able to enforce our rules consistently. We’ve benefited from feedback from various communities and cultures who use Twitter around the globe. Consistent feedback we’ve received includes:
That said, even with these improvements, we recognize we will still make mistakes. We are committed to continuing to work to further strengthen both our enforcement process and our appeals process to correct our mistakes and prevent similar ones moving forward.
Our trusted partners
We realize that we don’t have all the answers, so in addition to public feedback, we work in partnership with our Trust & Safety Council as well as other organizations around the world with deep subject matter expertise in this area.
As part of this update, we convened a global working group of third-party experts to help us think about how we could appropriately address dehumanizing speech around the complex categories of race, ethnicity, and national origin. These experts helped us better understand the challenges we would face and to ultimately answer questions like:
"Our work in local communities helps us think critically about ways to ensure social cohesion between the diverse communities we serve, institutions, and policymakers. As a trusted partner, we worked with Twitter to ensure that cultural and regional nuances relevant to migrant groups were accounted for in their hateful conduct policy update." — Roses of Peace, an interfaith organization that aims to build a cohesive and resilient Singapore
We’ll continue to build Twitter for the global community it serves and ensure your voices help shape our rules and how we work. As we continue to look for opportunities to evolve and expand our policies to better handle the challenges we’re currently facing, we’ll update you on what we learn and how we plan to address it. We’ll also continue to provide regular updates on all of the other work we’re doing to make Twitter a safer place for everyone via @TwitterSafety.
*Examples of research on the link between dehumanizing language and offline harm: