Reddit quarantined: can changing platform affordances reduce hateful material online?

: This paper studies the efficacy of the Reddit’s quarantine, increasingly implemented in the platform as a means of restricting and reducing misogynistic and other hateful material. Using the case studies of r/TheRedPill and r/Braincels, the paper argues the quarantine successfully cordoned off affected subreddits and associated hateful material from the rest of the platform. It did not, however, reduce the levels of hateful material within the affected spaces. Instead many users reacted by leaving Reddit for less regulated spaces, with Reddit making this hateful material someone else’s problem. The paper argues therefore that the success of the quarantine as a policy response is mixed.

including banning: any suggestive or sexual content featuring minors (u/reddit, 2012); the sharing of nude photos without the subject's consent (u/kn0thing, 2015 1 ); attacks and harassment of individuals (u/5days et al., 2015); the incitement of violence (u/landoflobsters, 2017); and attacks and harassment of broad social groups (u/landoflobsters, 2019). Reddit's argument, that it was a free speech site that would not intervene, has slowly come undone.
Reddit has frequently found itself in an uncomfortable position, straddling a fine line between maintaining a culture of free speech, while at the same time not allowing the flourishing of violence and hateful material (Copland, 2018). Reddit has faced dueling pressures from a user base, which finds Reddit's anything goes approach as core to the platform's appeal, versus other users who increasingly wish to constrain much of the more extreme material on the site (Massanari, 2015).
In doing so, in addition to the content policies identified above, Reddit has also implemented a unique approach to dealing with hateful content on the platform, that is the quarantine function.
Content moderation from large digital platforms, in particular in regards to hateful material, comprises an array of different techniques (Gillespie, 2018). As noted already, many platforms have become more active in deleting hateful content and banning or suspending the profiles of far-right speakers. In addition to this, many platforms have begun to engage in fact-checking processes in order to stop the spread of disinformation. Facebook, for example, has created an independent factchecking unit (Facebook, 2020), while Twitter made headlines by fact-checking a Tweet from President Donald Trump (BBC News, 2020). Platforms also rely on the labour of users themselves in order to moderate content (Matias, 2019). Many platforms, including Reddit, contain communities which self-moderate, with volunteer moderators regulating content based on community-defined rules. Users also frequently play the role of content monitors, reporting content to the platform for moderation to occur (Gillespie, 2018).
Reddit's quarantine function is unique from these approaches as it is designed to limit the accessibility of users to identified subreddits, and in turn to limit the spread of questionable content, while not banning the content outright. A quarantine uses the platform's affordances to discourage the spread of particular material and to encourage positive behaviour change within identified shocking subreddits.
1. On Reddit users create profile names, which are normally pseudonymous. Users are then identified by a u/(username) -i.e., u/kn0thing. I therefore reference content posted on Reddit by the username, identifying the author, followed by the year of the post or comment.
The quarantine therefore provides a useful moment to examine the efficacy of the use of bans and restrictions on particular content in both limiting the spread of hateful content as well as encouraging changes in behaviour within a platform.
Using a case study of two men's rights subreddits quarantined after the revamp of the function in September 2018 (u/landoflobsters, 2018), r/TheRedPill and r/ Braincels, this paper examines the efficacy of this approach in dealing with misogynistic content on the platform.
Using data from these two subreddits the paper argues that the quarantine was effective in limiting engagement with these subreddits and in turn limiting the exposure of the broader Reddit community to the misogynistic material within. In particular, the quarantine saw significant drops in activity within the targeted subreddits. However, I argue that despite this, the quarantine had little to no effect in the use of hateful language within the targeted subreddits, with Reddit failing to encourage remaining users to change their behaviour. Instead, the quarantine furthered animosity between users of these subreddits and Reddit overall, in particular resulting in concerted pushes for users to leave Reddit, to instead participate in self-moderated, and in turn less restrictive, platforms. Aligning with research on subreddit bans from Chandrasekharah, Pavalanathan, Srinivasan, Glynn, Eisenstein and Gilbert (2017, p. 18) I argue that the quarantine in turn, potentially made hateful material someone else's problem. While the policy reduced access to hateful material on the platform, it did not necessarily do so on the internet overall, instead potentially pushing it to less moderated spaces where it can continue unchecked.
As a policy response therefore, this paper argues that the quarantine had mixed results.

The quarantine
On 28 September 2018, Reddit announced a revamp of their quarantine function (u/landoflobsters, 2018). A quarantine implements a range of restrictions on a subreddit. Quarantined subreddits are unable to generate revenue, and their content does not appear on Reddit's front page, nor on other non-subscription based feeds.
Quarantined subreddits cannot be found via the search or recommendation function, meaning users need to either search for the subreddit via Google or know the subreddit URL to access it. Quarantined subreddits also do not appear on a Reddit user's subscription feed, requiring subscribers of that subreddit to note that it has been quarantined and then actively seek it out to once again put it in their feed.
When a user does attempt to gain access a warning is displayed (see Figure 1), informing users that the subreddit is dedicated to "shocking or highly offensive con-tent" and requiring the user to explicitly opt-in to participate.

FIGURE 1: Quarantine warning
As already noted, the implementation of the quarantine was part of a broader sweep of policy changes from the platform, ones which shifted Reddit away from their unlimited free-speech ethos. These shifts occurred primarily for two reasons.
First, Reddit has not been immune to the significant social pressure facing social media companies over their role in public discourse, particularly following the 2016 US presidential election (Marantz, 2019). This has resulted in the co-founders of Reddit Alexis Ohanian and Steve Huffman to publicly walk back much of their early ideas about how Reddit should operate as they reckoned with the influence of their platform (Marantz, 2019). In addition to this, Reddit has begun to rely more heavily on advertising for revenue (Castillo, 2018). This shift brought particular challenges, with advertisers expressing concerns about being associated with the hateful material that appears on the platform. As the forecasting director for eMarketer Monica Peart has argued: As a mix of forum and trending news site with a bit of social network, reddit has operated on an ' open internet' ethos. While that has yielded organic growth among a hard-to-reach audience, it has also meant a reality where controversial content is the norm. And in a news climate where missteps can tarnish results, that makes some digital advertisers nervous. (Peart, 2019) The quarantine is an attempt to restrict hateful material through an intervention into the affordances of a digital platform. As Davis and Chouinard (2016, p. 242) argue, an "affordance refers to the range of functions and constraints that an object provides for, and places upon, structurally situated subjects". In relation to studies of digital platforms, affordances are the functions of the platform that shape how it can, and cannot, be used. Davis and Chouinard (2016) provide categories to allow us to understand how affordances work. These mechanisms define the different ways in which the functions of an artefact define the usage of that artefact. Davis and Chouinard propose that affordances request, demand, allow, encourage, discourage, and refuse: Requests and demands refer to bids that the artefact places upon the subject. Encouragement, discouragement, and refusal refer to how the artefact responds to a subject's desired actions. Allow pertains to both bids placed upon on the subject and bids placed upon the artefact. (Davis and Chouinard, 2016, p. 242) The quarantine function represents an attempt from Reddit to use the architecture of the platform to both discourage and encourage particular behaviours from users. The intent of quarantining subreddits was twofold. Reddit aimed first to limit the spread of content from these spaces, discouraging engagement with identified shocking content. Secondarily however, Reddit also aimed to use the quarantine as a way to encourage positive change within a community, hoping that the process of shutting off communities would encourage users to change their behaviour as a way to re-engage with the broader space. As u/landoflobsters said: The purpose of quarantining a community is to prevent its content from being accidentally viewed by those who do not knowingly wish to do so, or viewed without appropriate context. We've also learned that quarantining a community may have a positive effect on the behavior of its subscribers by publicly signaling that there is a problem. This both forces subscribers to reconsider their behavior and incentivizes moderators to make changes. (u/landoflobsters, 2018) Case studies Since announcing the revamp of the quarantine function in September 2018, Reddit administrators became more active in using the function (Copland and Davis, 2020 This paper studies the quarantining of two manosphere subreddits, r/TheRedPill and r/Braincels, both of which were quarantined on the same day as the announcement of the revamp of the quarantine function in September 2018. r/ TheRedPill and r/Braincels are associated with a broad online community called the manosphere, which, primarily focused on discussions around sex and relationships, is broadly anti-feminist in its ideology (Marwick and Caplan, 2018). The manosphere is based in an ideology, described by Nicholas and Agius (2018) as masculinism, which argues that men have become subjugated by feminism, with this subjugation in particular impacting men's access to sexual and romantic relationships.
r/TheRedPill and r/Braincels engage in these questions in two different ways.
Members of r/TheRedPill focus on self-help and learning game; a range of techniques used to identify and pick up women confidently. Members of r/Braincels, part of a broad sub-community called incels, believe that due to a range of factors, primarily associated with their looks, that women refuse to engage in romantic or sexual relationships with them. These men are, by and large, resigned to their singledom (although some hold the more extreme view that they are owed sex by women), and come to manosphere forums to complain and bond about their lack of relationships with women, as well as to complain about women in general. Both subreddits have been identified in literature to contain high levels of misogynistic and other hateful language (e.g., Farrell et al., 2019;Van Valkenburgh, 2019).
It is this misogyny in particular which resulted in the administrators of Reddit quarantining the subreddits. When implementing the quarantine Reddit administrators sent the moderators of each subreddit messages informing them of why they have been quarantined and the avenues they had to appeal the decision. One of the volunteer moderators of r/TheRedPill, u/redpillschool, published the message sent to them. The message from Reddit Administrators was headlined, "This community is quarantined: It is dedicated to shocking or highly offensive content.
For information on positive masculinity, please see the resources available at Stony Brook University's Center for the Study of Men and Masculinities, " with the following explanation provided: In this case the quarantine was applied for the high degree of misogyny present in this subreddit. To be removed from quarantine, you may present an appeal here. The appeal should include a detailed accounting of changes to community moderation practices. (Appropriate changes may vary from community to community and could include techniques such as adding more moderators, replacing certain moderators, creating new rules, employing more aggressive auto-moderation tools, adjusting community styling, etc.) The appeal should also offer evidence of sustained, consistent enforcement of these changes over a period of at least one month, demonstrating meaningful community transformation. (u/redpillschool, 2018) Alongside the broader goal of restricting access to these subreddits, Reddit therefore articulated explicitly that it wanted to see a reduction of misogyny within these spaces, presenting this as the way in which both subreddits could exit their quarantine status.

Data analysis
Using these two case studies, the purpose of this paper is to examine the internal and external effects of this policy. First, I study whether the quarantine was an effective way to achieve Reddit's stated goals of discouraging engagement with hateful material and encouraging shifts towards more positive content within subreddits that foster this material. Beyond this, the article asks, can shifting the affordances of a platform for particular users help reduce misogynistic and hateful material on the internet external to Reddit?
As part of a broader PhD research project I have used the website pushshift.io to download all submissions and comments from both r/TheRedPill and r/Braincels from the start of November 2017 to the end of December 2018. The implementa-tion of the quarantine occurred during my data collection period, providing an opportunity to study the quarantine with an existent large data set. I have examined this data set in three different ways in order to understand the efficacy of the quarantine as a policy approach.

User activity level
I start by examining how the quarantine impacted the level of engagement within affected subreddits. This allows me to see whether Reddit achieved their first goal, which was to limit the exposure of the broader Reddit community from the hateful content within these spaces, and in turn to limit the spread of this material to other parts of Reddit. To analyse this, I created comment frequency histograms of the two affected subreddits ( We can see distinct trends in both histograms. Figure 2 is a histogram of the number of submissions and comments per day for r/TheRedPill. r/TheRedPill is established at the beginning of the data collection, with activity levels already high at this point. At its peak across the data collection r/TheRedPill has just over 1,500 submissions and comments on a single day. Submission and comment numbers on r/TheRedPill fluctuates widely on a day to day basis, with frequent highs and troughs. Activity level however drops significantly after the implementation of the quarantine. Despite continued peaks and troughs, these levels overall are approximately one half of what occurred before the quarantine.  Based on this analysis we can see that, at least in the immediate term, the quarantine achieved Reddit's first goal -i.e., to reduce engagement with particular subreddits and in turn exposure to these subreddits from the broader Reddit community. We can see that this occurred in both affected subreddits, both of which received a drop in activity of about 50%. This analysis therefore shows that the quarantine had an initial, immediate, success. However, while it shows that these sub-reddits were cordoned off effectively from the rest of Reddit, this does not necessarily prove that it reduced hateful content on Reddit overall. It is entirely possible that users of the affected subreddits simply shifted their hateful content to other parts of the platform. It is unfortunately outside of the scope of this paper to examine these potential broader shifts.

Misogynistic language analysis
While Reddit effectively cordoned off quarantine communities, significantly reducing engagement within them, the platform also had a second goal, which was to encourage users who stayed, to change their behaviour. Therefore, I will now shift to examining the discourse within the affected subreddits to see whether this occurred. What I am looking for here is whether there was a reduction in misogynistic and other hateful language within the two affected subreddits following the quarantine. Any reduction would suggest that the implementation of the quaran- LGBTIQ people, queer is also a term that has been reclaimed by many members of this community. This approach does not distinguish between these two uses. Despite this however, a linguistic inquiry approach can provide an insight into the type of language used in a community, and in this case whether there is a high prevalence of terms closely associated with misogynistic and other hateful material.
I have applied a linguistic approach to study the prevalence of misogynistic language with each subreddit in the two months prior to and after the implementation of the quarantine, totalling four months of data analysis. This is done to examine whether there were shifts in the use of hateful language across this period, and in turn whether Reddit was successful in encouraging users of the subreddit to change their behaviour and language. The results, presented in Tables 1 and 2, present the total number of submissions and comments that contain terms from the misogynistic dictionary, as well as the prevalence of each subcategory as a percentage of the total number of submissions and comments.  Prior to the quarantine, misogynistic language appeared in 24.74% and 27.55% of submissions and comments in r/TheRedPill and r/Braincels respectively. In the two months prior to the quarantine r/TheRedPill had a total of 9,504 submissions or comments containing misogynistic language, while r/Braincels had a total of 111,845. These vast differences largely represent the different volume of content within each subreddit. The most common form of misogyny within r/TheRedPill was hostility, being prevalent in 11.33% of submissions and comments. For r/ Braincels, stoicism appeared in 14.52% of submissions and comments. This is unsurprising as the stoicism category contains a number of terms directly associated with incel communities, including incel, volcel, ricecel, truecel, rope and cope.
Alongside the overall reduction in activity in both subreddits, there is a clear reduction in the raw number of submissions and comments containing misogynistic language in the immediate aftermath of the quarantine. r/TheRedPill sees a drop in the number of misogynistic submissions and comments from 9,504 to 4,710, and r/Braincels sees a similar drop from 111,854 to 58,907. However, this does not result in any shift in the density of this language within those who remained in the subreddit. Both subreddits see a slight increase in the percentage of submissions and comments containing misogynistic language with this increasing from 24.78% to 25.58% in r/TheRedPill and 27.55% to 27.97% in r/Braincels. Both of these shifts are too small however to suggest that we can draw any significant conclusions from these changes. Similar to the period prior to the implementation of the quar-antine, hostility and stoicism were the most common forms of misogyny in r/ TheRedPill and r/Braincels respectively.
Therefore, similar to engagement overall there is a sharp reduction-approximately a halving-in the raw number of submissions and comments containing misogynistic language within the subreddits. However, those who remained engaged with the subreddits did not change their language. The percentage of the misogynistic material stayed the same in both subreddits, suggesting that existent users continued using the same language as before, only at a smaller scale. Importantly however, there is also no significant increase in misogynistic language in either subreddit. While Reddit successfully reduced engagement with both subreddits from the broader Reddit community, they did not encourage changes in the use of language within either of these subreddits-either negatively or positively.

Responses to the quarantine
My first two areas of analysis studied the outcomes of the particular goals of the quarantine identified by Reddit itself. However, the impact of the quarantine did not just occur in these two areas. In particular, following its implementation, both

r/TheRedPill
The quarantine was met with immediate anger from members and moderators of r/TheRedPill, who argued that the quarantine represented an attack on their community and on men as a whole.
This manifested in two ways. First, subreddit moderators openly challenged and discredited the logic behind the quarantine and petitioned Reddit to either provide more evidence behind why r/TheRedPill deserved to be quarantined, or to overturn the implementation of the quarantine altogether. The most highly engaged of I challenge the admin, /u/spez, or anybody on their team to engage me in courteous, reasoned discussion to defend their assertions and endorsements this week, or to apologize and lift the quarantine. I promise to keep it 100% civil in an all-text debate of ideas with anybody on the reddit administration. I get the feeling nobody is going to be taking me up on this offer. This would be the action of somebody with integrity. I'm sure Michael Kimmel could make an argument that integrity is part of positive masculinity. /u/spez, you don't want to be toxic do you?(u/redpillschool, 2018b) The quarantine thus resulted in a significant split between moderators and users of r/TheRedPill and Reddit as a platform. This led to the second strategy used by r/ TheRedPill moderators. Through participating in attacks on men through the quarantine, r/TheRedPill moderators argued that Reddit could no longer be trusted as a space in which the community could gather. There are a number of comments in particular in the data set, in which users said that they believed that the quaran-tine was simply the starting point, and that Reddit would soon ban the subreddit altogether.
In response to these fears, the moderators of r/TheRedPill directed users to a selfmoderated platform outside of Reddit 2 . This is an independently established and self-moderated forum with similar site infrastructure to Reddit, but with a sole focus on Red Pill ideas. The forum is moderated by the same moderators as r/ TheRedPill. Users of r/TheRedPill were actively encouraged to make profiles with the same username they had on r/TheRedPill, ensuring continuity of discussion.
This push occurred primarily through a recurring comment by the AutoModerator (see Figure 5). Subreddit moderators have the capacity to configure AutoModerators to automatically remove particular posts or comments, or to reply to posts or comments with predetermined messages. Starting eleven days after the implementation of the quarantine, moderators of r/TheRedPill configured the AutoModerator to reply as the first, pinned, comment on every post informing users that the subreddit has been quarantined and directing users to the external Red Pill site (see Figure 4). This does not mean that users necessarily flooded to this outside platform, and unfortunately I am unable to conduct analysis to see how many actually followed the moderators directions and left Reddit for the external forum. This is because I do not have the relevant user data for this outside forum, and in turn cannot trace user migration from Reddit to this platform. However, research on previous bans of hateful subreddits have shown the bans have resulted in a large number of users of affected subreddits moving to different, often less regulated, platforms (Chandrasekharan et al., 2015;Newell et al., 2016). These shifts occurred without the concerted efforts to do so as we saw in r/TheRedPill following the quarantine.
While the specific data is not available in this case, it is reasonable to think that the quarantine resulted in some, or many, users of the subreddit moving to this external platform.

r/Braincels
While the responses from r/TheRedPill was coherent and organised, primarily emanating from moderators and other senior figures, the response from r/Braincels was blurrier. The first responses to the quarantine for example do not come from the moderators, but instead from regular users. Users quickly started asking questions and suggesting potential ways forward. Some for example, asked whether users should go elsewhere, others suggested it may end up being good for the subreddit as it would mean only dedicated incels would stay in the space, and some highlighted fears that the quarantine would eventually result in the subreddit being banned (a fear that was realised when r/Braincels was banned from Reddit in October 2019).
When the moderators engaged they did not suggest any means forward, instead providing basic information. The first engagement from moderators came from u/ Cristalline144hz, who on the day the quarantine was implemented posted a range of basic information about what it meant. This was followed by a lively discussion about how r/Braincels should respond. Similar to arguments prominent in r/ TheRedPill, some, such as u/Uqtpa argued that the quarantine represented the misandry of Reddit, which was unfairly targeting men. As they argued: It's ridiculous. There is far more misandry in subreddits such as againstmensrights, the- Here the quarantine is perceived as a positive development, and in fact, for some the implementation of the quarantine showed that the subreddit was on the right track. Some users argued that the quarantine would allow r/Braincels to stop being so "cucked" and in turn to return to its roots. As u/FireAlarm911 argued: I'd rather remain quarantined than "change the direction of the sub". It was already getting too cucked in here, conforming to their wishes will kill this sub more than quarantine, I'm sure they consider pretty much anything critical of women as "mysoginist".

Conclusion
This paper has studied the efficacy of the quarantine as a means of restricting and reducing hateful and misogynistic material on large digital platforms.
As already articulated, the goal of the quarantine from Reddit's perspective was twofold: to cordon off affected subreddits from the broader platform, in turn reducing access to hateful material, and to change discourse within affected subreddits in a more positive direction. Based on the analysis presented, the results of the quarantine are mixed. The quarantine resulted in a drop in engagement within both affected subreddits by approximately one half. This suggests that Reddit was successful in reaching its first goal. Based on the linguistic analysis however, those who continued to participate did not reduce their levels of misogyny, highlighting a failure in reaching goal two. This failure could, in part, be explained by how users responded to the quarantine in the first place. In both instances there was a large amount of anger and concern at the implementation of the quarantine, with this anger in particular manifesting in a campaign in r/TheRedPill to encourage users to move to an external platform. Some users in r/Braincels even celebrated the quarantine as a sign that the subreddit was heading in the right direction, arguing they would prefer to stay quarantined so they did not have to change their language. While some users did suggest the quarantine highlighted a need for change, these voices were in the minority, as highlighted by the lack of change in language within both affected subreddits.
What does this mean for the use of technical and editorial means to restrict hateful content on digital platforms? Digital platforms such as Reddit increasingly sit in an uncomfortable position. Large platforms face increasing pressure to ban or heavily moderate hateful material, often finding a need to strike a balance between pleasing users and advertisers, adhering to regulations, and maintaining their role as spaces of free speech (Gillespie, 2018).
In managing these tensions, platforms act as silos. The focus of the platform-wide moderators of Reddit is on managing hate speech on Reddit and Reddit alone.
There is little to no concern placed on the broader implications of moderation decisions on content on the web, overall. Acting as a silo therefore, Reddit can easily see the quarantine largely as a success. While the quarantine did not change the density of misogynistic material with either subreddit, it did reduce the levels of this material significantly in raw numbers. Reddit successfully isolated these communities from the rest of the platform, likely reducing the overall level of misogynistic material on the platform (although this platform-wide analysis has not been tested in this paper, meaning this is a speculative prediction).
This does not necessarily mean however that the quarantine reduced the levels of misogyny and other hateful material on the web overall. In particular, the quarantine resulted in a significant campaign from moderators of r/TheRedPill to push users to an external site, one managed entirely by Red Pill moderators. In this site users were, and remain, able to engage in misogynistic and other hateful behaviour with no external moderation at all. This site is restriction-less. While we do not have data as to how many users actually migrated to this platform, in the case of r/TheRedPill, the effect of the quarantine was potentially similar to previous bans of subreddits on the platform. In analysing these bans, Chandrasekharah et al. argued: "In a sense, Reddit has made these users (from banned subreddits) someone else's problem. To be clear, from a macro perspective, Reddit's actions likely did not make the internet safer or less hateful. One possible interpretation, given the evidence at hand, is that the ban drove the users from these banned subreddits to darker corners of the internet" (Chandrasekharah et al., 2017, p. 18).
These moves occurred alongside a significant debate within right-wing and extremist organisations, with a range of high-profile individuals, in particular in the United States, arguing the right should create and migrate to their own platforms (Marantz, 2019). A number of alternative platforms have already been created. This includes Gab, a right-wing version of Twitter, and Voat, which is seen as a replacement for Reddit (Marantz, 2019). A significant part of these shifts has been an articulated distrust in large digital media platforms who have been seen as biased and discriminatory to right-wing voices. While the response from r/TheRedPill and r/Braincels differed, this distrust was a strong theme throughout. Users in particular articulated concerns that the quarantine specifically targeted men and was misandrist, or that it was targeting some subreddits unfairly when others went unscathed. The quarantine, therefore either likely created or furthered distrust from users vis-à-vis Reddit as a platform.
In turn, while the quarantine likely reduced hateful material on Reddit, it has potentially created a situation in which members of the manosphere create, at least in regards to their engagement around manosphere ideas, a mini filter bubble of content and social connections that connect only with their own pre-defined beliefs and opinions (Pariser, 2012). While recent empirical research has called into question the widespread existence of filter bubbles online (Zuiderveen Borgesius, Trilling, Möller, Bodó, de Vreese and Helberger, 2016;Bruns, 2019), some research has shown they may exist at the extremes of the ideological spectrum (Rietzschel, 2017). In other words users seeking out manosphere material may increasingly only engage with that content on self-moderated forums, while using other platforms for other social and political connections. If this has occurred this presents new challenges, in particular as newer forums have much less content moderation than large digital platforms. In these spaces therefore, this material is able to flourish unchecked from the oversight either of platform moderators or other users. More research is required to examine the impact recent shifts to self-moderated forums from manosphere and far-right communities have on the creation and spread of hateful material on the web.
Overall therefore, the results of the quarantine as a policy prescription are mixed. For Reddit itself the quarantine would likely be seen as a success. This has been shown by its increased use since the revamp in 2018, with administrators in particular targeting high-profile subreddits such as r/The_Donald and r/ChapoTraphouse (before both were banned in 2020). However, this does not necessarily mean that it changed the perspectives of users affected by the policy or that it reduced levels of hateful material on the web overall. This paper provides a first step of analysis of this policy approach. More work is needed to further this understanding of the role of approaches such as quarantines as ways to reduce hateful content on the internet.
considered less than men, or some men being better than others by virtue of having traditional masculine qualities. • Flipping the Narrative encapsulates terms and expressions that refer to men being oppressed by women or (indirectly or directly) by other men. • Sexual Violence encapsulates any word explicitly connected with sexual violence (and nothing else). • Physical Violence encapsulates any work explicitly connected with physical violence that is not explicitly sexual. • Hostility includes violent verbs, and slurs that are not immediately racist or homophobic. However, if a verb is ambiguous (such as fucking), but it is made into a slur (such as fucker) it is coded as hostility. • Belittling encapsulates any word that is disrespectful or degrading of women's experiences. • Homophobia encapsulates any word related to being homosexual or that mocks being homosexual. This category does not distinguish between terms that have to do explicitly with women and those that have to do with men, for reasons that this was a modifying category to assess general violent attitudes. • Racism referred to any word that was supposed to represent a specific group of people based on where they are from or their perceived race or ethnicity. Xenophobia and racism are not separated in this category. We included terms like 'fresh off the boat' , 'paddy' (pejorative for Irish person) and 'kraut' (pejorative for German person), as well as terms referring specifically to race. The reason for this division is, similarly to the above, to simplify modifying categories about violent attitudes.