Introduction
In the vast ecosystem of online communities, language is more than just a tool for communication; it is a badge of identity. How we speak signals where we belong. Gamers use specific slang, academics have their jargon, and political groups have their distinct rhetoric. But what happens when members of a Hate Speech Group (HSG) leave their isolated echo chambers and enter mainstream discussions? Do they adapt their language to fit in, or do they carry their toxic patterns with them?
This is the core question behind the research paper “Style-Shifting Behaviour of the Manosphere on Reddit” by Jai Aggarwal and Suzanne Stevenson from the University of Toronto.
While much research has focused on detecting toxicity or radicalization, this study takes a nuanced sociolinguistic approach. It examines the “Manosphere”—a loose collection of misogynistic communities on Reddit—and investigates whether its members change their linguistic style when posting in neutral spaces like r/AskReddit, r/Funny, or r/News.
The findings are fascinating and slightly unsettling. They suggest that while hate group members do attempt to blend in, a distinct “Manospheric style” bleeds through, potentially altering the tone and health of mainstream communities. In this deep dive, we will unpack how the researchers quantified this “style,” how they tracked it across the platform, and what this means for the health of online social spaces.
Background: Style-Shifting and the Manosphere
To understand the paper’s contribution, we first need to define two key concepts: Style-Shifting and the Manosphere.
What is Style-Shifting?
In sociolinguistics, style-shifting refers to the way individuals adapt their language depending on the social context. You likely do this every day. The way you type a message to your best friend is probably different from how you write an email to your professor or boss. You shift your “style” to align with the norms of the environment you are in.
In online spaces, this happens constantly. A user might use formal, polite language in r/Science but switch to aggressive, slang-heavy language in a gaming subreddit. The researchers wanted to know: Does this rule apply to members of hate groups?
The Manosphere
The “Manosphere” on Reddit is a collection of subreddits united by a shared misogynistic worldview. This includes groups focused on “Pick Up Artistry,” “Men’s Rights Activism” (in its more radicalized forms), and “Incel” (Involuntary Celibate) ideologies. Previous research has characterized these groups by their high levels of toxicity and radicalization. However, this paper argues that their language isn’t just “toxic”—it has a specific style involving unique vocabulary, grammatical structures, and topical focuses.
The Methodology: Quantifying “Style”
The researchers faced a difficult challenge: How do you mathematically define a “linguistic style”? To answer their research questions, they needed to build a robust pipeline that could distinguish Manospheric speech from “baseline” Reddit speech.
1. Data Collection
The team used the Pushshift Reddit dataset to collect data from 2014 to 2017. They focused on two groups of users:
- Manospheric Authors: Users with at least 10 posts in identified Manosphere subreddits.
- Baseline Authors: Users who had never posted in the Manosphere.
Crucially, to ensure a fair comparison, they matched these authors based on activity levels. If a Manospheric author posted 50 times in r/News, they were compared against a Baseline author who also posted roughly 50 times in r/News.
2. Feature Engineering
The heart of this paper lies in its linguistic features. The researchers didn’t just look for “bad words.” They broke language down into three distinct categories:
A. Uncivil Language Features Using Google’s Perspective API and other sentiment analysis tools, they measured:
- Toxicity: How hateful or aggressive the text is.
- Negativity: The emotional valence of the text.
- Impoliteness: The lack of courteous markers in speech.
B. Syntactic Features (Grammar and Structure) Style is often hidden in how we say things, not just what we say. Using a text analysis tool called LIWC (Linguistic Inquiry and Word Count), the researchers extracted syntactic features. As shown in the table below, this includes function words like pronouns (I, you, she, he), punctuation, and sentence length. These features are powerful because they are often unconscious; a writer might hide their topic, but they rarely hide their grammar.

C. Semantic Features (Topics) Finally, they looked at what was being discussed. They used LIWC to categorize words into semantic themes like “Health,” “Money,” “Religion,” and most importantly for this study, “Female” and “Male” referents.

3. The Classifiers
With these features in hand, the researchers trained Logistic Regression models.
- Platform-Level Classifier: Trained to tell the difference between a post written inside the Manosphere and a post written outside of it.
- Subreddit-Specific Classifiers: Trained to tell the difference between the Manosphere and a specific target subreddit (e.g., Manosphere vs. r/Funny).
These models allowed the researchers to assign a “Manosphericness” score to any given post—a probability score from 0 to 1 indicating how much the post resembles the speech patterns of the Manosphere.

Table 1 above illustrates this scoring system perfectly. Post \(P_1\) is blatant hate speech and scores a 0.98. However, look at \(P_2\). It was posted in r/AskReddit, a neutral space. It scores 0.60, indicating that while it isn’t as extreme as \(P_1\), it still retains a “scent” of the Manosphere compared to the baseline response (\(P_3\), score 0.40).
Research Question 1: What Characterizes Manospheric Style?
Before measuring if the style shifts, the researchers had to define what the style is. They trained a model using all linguistic features to classify posts.

As Table 2 shows, using All 78 features yielded the highest accuracy (0.69). Interestingly, using only uncivil language (toxicity) was the worst predictor (0.56). This proves a critical point: Hate speech groups are defined by more than just hate. Their style is a complex mix of grammar, topic, and tone.
The Gender Obsession
To visualize exactly which features matter, the researchers analyzed feature importance.

The graph above is striking. The single most important predictor of Manospheric style is the use of Female semantic terms (words related to women). The blue bar (Manosphere) is nearly at max importance, while the Baseline (red bar) is significantly lower.
Other key features included:
- Male referents.
- “You” pronouns: Suggesting a confrontational, accusatory style (e.g., “You are the problem”).
- Toxicity: Important, but less so than the gendered topics.
The researchers concluded that Manospheric style is characterized by a heavy focus on gender (specifically women), confrontational framing, and toxicity.
Research Question 2: Do They Shift Their Style?
Now for the core behavioral question: When a Manospheric user posts in r/WorldNews or r/Funny, do they drop this style?
To test this, the authors compared three distributions of “Manosphericness” scores:
- Blue Line: Manospheric authors posting inside the Manosphere. (High Manosphericness expected).
- Red Line: Baseline authors posting in a specific subreddit (e.g., r/Funny). (Low Manosphericness expected).
- Green Line: Manospheric authors posting outside the Manosphere in that specific subreddit.
If Manospheric authors perfectly adapted (code-switched), the Green line would overlap with the Red line. They would sound just like everyone else.
If they didn’t adapt at all, the Green line would overlap with the Blue line. They would sound just like they do in their hate group.
The Result: They are in the middle.

Look closely at Figure 1.
- The Shift (Green vs. Blue): The Green line is significantly to the left of the Blue line. This proves that Manospheric authors do style-shift. They tone it down when they enter public spaces.
- The Spillover (Green vs. Red): However, the Green line is still to the right of the Red line. Even when they try to blend in, they consistently score higher on “Manosphericness” than the average user.
This pattern held true across almost every subreddit tested, as shown in the comprehensive figure below. Whether in r/Gaming, r/Politics, or r/AdviceAnimals, Manospheric authors are consistently distinct from the baseline population.

This gap between the Green and Red lines represents Stylistic Spillover. The hate group’s style is leaking into the broader internet.
Research Question 3: What Elements Spill Over?
We know something is carrying over, but what is it? Is it just that they use more swear words, or is it deeper?
The researchers performed a regression analysis to see which specific features Manospheric authors use more than Baseline authors in neutral spaces. Importantly, they controlled for the “Parent Post” (the post the user is replying to). This ensures that we aren’t just seeing Manospheric users replying to toxic threads; this measures their added contribution to the conversation.

Table 3 reveals the specific ingredients of the spillover:
- Toxicity, Impoliteness, Negativity: Manospheric authors are consistently more uncivil than baseline authors, even in neutral subs.
- Female Focus: They use significantly more words related to women.
- Male Focus: Interestingly, the coefficient is negative (-0.016). Outside the Manosphere, they talk less about men than the baseline, but they maintain their high focus on women.
Seeing it in Action
Numbers are useful, but examples make the findings real. Table 4 provides concrete examples of this spillover effect.

In the first example (Reply 1), a baseline user complains about “cologne” in r/AskMen. The Manospheric user, replying to the exact same prompt, pivots the conversation immediately to “putting women on a pedestal.”
In the second example (Reply 2) in r/Videos, a baseline user calls a woman a “chick” who needs a “reality check.” The Manospheric user escalates this drastically to violent imagery (“hit in the head…”).
These examples perfectly illustrate the statistical findings: the spillover manifests as an increased intensity of violence/toxicity and a relentless redirection of topics toward misogynistic grievances.
Conclusion and Implications
This research provides compelling evidence that the influence of online hate speech groups extends far beyond their own borders. The Manosphere has developed a distinct linguistic fingerprint—characterized by gendered obsession and confrontation—that its members carry with them like an accent.
While these users do engage in style-shifting (toning down their rhetoric when entering mainstream communities), they do not fully assimilate. Instead, they introduce higher levels of toxicity and misogynistic topical focus into communities like r/AskReddit, r/Movies, and r/News.
Why This Matters
- Community Health: This “spillover” acts as a low-level pollutant. Even if a post isn’t toxic enough to be banned by a moderator, it shifts the conversational norms of the community toward hostility and gender bias.
- Detection: Traditional moderation often looks for keywords or extreme slurs. This study shows that “style” is a more complex composite of grammar and topic. Detecting subtle influence requires more sophisticated models that look at how things are said, not just specific bad words.
- Radicalization: By bringing these topics into mainstream spaces, Manospheric actors may normalize their worldview, potentially serving as a gateway for recruiting new members from general-interest communities.
Aggarwal and Stevenson’s work reminds us that online communities are not islands. They are permeable, and the language developed in the darkest corners of the web eventually finds its way into the light. Understanding these linguistic currents is the first step toward mitigating their harmful effects.
](https://deep-paper.org/en/paper/file-3678/images/cover.png)