Skip to main content
Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
CSCW Conf Comput Support Coop Work. Author manuscript; available in PMC 2018 Jan 31.
Published in final edited form as:
CSCW Conf Comput Support Coop Work. 2017 Feb-Mar; 2017: 1217–1230.
doi: 10.1145/2998181.2998213
PMCID: PMC5791909
NIHMSID: NIHMS933916
PMID: 29399664

Anyone Can Become a Troll: Causes of Trolling Behavior in Online Discussions

Abstract

In online communities, antisocial behavior such as trolling disrupts constructive discussion. While prior work suggests that trolling behavior is confined to a vocal and antisocial minority, we demonstrate that ordinary people can engage in such behavior as well. We propose two primary trigger mechanisms: the individual’s mood, and the surrounding context of a discussion (e.g., exposure to prior trolling behavior). Through an experiment simulating an online discussion, we find that both negative mood and seeing troll posts by others significantly increases the probability of a user trolling, and together double this probability. To support and extend these results, we study how these same mechanisms play out in the wild via a data-driven, longitudinal analysis of a large online news discussion community. This analysis reveals temporal mood effects, and explores long range patterns of repeated exposure to trolling. A predictive model of trolling behavior shows that mood and discussion context together can explain trolling behavior better than an individual’s history of trolling. These results combine to suggest that ordinary people can, under the right circumstances, behave like trolls.

ACM Classification Keywords: H.2.8 Database Management: Database Applications–Data Mining, J.4 Computer Applications: Social and Behavioral Sciences
Author Keywords: Trolling, antisocial behavior, online communities

INTRODUCTION

As online discussions become increasingly part of our daily interactions [24], antisocial behavior such as trolling [37, 43], harassment, and bullying [82] is a growing concern. Not only does antisocial behavior result in significant emotional distress [1, 58, 70], but it can also lead to offline harassment and threats of violence [90]. Further, such behavior comprises a substantial fraction of user activity on many web sites [18, 24, 30] – 40% of internet users were victims of online harassment [27]; on CNN.com, over one in five comments are removed by moderators for violating community guidelines. What causes this prevalence of antisocial behavior online?

In this paper, we focus on the causes of trolling behavior in discussion communities, defined in the literature as behavior that falls outside acceptable bounds defined by those communities [9, 22, 37]. Prior work argues that trolls are born and not made: those engaging in trolling behavior have unique personality traits [11] and motivations [4, 38, 80]. However, other research suggests that people can be influenced by their environment to act aggressively [20, 41]. As such, is trolling caused by particularly antisocial individuals or by ordinary people? Is trolling behavior innate, or is it situational? Likewise, what are the conditions that affect a person’s likelihood of engaging in such behavior? And if people can be influenced to troll, can trolling spread from person to person in a community? By understanding what causes trolling and how it spreads in communities, we can design more robust social systems that can guard against such undesirable behavior.

This paper reports a field experiment and observational analysis of trolling behavior in a popular news discussion community. The former allows us to tease apart the causal mechanisms that affect a user’s likelihood of engaging in such behavior. The latter lets us replicate and explore finer grained aspects of these mechanisms as they occur in the wild. Specifically, we focus on two possible causes of trolling behavior: a user’s mood, and the surrounding discussion context (e.g., seeing others’ troll posts before posting).

Online experiment

We studied the effects of participants’ prior mood and the context of a discussion on their likelihood to leave troll-like comments. Negative mood increased the probability of a user subsequently trolling in an online news comment section, as did the presence of prior troll posts written by other users. These factors combined to double participants’ baseline rates of engaging in trolling behavior.

Large-scale data analysis

We augment these results with an analysis of over 16 million posts on CNN.com, a large online news site where users can discuss published news articles. One out of four posts flagged for abuse are authored by users with no prior record of such posts, suggesting that many undesirable posts can be attributed to ordinary users. Supporting our experimental findings, we show that a user’s propensity to troll rises and falls in parallel with known population-level mood shifts throughout the day [32], and exhibits cross-discussion persistence and temporal decay patterns, suggesting that negative mood from bad events linger [41, 45]. Our data analysis also recovers the effect of exposure to prior troll posts in the discussion, and further reveals how the strength of this effect depends on the volume and ordering of these posts.

Drawing on this evidence, we develop a logistic regression model that accurately (AUC=0.78) predicts whether an individual will troll in a given post. This model also lets us evaluate the relative importance of mood and discussion context, and contrast it with prior literature’s assumption of trolling being innate. The model reinforces our experimental findings – rather than trolling behavior being mostly intrinsic, such behavior can be mainly explained by the discussion’s context (i.e., if prior posts in the discussion were flagged), as well as the user’s mood as revealed through their recent posting history (i.e., if their last posts in other discussions were flagged).

Thus, not only can negative mood and the surrounding discussion context prompt ordinary users to engage in trolling behavior, but such behavior can also spread from person to person in discussions and persist across them to spread further in the community. Our findings suggest that trolling, like laughter, can be contagious, and that ordinary people, given the right conditions, can act like trolls. In summary, we:

  • present an experiment that shows that both negative mood and discussion context increases the likelihood of trolling,

  • validate these findings with a large-scale analysis of a large online discussion community, and

  • use these insights to develop a predictive model that suggests that trolling may be more situational than innate.

BACKGROUND

To begin, we review literature on antisocial behavior (e.g., aggression and trolling) and influence (e.g., contagion and cascading behavior), and identify open questions about how trolling spreads in a community.

Antisocial behavior in online discussions

Antisocial behavior online can be seen as an extension of similar behavior offline, and includes acts of aggression, harassment, and bullying [1, 43]. Online antisocial behavior increases anger and sadness [58], and threatens social and emotional development in adolescents [70]. In fact, the pain of verbal or social aggression may also linger longer than that of physical aggression [16].

Antisocial behavior can be commonly observed in online public discussions, whether on news websites or on social media. Methods of combating such behavior include comment ranking [39], moderation [53, 67], early troll identification [14, 18], and interface redesigns that encourage civility [51, 52]. Several sites have even resorted to completely disabling comments [28]. Nonetheless, on the majority of popular web sites which continue to allow discussions, antisocial behavior continues to be prevalent [18, 24, 30]. In particular, a rich vein of work has focused on understanding trolling on these discussion platforms [26, 37], for example discussing the possible causes of malicious comments [55].

A troll has been defined in multiple ways in previous literature – as a person who initially pretends to be a legitimate participant but later attempts to disrupt the community [26], as someone who “intentionally disrupts online communities” [77], or “takes pleasure in upsetting others” [47], or more broadly as a person engaging in “negatively marked online behavior” [37] or that “makes trouble” for a discussion forums’ stakeholders [9]. In this paper, similar to the latter studies, we adopt a definition of trolling that includes flaming, griefing, swearing, or personal attacks, including behavior outside the acceptable bounds defined by several community guidelines for discussion forums [22, 25, 35].1 In our experiment, we code posts manually for trolling behavior. In our longitudinal data analysis, we use posts that were flagged for unacceptable behavior as a proxy for trolling behavior.

Who engages in trolling behavior? One popular recurring narrative in the media suggests that trolling behavior comes from trolls: a small number of particularly sociopathic individuals [71, 77]. Several studies on trolling have focused on a small number of individuals [4, 9, 38, 80]; other work shows that there may be predisposing personality (e.g., sadism [11]) and biological traits (e.g., low baseline arousal [69]) to aggression and trolling. That is, trolls are born, not made.

Even so, the prevalence of antisocial behavior online suggests that these trolls, being relatively uncommon, are not responsible for all instances of trolling. Could ordinary individuals also engage in trolling behavior, even if temporarily? People are less inhibited in their online interactions [84]. The relative anonymity afforded by many platforms also deindividualizes and reduces accountability [95], decreasing comment quality [46]. This disinhibition effect suggests that people, in online settings, can be more easily influenced to act antisocially. Thus, rather than assume that only trolls engage in trolling behavior, we ask: RQ: Can situational factors trigger trolling behavior?

Causes of antisocial behavior

Previous work has suggested several motivations for engaging in antisocial behavior: out of boredom [86], for fun [80], or to vent [55]. Still, this work has been largely qualitative and non-causal, and whether these motivations apply to the general population remains largely unknown. Out of this broad literature, we identify two possible trigger mechanisms of trolling – mood and discussion context – and try to establish their effects using both a controlled experiment and a large-scale longitudinal analysis.

Mood

Bad moods may play a role in how a person later acts. Negative mood correlates with reduced satisfaction with life [79], impairs self-regulation [56], and leads to less favorable impressions of others [29]. Similarly, exposure to unrelated aversive events (e.g., higher temperatures [74] or secondhand smoke [41]) increases aggression towards others. An interview study found that people thought that malicious comments by others resulted from “anger and feelings of inferiority” [55].

Nonetheless, negative moods elicit greater attention to detail and higher logical consistency [78], which suggests that people in a bad mood may provide more thoughtful commentary. Prior work is also mixed on how affect influences prejudice and stereotyping. Both positive [10] and negative affect [34] can increase stereotyping, and thus trigger trolling [38]. Still, we expect the negative effects of negative mood in social contexts to outweigh these other factors.

Circumstances that influence mood may also modify the rate of trolling. For instance, mood changes with the time of day or day of week [32]. As negative mood rises at the start of the week, and late at night, trolling may vary similarly. “Time-outs” or allowing for a period of calming down [45] can also reduce aggression – users who wait longer to post after a bout of trolling may also be less susceptible to future trolling. Thus, we may be able to observe how mood affects trolling, directly through experimentation, and indirectly through observing factors that influence mood:

  • H1: Negative mood increases a user’s likelihood of trolling.

Discussion context

A discussion’s context may also affect what people contribute. The discussion starter influences the direction of the rest of the discussion [36]. Qualitative analyses suggest that people think online commenters follow suit in posting positive (or negative) comments [55]. More generally, standards of behavior (i.e., social norms) are inferred from the immediate environment [15, 20, 63]. Closer to our work is an experiment that demonstrated that less thoughtful posts led to less thoughtful responses [83]. We extend this work by studying amplified states of antisocial behavior (i.e., trolling) in both experimental and observational settings.

On the other hand, users may not necessarily react to trolling with more trolling. An experiment that manipulated the initial votes an article received found that initial downvotes tended to be corrected by the community [64]. Some users respond to trolling with sympathy or understanding [4], or apologies or joking [54]. Still, such responses are rarer [4].

Another aspect of a discussion’s context is the subject of discussion. In the case of discussions on news sites, the topic of an article can affect the amount of abusive comments posted [30]. Overall, we expect that previous troll posts, regardless of who wrote them, are likely to result in more subsequent trolling, and that the topic of discussion also plays a role:

  • H2: The discussion context (e.g., prior troll posts by other users) affects a user’s likelihood of trolling.

Influence and antisocial behavior

That people can be influenced by environmental factors suggests that trolling could be contagious – a single user’s out-burst might lead to multiple users participating in a flame war. Prior work on social influence [5] has demonstrated multiple examples of herding behavior, or that people are likely to take similar actions to previous others [21, 62, 95]. Similarly, emotions and behavior can be transferred from person to person [6, 13, 31, 50, 89]. More relevant is work showing that getting downvoted leads people to downvote others more and post content that gets further downvoted in the future [17].

These studies generally point toward a “Broken Windows” hypothesis, which postulates that untended behavior can lead to the breakdown of a community [92]. As an unfixed broken window may create a perception of unruliness, comments made in poor taste may invite worse comments. If antisocial behavior becomes the norm, this can lead a community to further perpetuate it despite its undesirability [91].

Further evidence for the impact of antisocial behavior stems from research on negativity bias – that negative traits or events tend to dominate positive ones. Negative entities are more contagious than positive ones [75], and bad impressions are quicker to form and more resistant to disconfirmation [7]. Thus, we expect antisocial behavior is particularly likely to be influential, and likely to persist. Altogether, we hypothesize:

  • H3: Trolling behavior can spread from user to user.

We test H1 and H2 using a controlled experiment, then verify and extend our results with an analysis of discussions on CNN.com. We test H3 by studying the evolution of discussions on CNN.com, finally developing an overall model for how trolling might spread from person to person.

EXPERIMENT: MOOD AND DISCUSSION CONTEXT

To establish the effects of mood and discussion context, we deployed an experiment designed to replicate a typical online discussion of a news article.

Specifically, we measured the effect of mood and discussion context on the quality of the resulting discussion across two factors: a) PosMood or NegMood: participants were either exposed to an unrelated positive or negative prior stimulus (which in turn affected their prevailing mood), and b) PosContext or NegContext: the initial posts in the discussion thread were either benign (or not troll-like), or troll-like. Thus, this was a two-by-two between-subjects design, with participants assigned in a round robin to each of the four conditions.

We evaluated discussion quality using two measures: a) trolling behavior, or whether participants wrote more troll–like posts, and b) affect, or how positive or negative the resulting discussion was, as measured using sentiment analysis. If negative mood (NegMood) or troll posts (NegContext) affects the probability of trolling, we would expect these conditions to reduce discussion quality.

Experimental Setup

The experiment consisted of two main parts – a quiz, followed by a discussion – and was conducted on Amazon Mechanical Turk (AMT). Past work has also recruited workers to participate in experiments with online discussions [60]. Participants were restricted to residing in the US, only allowed to complete the experiment once, and compensated $2.00, for an hourly rate of $8.00. To avoid demand characteristics, participants were not told of the experiment’s purpose prior, and were only instructed to complete a quiz, and then participate in an online discussion. After the experiment, participants were debriefed and told of its purpose (i.e., to measure the impact of mood and trolling in discussions). The experimental protocol was reviewed and conducted under IRB Protocol #32738.

Quiz (PosMood or NegMood)

The goal of the quiz was to see if participants’ mood prior to participating in a discussion had an effect on subsequent trolling. Research on mood commonly involves giving people negative feedback on tasks that they perform in laboratory experiments regardless of their actual performance [48, 93, 33]. Adapting this to the context of AMT, where workers care about their performance on tasks and qualifications (which are necessary to perform many higher-paying tasks), participants were instructed to complete an experimental test qualification that was being considered for future use on AMT. They were told that their performance on the quiz would have no bearing on their payment at the end of the experiment.

The quiz consisted of 15 open-ended questions, and included logic, math, and word problems (e.g., word scrambles) (Figure 1a). In both conditions, participants were given five minutes to complete the quiz, after which all input fields were disabled and participants forced to move on. In both the PosMood and NegMood conditions, the composition and order of the types of questions remained the same. However, the NegMood condition was made up of questions that were substantially harder to answer within the time limit: for example, unscramble “DEANYON” (NegMood) vs. “PAPHY” (PosMood). At the end of the quiz, participants’ answers were automatically scored, and their final score displayed to them. They were told whether they performed better, at, or worse than the “average”, which was fixed at eight correct questions. Thus, participants were expected to perform well in the PosMood condition and receive positive feedback, and expected to perform poorly in the NegMood condition and receive negative feedback, being told that they were performing poorly, both absolutely and relatively to other users. While users in the PosMood condition can still perform poorly, and users in the NegMood condition perform well, this only reduces the differences later observed.

An external file that holds a picture, illustration, etc.
Object name is nihms933916f1.jpg

To understand how a person’s mood and discussion’s context (i.e., prior troll posts) affected the quality of a discussion, we conducted an experiment that varied (a) how difficult a quiz, given prior to participation in the discussion, was, as well as (b) whether the initial posts in a discussion were troll posts or not.

To measure participants’ mood following the quiz, and acting as a manipulation check, participants then completed 65 Likert-scale questions on how they were feeling based on the Profile of Mood States (POMS) questionnaire [61], which quantifies mood on six axes such as anger and fatigue.

Discussion (PosContext or NegContext)

Participants were then instructed to take part in an online discussion, and told that we were testing a comment ranking algorithm. Here, we showed participants an interface similar to what they might see on a news site — a short article, followed by a comments section. Users could leave comments, reply to others’ comments, or upvote and downvote comments (Figure 1b). Participants were required to leave at least one comment, and told that their comments may be seen by other participants. Each participant was randomly assigned a username (e.g., User1234) when they commented. In this experiment, we showed participants an abridged version of an article arguing that women should vote for Hillary Clinton instead of Bernie Sanders in the Democratic primaries leading up to the 2016 US presidential election [42]. In the NegContext condition, the first three comments were troll posts, e.g.,:

Oh yes. By all means, vote for a Wall Street sellout – a lying, abuse-enabling, soon-to-be felon as our next President. And do it for your daughter. You’re quite the role model.

In the PosContext, they were more innocuous:

I’m a woman, and I don’t think you should vote for a woman just because she is a woman. Vote for her because you believe she deserves it.

These comments were abridged from real comments posted by users in comments in the original article, as well as other online discussion forums discussing the issue (e.g., Reddit).

To ensure that the effects we observed were not path-dependent (i.e., if a discussion breaks down by chance because of a single user), we created eight separate “universes” for each condition [76], for a total of 32 universes. Each universe was seeded with the same comments, but were otherwise entirely independent. Participants were randomized between universes within each condition. Participants assigned to the same universe could see and respond to other participants who had commented prior, but not interact with participants from other universes.

Measuring discussion quality

We evaluated discussion quality in two ways: if subsequent posts written exhibited trolling behavior, or if they contained more negative affect. To evaluate whether a post was a troll post or not, two experts (including one of the authors) independently labeled posts as being troll or non-troll posts, blind to the experimental conditions, with disagreements resolved through discussion. Both experts reviewed CNN.com’s community guidelines [22] for commenting – posts that were offensive, irrelevant, or designed to elicit an angry response, whether intentional or not, were labeled as trolling. To measure the negative affect of a post, we used LIWC [68] (Vader [40] gives similar results).

Results

667 participants (40% female, mean age 34.2, 54% Democrat, 25% Moderate, 21% Republican) completed the experiment, with an average of 21 participants in each universe. In aggregate, these workers contributed 791 posts (with an average of 37.8 words written per post) and 1392 votes.

Manipulation checks

First we sought to verify that the quiz did affect participants’ mood. On average, participants in the PosMood condition obtained 11.2 out of 15 questions correct, performing above the stated “average” score of 8. In contrast, participants in the NegMood condition answered only an average of 1.9 questions correctly, performing significantly worse (t(594)=63.2, p<0.001 using an unequal variances t-test), and below the stated “average”. Correspondingly, the post-quiz POMS questionnaire confirmed that participants in the NegMood condition experienced higher mood disturbance on all axes, with higher anger, confusion, depression, fatigue, and tension scores, and a lower vigor score (t(534)>7.0, p<0.001). Total mood disturbance, where higher scores correspond to more negative mood, was 12.2 for participants in the PosMood condition (comparable to a baseline level of disturbance measured among athletes [85]), and 40.8 in the NegMood condition. Thus, the quiz put participants into a more negative mood.

Verifying that the initial posts in the NegContext condition were perceived as being more troll-like than those in the PosContext condition, we found that the initial posts in the NegContext condition were less likely to be upvoted (36% vs. 90% upvoted for PosContext, t(507)=15.7, p<0.001).

Negative mood and negative context increase trolling behavior

Table 1 shows how the proportion of troll posts and negative affect (measured as the proportion of negative words) differ in each condition. The proportion of troll posts was highest in the (NegMood, NegContext) condition with 68% troll posts, drops in both the (NegMood, PosContext) and (PosMood, NegContext) conditions with 47% and 49% each, and is lowest in the (PosMood, PosContext) condition with 35%. For negative affect, we observe similar differences.

Table 1

The proportion of user-written posts that were labeled as trolling (and proportion of words with negative affect) was lowest in the (PosMood, PosContext) condition, and highest, and almost double, in the (NegMood, NegContext) condition (highlighted in bold).

Proportion of Troll PostsNegative Affect (LIWC)
PosMoodNegMoodPosMoodNegMood
PosContext35%49%1.1%1.4%
NegContext47%68%2.3%2.9%

Fitting a mixed effects logistic regression model, with the two conditions as fixed effects, an interaction between the two conditions, user as a random effect, and whether a contributed post was trolling or not as the outcome variable, we do observe a significant effect of both NegMood and NegContext (p<0.05) (Table 2). These results confirm both H1 and H2, that negative mood and the discussion context (i.e., prior troll posts) increase a user’s likelihood of trolling. Negative mood increases the odds of trolling by 89%, and the presence of prior troll posts increases the odds by 68%. A mixed model using MCMC revealed similar effects (p<0.05), and controlling for universe, gender, age, or political affiliation also gave similar results. Further, the effect of a post’s position in the discussion on trolling was not significant, suggesting that trolling tends to persist in the discussion.

Table 2

A mixed effects logistic regression reveals a significant effect of both NegMood and NegContext on troll posts (*: p<0.05, **: p<0.01, ***: p<0.001). In other words, both negative mood and the presence of initial troll posts increases the probability of trolling.

Fixed EffectsCoef.SEz
(Intercept)−0:70***0:17−4:23
NegMood0:64**0:242:66
NegContext0:52*0:232:38
NegMood × NegContext0:410:331:23

Random EffectsVar.SE
User0:410:64

With the proportion of words with negative affect as the outcome variable, we observed a significant effect of NegContext (p<0.05), but not of NegMood – such measures may not accurately capture types of trolling such as sarcasm or off-topic posting. There was no significant effect of either factor on positive affect.

Examples of troll posts

Contributed troll posts comprised a relatively wide range of antisocial behavior: from outright swearing (“What a dumb c***”) and personal attacks (“You’re and idiot and one of the things that’s wrong with this country.”) to veiled insults (“Hillary isn’t half the man Bernie is!!! lol”), sarcasm (“You sound very white, and very male. Must be nice.”), and off-topic statements (“I think Ted Cruz has a very good chance of becoming president.”). In contrast, non-troll posts tended to be more measured, regardless of whether they agreed with the article (“Honestly I agree too. I think too many people vote for someone who they identify with rather than someone who would be most qualified.”).

Other results

We observed trends in the data. Both conditions reduced the number of words written relative to the control condition: 44 words written in the (PosMood, PosContext) vs. 29 words written in the (NegMood, NegContext) condition. Also, the percentage of upvotes on posts written by other users (i.e., excluding the initial seed posts) was lower: 79% in the (PosMood, PosContext) condition vs. 75% in the (NegMood, NegContext) condition. While suggestive, neither effect was significant.

Discussion

Why did NegContext and NegMood increase the rate of trolling? Drawing on prior research explaining the mechanism of contagion [89], participants may have an initial negative reaction to reading the article, but are unlikely to bluntly externalize them because of self-control or environmental cues. NegContext provides evidence that others had similar reactions, making it more acceptable to also express them. NegMood further accentuates any perceived negativity from reading the article and reduces self-inhibition [56], making participants more likely to act out.

Limitations

In this experiment, like prior work [60, 83], we recruited participants to participate in an online discussion, and required each to post at least one comment. While this enables us isolate both mood and discussion context (which is difficult to control for in a live Reddit discussion for example) and further allows us to debrief participants afterwards, payment may alter the incentives to participate in the discussion. Users also were commenting pseudonymously via randomly generated usernames, which may reduce overall comment quality [46]. Different initial posts may also elicit different subsequent posts. While our analyses did not reveal significant effects of demographic factors, future work could further examine their impact on trolling. For example, men may be more susceptible to trolling as they tend to be more aggressive [8]. Anecdotally, several users who identified as Republican trolled the discussion with irrelevant mentions of Donald Trump (e.g., “I’m a White man and I’m definitely voting for Donald Trump!!!”). Understanding the effects of different types of trolling (e.g., swearing vs. sarcasm) and user motivations for such trolling (e.g., just to rile others up) also remains future work. Last, different articles may be trolled to different extents [30], so we examine the effect of article topic in our subsequent analyses.

Overall, we find that both mood and discussion context significantly affect a user’s likelihood of engaging in trolling behavior. For such effects to be observable, a substantial proportion of the population must have been susceptible to trolling, rather than only a small fraction of atypical users – suggesting that trolling can be generally induced. But do these results generalize to real-world online discussions? In the subsequent sections, we verify and extend our results with an analysis of CNN.com, a large online news discussion community. After describing this dataset, we study how trolling behavior tracks known daily mood patterns, and how mood persists across multiple discussions. We again find that the initial posts of discussions have a significant effect on subsequent posts, and study the impact of the volume and ordering of multiple troll posts on subsequent trolling. Extending our analysis of discussion context to include the accompanying article’s topic, we find that it too mediates trolling behavior.

DATA: INTRODUCTION

CNN.com is a popular American news website where editors and journalists write articles on a variety of topics (e.g., politics and technology), which users can then discuss. In addition to writing and replying to posts, users can up- and down-vote, as well as flag posts (typically for abuse or violations of the community guidelines [22]). Moderators can also delete posts or even ban users, in keeping with these guidelines. Disqus, a commenting platform that hosted these discussions on CNN.com, provided us with a complete trace of user activity from December 2012 to August 2013, consisting of 865,248 users (20,197 banned), 16,470 discussions, and 16,500,603 posts, of which 571,662 (3.5%) were flagged and 3,801,774 (23%) were deleted. Out of all flagged posts, 26% were made by users with no prior record of flagging in previous discussions; also, out of all users with flagged posts who authored at least ten posts, 40% had less than 3.5% of their posts flagged (the baseline probability of a random post being flagged on CNN). These observations suggest that ordinary users are responsible for a significant amount of trolling behavior, and that many may have just been having a bad day.

In studying behavior on CNN.com, we consider two main units of analysis: a) a discussion, or all the posts that follow a given news article, and b) a sub-discussion, or a top-level post and any replies to that post. We make this distinction as discussions may reach thousands of posts, making it likely that users may post in a discussion without reading any previous responses. In contrast, a sub-discussion necessarily involves replying to a previous post, and would allow us to better study the effects of people reading and responding to each other.

In our subsequent analyses, we filter banned users (of which many tend to be clearly identifiable trolls [18]), as well as any users who had all of their posts deleted, as we are primarily interested in studying the effects of mood and discussion context on the general population.

We use flagged posts (posts that CNN.com users marked for violating community guidelines) as our primary measure of trolling behavior. In contrast, moderator deletions are typically incomplete: moderators miss some legitimate troll behavior and tend to delete entire discussions as opposed to individual posts. Likewise, written negative affect misses sarcasm and other trolling behaviors that do not involve common negative words, and downvoting may simply indicate disagreement. To validate this approach, two experts (including one of the authors) labeled 500 posts (250 flagged) sampled at random, blind to whether each post was flagged, using the same criteria for trolling as for the experiment. Comparing the expert labels with post flags from the dataset, we obtained a precision of 0.66 and recall of 0.94, suggesting that while some troll posts remain unflagged, almost all flagged posts are troll posts. In other words, while instances of trolling behavior go unnoticed (or are ignored), when a post is flagged, it is highly likely that trolling behavior did occur. So, we use flagged posts as a primary estimate of trolling behavior in our analyses, complementing our analysis with other signals such as negative affect and downvotes. These signals are correlated: flagged posts are more likely than non-flagged posts to have greater negative affect (3.7% vs. 3.4% of words, Cohen’s d=0.06, t=40, p<0.001), be downvoted (58% vs. 30% of votes, d=0.76, t=531, p<0.001), or be deleted by a moderator (79% vs. 21% of posts, d=1.4, t=1050, p<0.001).

DATA: UNDERSTANDING MOOD

In the earlier experiment, we showed that bad mood increases the probability of trolling. In this section, using large-scale and longitudinal observational data, we verify and expand on this result. While we cannot measure mood directly, we can study its known correlates. Seasonality influences mood [32], so we study how trolling behavior also changes with the time of day or day of week. Aggression can linger beyond an initial unpleasant event [41], thus we also study how trolling behavior persists as a user participates in multiple discussions.

Happy in the day, sad at night

Prior work that studied changes in linguistic affect on Twitter demonstrated that mood changes with the time of the day, and with the day of the week – positive affect peaks in the morning, and during weekends [32]. If mood changes with time, could trolling be similarly affected? Are people more likely to troll later in the day, and on weekdays? To evaluate the impact of the time of day or day of week on mood and trolling behavior, we track several measures that may indicate troll-like behavior: a) the proportion of flagged posts (or posts reported by other users as being abusive), b) negative affect, and c) the proportion of downvotes on posts (or the average fraction of downvotes on posts that received at least one vote).

Figures 2a and 2b show how each of these measures changes with the time of day and day of week, respectively, across all posts. Our findings corroborate prior work – the proportion of flagged posts, negative affect, and the proportion of downvotes are all lowest in the morning, and highest in the evening, aligning with when mood is worst [32]. These measures also peak on Monday (the start of the work week in the US).

An external file that holds a picture, illustration, etc.
Object name is nihms933916f2.jpg

Like negative mood, indicators of trolling peak (a) late at night, and (b) early in the work week, supporting a relation between mood and trolling. Further, (c) the shorter the time between a user’s subsequent posts in unrelated discussions, where the first post is flagged, the more likely the second will also be flagged, suggesting that negative mood may persist for some time.

Still, trolls may simply wake up later than normal users, or post on different days. To understand how the time of day and day of week affect the same user, we compare these measures for the same user in two different time periods: from 6 am to 12 pm and from 11 pm to 5 am, and on two different days: Monday and Friday (i.e., early or late in the work week). A paired t-test reveals a small, but significant increase in negative behavior between 11 pm and 5 am (flagged posts: 4.1% vs. 4.3%, d=0.01, t(106300)=2.79, p<0.01; negative affect: 3.3% vs. 3.4%, t(106220)=3.44, d=0.01, p<0.01; downvotes: 20.6% vs. 21.4%, d=0.02, t(26390)=2.46, p<0.05). Posts made on Monday also show more negative behavior than posts made on Friday (d≥0.02, t>2.5, p<0.05). While these effects may also be influenced by the type of news that gets posted at specific times or days, limiting our analysis to just news articles categorized as “US” or “World”, the two largest sections, we continue to observe similar results.

Thus, even without direct user mood measurements, patterns of trolling behavior correspond predictably with mood.

Anger begets more anger

Negative mood can persist beyond the events that brought about those feelings [44]. If trolling is dependent on mood, we may be able to observe the aftermath of user outbursts, where negative mood might spill over from prior discussions into subsequent, unrelated ones, just as our experiment showed that negative mood that resulted from doing poorly on a quiz affected later commenting in a discussion. Further, we may also differentiate the effects that stem from actively engaging in negative behavior in the past, versus simply being exposed to negative behavior. Correspondingly, we ask two questions, and answer them in turn. First, a) if a user wrote a troll post in a prior discussion, how does that affect their probability of trolling in a subsequent, unrelated discussion? At the same time, we might also observe indirect effects of trolling: b) if a user participated in a discussion where trolling occurred, but did not engage in trolling behavior themselves, how does that affect their probability of trolling in a subsequent, unrelated discussion?

To answer the former, for a given discussion, we sample two users at random, where one had a post which was flagged, and where one had a post which was not flagged. We ensure that these two users made at least one post prior to participating in the discussion, and match users on the total number of posts they wrote prior to the discussion. As we are interested in these effects on ordinary users, we also ensure that neither of these users have had any of their posts flagged in the past. We then compare the likelihood of each user’s next post in a new discussion also being flagged. We find that users who had a post flagged in a prior discussion were twice as likely to troll in their next post in a different discussion (4.6% vs. 2.1%, d=0.14, t(4641)=6.8, p<0.001) (Figure 3a). We obtain similar results even when requiring these users to also have no prior deleted posts or longer histories (e.g., if they have written at least five posts prior to the discussion).

An external file that holds a picture, illustration, etc.
Object name is nihms933916f3.jpg

Suggesting that negative mood may persist across discussions, users with no prior history of flagged posts, who either (a) make a post in a prior unrelated discussion that is flagged, or (b) simply participates in a sub-discussion in a prior discussion with at least one flagged post, without themselves being flagged, are more likely to be subsequently flagged in the next discussion they participate in. Demonstrating the effect of discussion context, (c) discussions that begin with a flagged post are more likely to have a greater proportion of flagged posts by other users later on, as do (d) sub-discussions that begin with a flagged post.

Next, we examine the indirect effect of participating in a “bad” discussion, even when the user does not directly engage in trolling behavior. We again sample two users from the same discussion, but where each user participated in a different sub-discussion: one sub-discussion had at least one other post by another user flagged, and the other sub-discussion had no flagged posts. Again, we match users on the number of posts they wrote in the past, and ensure that these users have no prior flagged posts (including in the sampled discussions). We then compare the likelihood of each user’s next post in a new discussion being flagged. Here, we also find that users who participated in a prior discussion with at least one flagged post were significantly more likely to subsequently author a post in an new discussion that would be flagged (Figure 3b). However, this effect is significantly weaker (2.2% vs. 1.7%, d=0.04, t(7321)=2.7, p<0.01).

Thus, both trolling in a past discussion, as well as participating in a discussion where trolling occurred, can affect whether a user trolls in the future discussion. These results suggest that negative mood can persist and transmit trolling norms and behavior across multiple discussions, where there is no similar context to draw on. As none of the users we analyzed had prior flagged posts, this effect is unlikely to arise simply because some users were just trolls in general.

Time heals all wounds

One typical anger management strategy is to use a “time-out” to calm down [45]. Thus, could we minimize negative mood carrying over to new discussions by having users wait longer before making new posts? Assuming that a user is in a negative mood (as indicated by writing a post that is flagged), the time elapsed until the user’s next post may correlate with the likelihood of subsequent trolling. In other words, we might expect that the longer time the time between posts, the greater the temporal distance from the origin of the negative mood, and hence the lower the likelihood of trolling.

Figure 2c shows how the probability of a user’s next post being flagged changes with the time since that user’s last post, assuming that the previous post was flagged. So as not to confuse the effects of the initial post’s discussion context, we ensure that the user’s next post is made in a new discussion with different other users. The probability of being flagged is high when the time between these two subsequent posts is short (five minutes or less), suggesting that a user might still be in a negative mood persisting from the initial post. As more time passes, even just ten minutes, the probability of being flagged gradually decreases. Nonetheless, users with better impulse control may wait longer before posting again if they are angry, and isolating this effect would be future work. Our findings here lend credence to the rate-limiting of posts that some forums have introduced [2].

DATA: UNDERSTANDING DISCUSSION CONTEXT

From our experiment, we identified mood and discussion context as influencing trolling. The previous section verified and extended our results on mood; in this section, we do the same for discussion context. In particular, we show that posts are more likely to be flagged if others’ prior posts were also flagged. Further, the number and ordering of flagged posts in a discussion affects the probability of subsequent trolling, as does the topic of the discussion.

FirST!!1

How strongly do the initial posts to a discussion affect the likelihood of subsequent posts to troll? To measure the effect of the initial posts on subsequent discussions, we first identified discussions of at least 20 posts, separating them into those with their first post flagged and those without their first post flagged. We then used propensity score matching to create matched pairs of discussions where the topic of the article, the day of week the article was posted, and the total number of posts are controlled for [72]. Thus, we end up with pairs of discussions on the same topic, started on the same day of the week, and with similar popularity, but where one discussion had its first post flagged, while the other did not. We then compare the probability of the subsequent posts in the discussion being flagged. As we were interested in the impact of the initial post on other ordinary users, we excluded any posts written by the user who made the initial post, posts by users who replied (directly or indirectly) to that post, and posts by users with prior flagged or deleted posts in previous discussions.

After an initial flagged post, we find that subsequent posts by other users were more likely to be flagged, than if the initial post was not flagged (3.1% vs. 1.7%, d=0.32, t(1545)=9.1, p<0.001) (Figure 3c). This difference remains significant even when only considering posts made in the second half of a discussion (2.1% vs. 1.3%, d=0.19, t(1545)=5.4, p<0.001). Comparing discussions where the first three posts were all flagged to those where none of these posts were flagged (similar to NegContext vs. PosContext in our experiment), the gap widens (7.1% vs. 1.7%, d=0.61, t(113)=4.6, p<0.001).

Nonetheless, as these different discussions were on different articles, some articles, even within the same topic, may have been more inflammatory, increasing the overall rate of flagging. To control for the article being discussed, we also look at sub-discussions (a top-level post and all of its replies) within the same discussion. Sub-discussions tend to be closer to actual conversations between users as each subsequent post is an explicit reply to another post in the chain, as opposed to considering the discussion as a whole where users can simply leave a comment without reading or responding to anyone else. From each discussion we select two sub-discussions at random, where one sub-discussion’s top-level post was flagged, and where the other’s was not, and only considered posts not written by the users who started these sub-discussions. Again, we find that sub-discussions whose top-level posts were flagged were significantly more likely to result in more flagging later in that sub-discussion (9.6% vs. 5.9%, d=0.16, t(501)=3.9, p<0.001) (Figure 3d).

Altogether, these results suggest that the initial posts in a discussion set a strong, lasting precedent for later trolling.

From bad to worse: sequences of trolling

By analyzing the volume and ordering of troll posts in a discussion, we can better understand how discussion context and trolling behavior interact. Here, we study sub-discussions at least five posts in length, and separately consider posts written by users new to the sub-discussion and posts written by users who have posted before in the sub-discussion to control for the impact of having already participated in the discussion.

Do more troll posts increase the likelihood of future troll posts? Figure 4a shows that as the number of flagged posts among the first four posts increases, the probability that the fifth post is also flagged increases monotonically. With no prior flagged posts, the chance of the fifth post by a new user to the sub-discussion being flagged is just 2%; with one other flagged post, this jumps to 7%; with four flagged posts, the odds of the fifth post also being flagged are almost one to one (49%). These pairwise differences are all significant with a Holm correction (χ2(1)>7.6, p<0.01). We observe similar trends for users new to the sub-discussion, as well as users that had posted previously, with the latter group of users more likely to be subsequently flagged.

An external file that holds a picture, illustration, etc.
Object name is nihms933916f4.jpg

In discussions with at least five posts, (a) the probability that a post is flagged monotonically increases with the number of prior flagged posts in the discussion. (b) If only one of the first four posts was flagged, the fifth post is more likely to be flagged if that flagged post is closer in position. (c) The topic of a discussion also influences the probability of a post being flagged.

Further, does a troll post made later in a discussion, and closer to where a user’s post will show up, have a greater impact than a troll post made earlier on? Here, we look at discussions of at least five posts where there was exactly one flagged post among the first four, and where that flagged post was not written by the fifth post’s author. In Figure 4b, the closer in position the flagged post is to the fifth post, the more likely that post is to be flagged. For both groups of users, the fifth post in a discussion is more likely to be flagged if the fourth post was flagged, as opposed to the first (χ2(1)>6.9, p<0.01).

Beyond the presence of troll posts, their conspicuousness in discussions substantially affects if new discussants troll as well. These findings, together with our previous results showing how simply participating in a previous discussion having a flagged post raises the likelihood of future trolling behavior, support H3: that trolling behavior spreads from user to user.

Hot-button issues push users’ buttons?

How does the subject of a discussion affect the rate of trolling? Controversial topics (e.g., gender, GMOs, race, religion, or war) may divide a community [57], and thus lead to more trolling. Figure 4c shows the average rate of flagged posts of articles belonging to different sections of CNN.com.

Post flagging is more frequent in the health, justice, showbiz, sport, US, and world sections (near 4%), and less frequent in the opinion, politics, tech, and travel sections (near 2%). Flagging may be more common in the health, justice, US, and world sections because these sections tend to cover controversial issues: a linear regression unigram model using the titles of articles to predict the proportion of flagged posts revealed that “Nidal” and “Hasan” (the perpetrator of the 2009 Fort Hood shooting) were among the most predictive words in the justice section. For the showbiz and sport sections, inter-group conflict may have a strong effect (e.g., fans of opposing teams) [81]. Though political issues in the US may appear polarizing, the politics section has one of the lowest rates of post flagging, similar to tech. Still, a deeper analysis of the interplay of these factors (e.g., personal values, group membership, and topic) with trolling remains future work.

The relatively large variation here suggests that the topic of a discussion influences the baseline rate of trolling, where hot-button topics spark more troll posts.

Summary

Through experimentation and data analysis, we find that situational factors such as mood and discussion context can induce trolling behavior, answering our main research question (RQ). Bad mood induces trolling, and trolling, like mood, varies with time of day and day of week; bad mood may also persist across discussions, but its effect diminishes with time. Prior troll posts in a discussion increase the likelihood of future troll posts (with an additive effect the more troll posts there are), as do more controversial topics of discussion.

A MODEL OF HOW TROLLING SPREADS

Thus far, our investigation sought to understand whether ordinary users engage in trolling behavior. In contrast, prior work suggested that trolling is largely driven by a small population of trolls (i.e., by intrinsic characteristics such as personality), and our evidence suggests complementary hypotheses – that mood and discussion context also affect trolling behavior. In this section, we construct a combined predictive model to understand the relative strengths of each explanation.

We model each explanation through features in the CNN.com dataset. First, the impact of mood on trolling behavior can be modeled indirectly using seasonality, as expressed through time of day and day of week; and a user’s recent posting history (outside of the current discussion), in terms of the time elapsed since the last post and whether the user’s previous post was flagged. Second, the effect of discussion context can be modeled using the previous posts that precede a user’s in a discussion (whether any of the previous five posts in the discussion were flagged, and if they were written by the same user); and the topic of discussion (e.g., politics). Third, to evaluate if trolling may be innate, we use a user’s User ID to learn each user’s base propensity to troll, and the user’s overall history of prior trolling (the total number and proportion of flagged posts accumulated).

Our prediction task is to guess whether a user will write a post that will get flagged, given features relating to the discussion or user. We sampled posts from discussions at random (N=116,026), and balance the set of users whose posts are later flagged and users whose posts are not flagged, so that random guessing results in 50% accuracy. To understand trolling behavior across all users, this analysis was not restricted to users who did not have their posts previously flagged. We use a logistic regression classifier, one-hot encoding features (e.g., time of day) as appropriate. A random forest classifier gives empirically similar results.

Our results suggest that trolling is better explained as situational (i.e., a result of the user’s environment) than as innate (i.e., an inherent trait). Table 3 describes performance on this prediction task for different sets of features. Features relating to discussion context perform best (AUC=0.74), hinting that context alone is sufficient in predicting trolling behavior; the individually most predictive feature was whether the previous post in the discussion was flagged. Discussion topic was somewhat informative (0.58), with the most predictive feature being if the post was in the opinion section. In the experiment, mood produced a stronger effect than discussion context. However, here we cannot measure mood directly, so its feature sets (seasonality and recent user history) were weaker (0.60 and 0.53 respectively). Most predictive was if the user’s last post in a different discussion was flagged, and if the post was written on Friday. Modeling each user’s probability of trolling individually, or by measuring all flagged posts over their lifetime was moderately predictive (0.66 in either case). Further, user features do not improve performance beyond the using just the discussion context and a user’s recent history. Combining previous posts with recent history (0.77) resulted in performance nearly as good as including all features (0.78). We continue to observe strong performance when restricting our analysis only to posts by users new to a discussion (0.75), or to users with no prior record of reported or deleted posts (0.70). In the latter case, it is difficult to detect trolling behavior without discussion context features (<0.56).

Table 3

In predicting trolling in a discussion, features relating to the discussion’s context are most informative, followed by user-specific and mood features. This suggests that while some users are inherently more likely to troll, the context of a discussion plays a greater role in whether trolling actually occurs. The number of binary features is in parentheses.

Feature SetAUC
Mood
Seasonality (31)0.53
Recent User History (4)0.60

Discussion Context
Previous Posts (15)0.74
Article Topic (13)0.58

User-specific
Overall User History (2)0.66
User ID (45895)0.66

Combined
Previous Posts + Recent User History (19)0.77
All Features0.78

Overall, we find that the context in which a post is made is a strong predictor of a user later trolling, beyond their intrinsic propensity to troll. A user’s recent posting history is also predictive, suggesting that mood carries over from previous discussions, and that past trolling predicts future trolling.

DISCUSSION

While prior work suggests that some users may be born trolls and innately more likely to troll others, our results show that ordinary users will also troll when mood and discussion context prompt such behavior.

The spread of negativity

If trolling behavior can be induced, and can carry over from previous discussions, could such behavior cascade and lead to the community worsening overall over time? Figure 5 shows that on CNN.com, the proportion of flagged posts and proportion of users with flagged posts are rising over time. These upward trends suggest that trolling behavior is becoming more common, and that a growing fraction of users are engaging in such behavior. Comparing posts made in the first half and second half of the CNN.com dataset, the proportion of flagged posts and proportion of users with flagged posts increased (0.03 vs. 0.04 and 0.09 vs. 0.12, p<0.001). There may be several explanations for this (e.g., that users joining later are more susceptible to trolling), but our findings, together with prior work showing that negative norms can be reinforced [91] and that downvoted users go on to downvote others [17], suggest that negative behavior can persist in and permeate a community when left unchecked.

An external file that holds a picture, illustration, etc.
Object name is nihms933916f5.jpg

On CNN.com, the proportion of flagged posts, as well as users with flagged posts, is increasing over time, suggesting that trolling behavior can spread and be reinforced.

Designing better discussion platforms

The continuing endurance of the idea that trolling is innate may be explained using the fundamental attribution error [73]: people tend to attribute a person’s behavior to their internal characteristics rather than external factors – for example, interpreting snarky remarks as resulting from general mean-spiritedness (i.e., their disposition), rather than a bad day (i.e., the situation that may have led to such behavior). This line of reasoning may lead communities to incorrectly conclude that trolling is caused by people who are unquestionably trolls, and that trolling can be eradicated by banning these users. However, not only are some banned users likely to be ordinary users just having a bad day, but such an approach also does little to curb such situational trolling, which many ordinary users may be susceptible to. How might we design discussion platforms that minimize the spread of trolling behavior?

Inferring mood through recent posting behavior (e.g., if a user just participated in a heated debate) or other behavioral traces such as keystroke movements [49], and selectively enforcing measures such as post rate-limiting [2] may discourage users from posting in the heat of the moment. Allowing users to retract recently posted comments may help minimize regret [88]. Alternatively, reducing other sources of user frustration (e.g., poor interface design or slow loading times [12]) may further temper aggression.

Altering the context of a discussion (e.g., by hiding troll comments and prioritizing constructive ones) may increase the perception of civility, making users less likely to follow suit in trolling. To this end, one solution is to rank comments using user feedback, typically by allowing users to up- and down-vote content, which reduces the likelihood of subsequent users encountering downvoted content. But though this approach is scalable, downvoting can cause users to post worse comments, perpetuating a negative feedback loop [17]. Selectively exposing feedback, where positive signals are public and negative signals are hidden, may enable context to be altered without adversely affecting user behavior. Community norms can also influence a discussion’s context: reminders of ethical standards or past moral actions (e.g., if users had to sign a “no trolling” pledge before joining a community) can also increase future moral behavior [59, 65].

Limitations and future work

Though our results do suggest the overall effect of mood on trolling behavior, a more nuanced understanding of this relation should require improved signals of mood (e.g., by using behavioral traces as described earlier). Models of discussions that account for the reply structure [3], changes in sentiment [87], and the flow of ideas [66, 94] may provide deeper insight into the effect of context on trolling behavior.

Different trolling strategies may also vary in prevalence and severity (e.g., undirected swearing vs. targeted harassment and bullying). Understanding the effects of specific types of trolling may also allow us to design measures better targeted to the specific behaviors that may be more pertinent to deal with. The presence of social cues may also mediate the effect of these factors: while many online communities allow their users to use pseudonyms, reducing anonymity (e.g., through the addition of voice communication [23] or real name policies [19]) can reduce bad behavior such as swearing, but may also reduce the overall likelihood of participation [19]. Finally, differentiating the impact of a troll post and the intent of its author (e.g., did its writer intend to hurt others, or were they just expressing a different viewpoint? [55]) may help separate undesirable individuals from those who just need help communicating their ideas appropriately.

Future work could also distinguish different types of users who end up trolling. Prior work that studied users banned from communities found two distinct groups – users whose posts were consistently deleted by moderators, and those whose posts only started to get deleted just before they were banned [18]. Our findings suggest that the former type of trolling may have been innate (i.e., the user was constantly trolling), while the latter type of trolling may have been situational (i.e., the user was involved in a heated discussion).

CONCLUSION

Trolling stems from both innate and situational factors –where prior work has discussed the former, this work focuses on the latter, and reveals that both mood and discussion context affect trolling behavior. This suggests the importance of different design affordances to manage either type of trolling. Rather than banning all users who troll and violate community norms, also considering measures that mitigate the situational factors that lead to trolling may better reflect the reality of how trolling occurs.

Acknowledgments

We would like to thank Chloe Kliman-Silver, Disqus for the data used in our observational study, and our reviewers for their helpful comments. This work was supported in part by a Microsoft Research PhD Fellowship, a Google Research Faculty Award, NSF Grant IIS-1149837, ARO MURI, DARPA NGS2, SDSI, Boeing, Lightspeed, SAP, and Volkswagen.

Footnotes

1In contrast to cyberbullying, defined as behavior that is repeated, intended to harm, and targeted at specific individuals [82], this definition of trolling encompasses a broader set of behaviors that may be one-off, unintentional, or untargeted.

References

1. Akbulut Yavuz, Sahin Yusuf Levent, Eristi Bahadir. Cyberbullying Victimization among Turkish Online Social Utility Members. Educ Technol Soc 2010 [Google Scholar]
2. Atwood Jeff. Why is there a topic reply limit for new users? 2013 http://bit.ly/1XDLk8a.
3. Backstrom Lars, Kleinberg Jon, Lee Lillian, Danescu-Niculescu-Mizil Cristian. Characterizing and curating conversation threads: expansion, focus, volume, re-entry. Proc. WSDM.2013. [Google Scholar]
4. Baker Paul. Moral panic and alternative identity construction in Usenet. J Comput Mediat Commun 2001 [Google Scholar]
5. Banerjee Abhijit V. A simple model of herd behavior. Q J Econ 1992 [Google Scholar]
6. Barsade Sigal G. The ripple effect: Emotional contagion and its influence on group behavior. Adm Sci Q 2002 [Google Scholar]
7. Baumeister Roy F, Bratslavsky Ellen, Finkenauer Catrin, Vohs Kathleen D. Bad is stronger than good. Rev Gen Psychol 2001 [Google Scholar]
8. Berkowitz Leonard. Aggression: Its causes, consequences, and control 1993 [Google Scholar]
9. Binns Amy. DON’T FEED THE TROLLS! Managing troublemakers in magazines’ online communities. Journalism Practice 2012 [Google Scholar]
10. Bodenhausen Galen V, Kramer Geoffrey P, Süsser Karin. Happiness and stereotypic thinking in social judgment. J Pers Soc Psychol 1994 [Google Scholar]
11. Buckels Erin E, Trapnell Paul D, Paulhus Delroy L. Trolls just want to have fun. Pers Individ Dif 2014 [Google Scholar]
12. Ceaparu Irina, Lazar Jonathan, Bessiere Katie, Robinson John, Shneiderman Ben. Determining causes and severity of end-user frustration. Int J Hum Comput Interact 2004 [Google Scholar]
13. Centola Damon. The spread of behavior in an online social network experiment. Science 2010 [PubMed] [Google Scholar]
14. Chancellor Stevie, Lin Zhiyuan Jerry, De Choudhury Munmun. This Post Will Just Get Taken Down: Characterizing Removed Pro-Eating Disorder Social Media Content. Proc. SIGCHI.2016. [Google Scholar]
15. Chang Daphne, Krupka Erin L, Adar Eytan, Acquisti Alessandro. Engineering Information Disclosure: Norm Shaping Designs. Proc. SIGCHI.2016. [Google Scholar]
16. Chen Zhansheng, Williams Kipling D, Fitness Julie, Newton Nicola C. When Hurt Will Not Heal: Exploring the Capacity to Relive Social and Physical Pain. Psychol Sci 2008 [PubMed] [Google Scholar]
17. Cheng Justin, Danescu-Niculescu-Mizil Cristian, Leskovec Jure. How community feedback shapes user behavior. Proc. ICWSM.2014. [Google Scholar]
18. Cheng Justin, Danescu-Niculescu-Mizil Cristian, Leskovec Jure. Antisocial Behavior in Online Discussion Communities. Proc. ICWSM.2015. [Google Scholar]
19. Cho Daegon, Acquisti Alessandro. The more social cues, the less trolling? An empirical study of online commenting behavior. Proc. WEIS.2013. [Google Scholar]
20. Cialdini Robert B, Goldstein Noah J. Social influence: Compliance and conformity. Annu Rev Psychol 2004 [PubMed] [Google Scholar]
21. Cialdini Robert B, Reno Raymond R, Kallgren Carl A. A focus theory of normative conduct: recycling the concept of norms to reduce littering in public places. J Pers Soc Psychol 1990 [Google Scholar]
22. CNN. Community Guidelines. n.d http://cnn.it/1WAhxh5.
23. Davis John P, Farnham Shelly, Jensen Carlos. Decreasing online ‘bad’ behavior. Proc. SIGCHI Extended Abstracts.2002. [Google Scholar]
24. Diakopoulos Nicholas, Naaman Mor. Towards quality discourse in online news comments. Proc. CSCW.2011. [Google Scholar]
25. Discourse. This is a Civilized Place for Public Discussion. n.d http://bit.ly/1TM8K5x.
26. Donath Judith S. Identity and deception in the virtual community. Communities in Cyberspace 1999 [Google Scholar]
27. Duggan Maeve. Online Harassment. Pew Research Center; 2014. [Google Scholar]
28. Finley Klint. A brief history of the end of the comments. 2015 http://bit.ly/1Wz2OUg.
29. Forgas Joseph P, Bower Gordon H. Mood effects on person-perception judgments. J Pers Soc Psychol 1987 [PubMed] [Google Scholar]
30. Gardiner Becky, Mansfield Mahana, Anderson Ian, Holder Josh, Louter Daan, Ulmanu Monica. The dark side of Guardian comments. Guardian 2016 [Google Scholar]
31. Gino Francesca, Ayal Shahar, Ariely Dan. Contagion and differentiation in unethical behavior: the effect of one bad apple on the barrel. Psychol Sci 2009 [PubMed] [Google Scholar]
32. Golder Scott A, Macy Michael W. Diurnal and seasonal mood vary with work, sleep, and daylength across diverse cultures. Science 2011 [PubMed] [Google Scholar]
33. González-Ibáñez Roberto, Shah Chirag. Investigating positive and negative affects in collaborative information seeking: A pilot study report. Proc. ASIST.2012. [Google Scholar]
34. Greenberg Jeff, Simon Linda, Pyszczynski Tom, Solomon Sheldon, Chatel Dan. Terror management and tolerance: Does mortality salience always intensify negative reactions to others who threaten one’s worldview? J Pers Soc Psychol 1992 [PubMed] [Google Scholar]
35. The Guardian. Community Standards and Participation Guidelines. n.d http://bit.ly/1YtIxfW.
36. Hara Noriko, Bonk Curtis Jay, Angeli Charoula. Content analysis of online discussion in an applied educational psychology course. Instr Sci 2000 [Google Scholar]
37. Hardaker Claire. Trolling in asynchronous computer-mediated communication: From user discussions to academic definitions. J Politeness Res 2010 [Google Scholar]
38. Herring Susan, Job-Sluder Kirk, Scheckler Rebecca, Barab Sasha. Searching for safety online: Managing “trolling” in a feminist forum. The Information Society 2002 [Google Scholar]
39. Hsu Chiao-Fang, Khabiri Elham, Caverlee James. Ranking comments on the social web. Proc. CSE.2009. [Google Scholar]
40. Hutto Clayton J, Gilbert Eric. Vader: A parsimonious rule-based model for sentiment analysis of social media text. Proc. ICWSM.2014. [Google Scholar]
41. Jones John W, Anne Bogat G. Air pollution and human aggression. Psychol Rep 1978 [PubMed] [Google Scholar]
42. June Laura. I’m Voting for Hillary Because of My Daughter. 2015 http://thecut.io/1rNEeBJ.
43. Kayany Joseph M. Contexts of uninhibited online behavior: Flaming in social newsgroups on Usenet. J Am Soc Inf Sci 1998 [Google Scholar]
44. Keltner Dacher, Ellsworth Phoebe C, Edwards Kari. Beyond simple pessimism: effects of sadness and anger on social perception. J Pers Soc Psychol 1993 [PubMed] [Google Scholar]
45. Kendall Philip C, Robert Nay W, Jeffers John. Timeout duration and contrast effects: A systematic evaluation of a successive treatments design. Behavior Therapy 1975 [Google Scholar]
46. Kilner Peter G, Hoadley Christopher M. Anonymity options and professional participation in an online community of practice. Proc. CSCL.2005. [Google Scholar]
47. Kirman Ben, Lineham Conor, Lawson Shaun. Exploring mischief and mayhem in social computing or: how we learned to stop worrying and love the trolls. Proc. SIGCHI Extended Abstracts.2012. [Google Scholar]
48. Knobloch-Westerwick Silvia, Alter Scott. Mood adjustment to social situations through mass media use: How men ruminate and women dissipate angry moods. Hum Commun Res 2006 [Google Scholar]
49. Kołakowska Agata. A review of emotion recognition methods based on keystroke dynamics and mouse movements. Proc. HSI.2013. [Google Scholar]
50. Kramer Adam DI, Guillory Jamie E, Hancock Jeffrey T. Experimental evidence of massive-scale emotional contagion through social networks. PNAS 2014 [PMC free article] [PubMed] [Google Scholar]
51. Kriplean Travis, Morgan Jonathan, Freelon Deen, Borning Alan, Bennett Lance. Supporting reflective public thought with considerit. Proc. CSCW.2012a. [Google Scholar]
52. Kriplean Travis, Toomim Michael, Morgan Jonathan, Borning Alan, Ko Andrew. Is this what you meant? Promoting listening on the web with reflect. Proc. SIGCHI.2012b. [Google Scholar]
53. Lampe Cliff, Resnick Paul. Slash (dot) and burn: distributed moderation in a large online conversation space. Proc. SIGCHI.2004. [Google Scholar]
54. Lee Hangwoo. Behavioral strategies for dealing with flaming in an online forum. Sociol Q 2005 [Google Scholar]
55. Lee So-Hyun, Kim Hee-Woong. Why people post benevolent and malicious comments online. Commun ACM 2015 [Google Scholar]
56. Leith Karen Pezza, Baumeister Roy F. Why do bad moods increase self-defeating behavior? Emotion, risk tasking, and self-regulation. J Pers Soc Psychol 1996 [PubMed] [Google Scholar]
57. Levine Jerome M, Murphy Gardner. The learning and forgetting of controversial material. J Abnorm Soc Psychol 1943 [Google Scholar]
58. Li Tanya Beran Qing. Cyber-harassment: A study of a new method for an old behavior. J Educ Comput Res 2005 [Google Scholar]
59. Mazar Nina, Amir On, Ariely Dan. The dishonesty of honest people: A theory of self-concept maintenance. J Mark Res 2008 [Google Scholar]
60. McInnis Brian James, Murnane Elizabeth Lindley, Epstein Dmitry, Cosley Dan, Leshed Gilly. One and Done: Factors affecting one-time contributors to ad-hoc online communities. Proc. CSCW.2016. [Google Scholar]
61. McNair Douglas M. Manual profile of mood states 1971 [Google Scholar]
62. Milgram Stanley, Bickman Leonard, Berkowitz Lawrence. Note on the drawing power of crowds of different size. J Pers Soc Psychol 1969 [Google Scholar]
63. Milgram Stanley, Van den Haag Ernest. Obedience to authority 1978 [Google Scholar]
64. Muchnik Lev, Aral Sinan, Taylor Sean J. Social influence bias: A randomized experiment. Science 2013 [PubMed] [Google Scholar]
65. Nathan Matias J. Posting Rules in Online Discussions Prevents Problems & Increases Participation. 2016 http://bit.ly/2foNQ3K.
66. Niculae Vlad, Danescu-Niculescu-Mizil Cristian. Conversational markers of constructive discussions. Proc. NAACL.2016. [Google Scholar]
67. Park Deokgun, Sachar Simranjit, Diakopoulos Nicholas, Elmqvist Niklas. Supporting Comment Moderators in Identifying High Quality Online News Comments. Proc. SIGCHI.2016. [Google Scholar]
68. Pennebaker James W, Francis Martha E, Booth Roger J. Linguistic inquiry and word count: LIWC 2001 2001 [Google Scholar]
69. Raine Adrian. Annotation: The role of prefrontal deficits, low autonomic arousal, and early health factors in the development of antisocial and aggressive behavior in children. J Child Psychol Psychiatry 2002 [PubMed] [Google Scholar]
70. Raskauskas Juliana, Stoltz Ann D. Involvement in traditional and electronic bullying among adolescents. Dev Psychol 2007 [PubMed] [Google Scholar]
71. Rensin Emmett. Confessions of a former internet troll. Vox 2014 [Google Scholar]
72. Rosenbaum Paul R, Rubin Donald B. The central role of the propensity score in observational studies for causal effects. Biometrika 1983 [Google Scholar]
73. Ross Lee. The intuitive psychologist and his shortcomings: Distortions in the attribution process. Adv Exp Soc Psychol 1977 [Google Scholar]
74. Rotton James, Frey James. Air pollution, weather, and violent crimes: concomitant time-series analysis of archival data. J Pers Soc Psychol 1985 [PubMed] [Google Scholar]
75. Rozin Paul, Royzman Edward B. Negativity bias, negativity dominance, and contagion. Pers Soc Psychol Rev 2001 [Google Scholar]
76. Salganik Matthew J, Dodds Peter Sheridan, Watts Duncan J. Experimental study of inequality and unpredictability in an artificial cultural market. Science 2006 [PubMed] [Google Scholar]
77. Schwartz Mattathias. NY Times Magazine. 2008. The trolls among us. [Google Scholar]
78. Schwarz Norbert, Bless Herbert. Happy and mindless, but sad and smart? The impact of affective states on analytic reasoning. Emotion and Social Judgments 1991 [Google Scholar]
79. Schwarz Norbert, Clore Gerald L. Mood, misattribution, and judgments of well-being: informative and directive functions of affective states. J Pers Soc Psychol 1983 [Google Scholar]
80. Shachaf Pnina, Hara Noriko. Beyond vandalism: Wikipedia trolls. J Inf Sci 2010 [Google Scholar]
81. Sherif Muzafer, Harvey Oliver J, Jack White B, Hood William R, Sherif Carolyn W, et al. Intergroup conflict and cooperation: The Robbers Cave experiment 1961 [Google Scholar]
82. Slonje Robert, Smith Peter K, FriséN Ann. The nature of cyberbullying, and strategies for prevention. Comput Hum Behav 2013 [Google Scholar]
83. Sukumaran Abhay, Vezich Stephanie, McHugh Melanie, Nass Clifford. Normative influences on thoughtful online participation. Proc. SIGCHI.2011. [Google Scholar]
84. Suler John. The online disinhibition effect. Cyberpsychol Behav 2004 [PubMed] [Google Scholar]
85. Terry Peter C, Lane Andrew M. Normative values for the Profile of Mood States for use with athletic samples. J Appl Sport Psychol 2000 [Google Scholar]
86. Varjas Kris, Talley Jasmaine, Meyers Joel, Parris Leandra, Cutts Hayley. High school students perceptions of motivations for cyberbullying: An exploratory study. West J Emerg Med 2010 [PMC free article] [PubMed] [Google Scholar]
87. Wang Lu, Cardie Claire. A Piece of My Mind: A Sentiment Analysis Approach for Online Dispute Detection. Proc. ACL.2014. [Google Scholar]
88. Wang Yang, Norcie Gregory, Komanduri Saranga, Acquisti Alessandro, Leon Pedro Giovanni, Cranor Lorrie Faith. I regretted the minute I pressed share: A qualitative study of regrets on Facebook. Proc. SUPS.2011. [Google Scholar]
89. Wheeler Ladd. Toward a theory of behavioral contagion. Psychol Rev 1966 [Google Scholar]
90. Wiener David. Negligent publication of statements posted on electronic bulletin boards: Is there any liability left after Zeran? Santa Clara L Rev 1998 [Google Scholar]
91. Willer Robb, Kuwabara Ko, Macy Michael W. The False Enforcement of Unpopular Norms. Am J Sociol 2009 [PubMed] [Google Scholar]
92. Wilson James Q, Kelling George L. Broken Windows. Atlantic Monthly 1982 [Google Scholar]
93. Wyland Carrie L, Forgas Joseph P. On bad mood and white bears: The effects of mood state on ability to suppress unwanted thoughts. Cognition and Emotion 2007 [Google Scholar]
94. Zhang Justine, Kumar Ravi, Ravi Sujith, Danescu-Niculescu-Mizil Cristian. Conversational flow in Oxford-style debates. Proc. NAACL.2016. [Google Scholar]
95. Zimbardo Philip G. The human choice: Individuation, reason, and order versus deindividuation, impulse, and chaos. Nebr Symp Motiv 1969 [Google Scholar]
Feedback