key: cord-0678562-tgksf81t authors: Sipka, Andrea; Hannak, Aniko; Urman, Aleksandra title: Comparing the Language of QAnon-related content on Parler, Gab, and Twitter date: 2021-11-22 journal: nan DOI: nan sha: 6605e18e80121b32fbf4610e4153f179f8a35254 doc_id: 678562 cord_uid: tgksf81t Parler, a"free speech"platform popular with conservatives, was taken offline in January 2021 due to the lack of moderation of hateful and QAnon- and other conspiracy-related content that allegedly made it instrumental in the organisation of the storming of the US Capitol on January 6. However, Parler co-existed with other social media platforms, and comparative studies are needed to draw conclusions about the prevalence of anti-social language, hate speech, or conspiracy theory content on the platform. We address this through a cross-platform comparison of posts related to QAnon. We compare posts with the hashtag #QAnon on Parler over a month-long period with posts on Twitter and Gab. In our analysis, Parler emerges as the platform with the highest average toxicity of the posts, though this largely stems from the distinctive way hashtags are used on this platform. Gab has the highest proportion of #QAnon posts with hate terms, and Parler and Twitter are similar in this respect. On all three platforms, posts mentioning female political figures, Democrats, or Donald Trump have more anti-social language than posts mentioning male politicians, Republicans, or Joe Biden. An analysis of entities mentioned in posts revealed differences in content - with Twitter posts mentioning prominent figures related to QAnon, while Parler and Gab posts mention entities related to conspiracies. Narrative analysis indicates that the discussion on Twitter centres on QAnon and is critical of it, Parler focuses on supporting Donald Trump, while on Gab the discussion focuses on more conspiratorial content, in relation to both Trump and other political figures. On January 6, 2021, a group of supporters of the President of the United States Donald Trump stormed the US Capitol building, interrupting a joint congressional session that was counting the electoral votes following the November 2020 presidential election and was due to confirm Joe Biden as the winner. For two months prior to the storming, Donald Trump repeatedly claimed that the election was being stolen. Trump's supporters started movements on multiple social networks (e.g. #StopTheSteal) and held protests across the United States, some of which turned violent. The culmination of that movement, the Capitol riot itself, resulted in five deaths, with more than 140 injured, and was the subject of the second impeachment trial of Donald Trump. Following the riot, various media outlets made attempts to uncover how the riot had been organised. Many reputable media outlets identified Parler, a social networking site launched in 2018 as a "free speech alternative" to more established platforms such as Twitter, as the place where the organisation took place (Frenkel 2021; Timberg and Harwell 2021) . In connection to these allegations, in the days following the riot, a growing number of service providers pulled support from Parler, and it was removed from the most popular app stores. This ultimately resulted in Parler being taken offline on January 10, 2021, due to the inability to host the service. Parler returned online on February 15, 2021, following the platform's migration to a set of alternative service providers, although with all the previous data removed. Since its inception, Parler was increasingly popular with users on the right of the political spectrum, with many conservative commentators and politicians joining and advocating for the platform between 2018 and 2020. The 8th most popular hashtag on Parler prior to the shutdown was #QAnon (Aliapoulios et al. 2021a ), used to denote a far-right conspiracy theory suggesting that "Trump has been battling against a satan worshipping global child sextrafficking ring and an anonymous source called 'Q' is cryptically providing secret information about the ring" (Zuckerman 2019). QAnon-related hashtags were amongst the top hashtags used in profile descriptions of users spreading misinformation related to the 2020 presidential election on Twitter (EIP 2020), and among the top 10 hashtags in a sample of 600 million US election-related tweets (Ferrara et al. 2020 ). The followers of QAnon have been spreading conspiracy theories relating to elections, and have participated in the storming of the Capitol (most notably, "QAnon Shaman" Jake Angeli) (US Department of Justice 2021). For this reason, we analyse posts containing the hashtag #QAnon in the period between December 7, 2020, and January 10, 2021, as it encompasses multiple events of interest: the "Stop the steal" movement, the US Senate run-off election in Georgia, multiple protests of election results, including the January 6 riot, the confirmation of Biden as the election winner, Twitter banning Trump, and the announcement of the (temporary) shutdown of Parler. Despite the attention Parler has garnered in relation to the January 6 riot, and the consequences the platform has suffered in connection to the claims about the insufficient moderation or prevention of spread of dangerous and egregious content, Parler was not the only social media platform used by the rioters, as evidenced by the court documents in cases brought against the participants (Brewster 2021) . Parler exists in a large ecosystem of platforms, which have varied functionalities and moderation policies. While many researchers have studied political movements on social media platforms, few studies have been published on Parler. To make a fair judgment about the difference in discourse, tone, and anti-social language, the scientific community needs to conduct studies comparing multiple similar platforms, yet comparative studies are rare. We aim to partially address the existing gap with the present study encompassing three platforms -Parler, Twitter, and Gab. We choose Twitter and Gab as comparison points to Parler for two reasons. Firstly, the three platforms are very similar in basic functionality, being primarily microblogging platforms with the same mechanism of posting and engaging with the content, which makes them comparable. We do not include other platforms with fundamentally different architectures (e.g. Facebook or Reddit) as that would decrease the internal validity of the analysis. Secondly, as insufficient content moderation was frequently cited as a reason for removing Parler's access to services, a comparison with platforms that have opposing types of moderation policies is particularly relevant -Twitter has been engaging in moderation of QAnon content (TwitterSafety 2020), while Gab is marketing itself in similar ways Parler did ("A social network that champions free speech, individual liberty and the free flow of information online"), and has been reported as one of the platforms that Parler users migrated to following the shutdown (Wilson 2021) . Our aim is to compare Parler, Twitter, and Gab posts that have used the same hashtag in the same time period, to gain an understanding of differences and similarities between the three platforms in the context of the discourse surrounding #QAnon in the run-up, and for a few days following, the Capitol riot. We analyse the volume of posting and the number of users who post to gauge the activity and interest levels across three platforms related to QAnon. In addition, as lack of moderation of threatening and harmful content was a frequent criticism of Parler, we analyse multiple aspects of anti-social language (such as hateful, threatening, and toxic language). This allows us to draw conclusions on how Parler compares to similar moderated and unmoderated platforms. We look at named entities used across the platforms, focusing on similarities and differences, to increase understanding of which figures, events and similar entities are mentioned in posts. In particular, as the period analysed includes major political events in the US, we test if the language differs in posts mentioning groups of political figures. We look at the posts about politicians from the opposite political parties (as Gab and Parler are considered right-wing), and posts mentioning politicians of different genders 1 (as previous re-search reported higher levels of incivility on social media towards women in politics (Rheault, Rayment, and Musulan 2019) ). Finally, we analyse the prevalent narratives on the three platforms to gain a deeper understanding of differences and similarities in the conversation. We formulate these aims as research questions: • RQ1: How does the volume of #QAnon posts, and the number of users posting, vary across the three platforms? • RQ2: How do the three platforms compare with respect to the prevalence of anti-social language? • RQ3: Are there differences in the relative prevalence of themes and political figures mentioned between the three platforms, and in the use of anti-social language related to those mentions? • RQ4: What are the prevalent narratives on the three platforms, and how do they compare? For both Parler and Gab, we used unofficial APIs to collect the data. This data was, at the time of collection, publicly available to anyone with an account on the platform (Parler), or anyone on the internet (Gab). Our data collection was non-intrusive as it did not affect users or platforms. The collection of Twitter data is in line with the platform's Terms of use, through the official API. For all three platforms, while the original data included usernames, bios, and similar information which could lead to the identification of individuals, we have excluded these fields from the analysis, thus anonymising the data. The posts were only collected for accounts that were not private at the time. Twitter Twitter is one of the most, if not the most, studied social network. Regarding QAnon-related content in particular, a study of banned Twitter users found that "QAnon is well-positioned at the centre of the political hashtag community" of banned users (Chowdhury et al. 2020) . Numerous studies have been conducted on hate speech on Twitter. For the present research, of particular relevance are studies of hate speech related to violent or political events. A study of hate speech and white nationalist language on Twitter around and after the 2016 US presidential election reveals "no evidence of an increase in hate speech before or after the election", while noting that there are short time periods where the level of hate rises, finding "evidence of tens of thousands of tweets containing hate speech and white nationalist rhetoric on Twitter" (Siegel et al. 2019) . Gab To the best of our knowledge, no studies examining QAnon on Gab have been conducted, although a study of topic evolution on the platform found that by the start of 2018, the discourse on Gab has switched to alt-right political topics, and in particular posts related to QAnon (McIlroy-Young and Anderson 2019). Related to language on the platform, previous research on Gab found that 5.4% of Gab posts contain at least one hate word, and that the most prevalent points of discussion on Gab are news, events, and conspiracy theories (Zannettou et al. 2018 ). 90% of posts on Gab were found to have toxicity scores less than 0.7 (on a scale from 0 to 1), although the authors note that there are users who "abuse the lack of moderation to spread hate" (Lima et al. 2018) . Parler Most of the research on Parler is still undergoing peer review, and to the best of our knowledge, nobody examined QAnon or anti-social language on the platform. However, studies conducted include reports of the number of users on Parler having more than doubled within weeks around the 2020 US election, and that Parler "has emerged as a space in which accounts that have been suspended by Twitter Safety continue to communicate with their audiences" (Thiel et al. 2021) . A study supplementary to a release of a large Parler dataset has found that Parler has experienced growth in user base in close proximity to "online censorship on mainstream platforms like Twitter, as well as events related to US politics" (Aliapoulios et al. 2021a ). In addition, the study reports that QAnon is the 8th most popular hashtag on Parler. Cross-platform studies Cross-platform studies allow for a fair comparison and deepen our understanding of the roles different platforms play in the ecosystem of information. The latter is particularly important when it relates to phenomena damaging to society (e.g. conspiracy theories) or to individuals (e.g. anti-social language). However, not many studies have been published in the context of conspiracyrelated content, anti-social language or hate speech. Notable exceptions include an analysis and comparison of language related to QAnon, across sites collecting posts claimed to be written by "Q", Twitter, 4chan, 8chan, Reddit, and Voat (Aliapoulios et al. 2021b ). The study suggested that the QAnon community can find a home for their content on mainstream platforms, and that bans on one platform "do little to slow growth on others". The same study reports that posts written by "Q" are less toxic than posts by QAnon communities on sites such as Voat and 4chan, although the study does not capture the same measure for Twitter. Several comparative studies have been published on social media engagement with the Covid-19 pandemic, some focusing on conspiratorial content. A study of Covid-19related conspiracy theories on 8kun and Gab used 8kun's most prominent QAnon related board "/qresearch" to collect pandemic-related conspiracy theories (Zeng and Schäfer 2021) . The study found that 24% of the Covid-related posts on Gab contain conspiracy theories, and that 57% of randomly selected user profiles contain conspiracy theory content, such as QAnon, although that the prevalence of such content is higher on 8kun. Another study has focused on the effect of moderation, contrasting Facebook, Twitter, Reddit, and 4chan (Papakyriakopoulos, Serrano, and Hegelich 2020) , finding that "content moderation on Twitter was less effective than on the other platforms", potentially attributed to the fact that content on Twitter spreads within the first hours of it being posted. Finally, a study on the emergence of sinophobic behaviour on web communities in the face of the pandemic on Twitter and 4chan has focused on content analysis, including racist slurs, as a type of hate speech (Tahmasbi et al. 2021 ). Other existing scholarship involving multiple platforms and hate speech includes a comparative study of hate speech on Twitter and Reddit around attacks involving Arabs and Muslims as perpetrators or victims (Olteanu et al. 2018 ). The authors observed that "extremist violence tends to lead to an increase in online hate speech", with the biggest increase seen in messages advocating for the violence. A comparison of Gab and 4chan's Politically Incorrect board (/pol/) in the context of antisemitism, found evidence of increasing antisemitism around political events such as the US presidential election of 2016 (Zannettou et al. 2020b ). Finally, Gab was reported to have a "high percentage of news shared to be from untrustworthy sources (48.7%), compared to 8.7% for Twitter" (Wang et al. 2021 ). To compare the discourse related to the #QAnon, we have collected the data from all three platforms, including all posts which have been posted with the #QAnon hashtag (not case sensitive) in the period between December 7, 2020 (15:44:33) and January 10, 2021 (20:22:15) CET. We obtained Parler data via Parlance API (Castle Lemongrab 2020). While unofficial, we confirmed that the API has allowed us to access the same data that would have been accessible if we had used Parler's search feature to search for the same hashtag. The data collection started on January 6 and ended when Parler went offline, at 08:59 CET on January 12. We collected data for the same hashtag, for the same time period from Gab using garc library (Stevens 2020) , and from Twitter using Twitter's Academic full archive search API, via the academictwitteR library (Twitter 2021; Barrie and Ho 2021) . The data for all three platforms included the text of the post, as well as metadata (such as information on author, time, date). Sampling strategy #QAnon is not the only hashtag used by QAnon supporters. A notable alternative, #wwg1wga ("where we go one, we go all"), was reportedly more popular than #QAnon on Parler, although in a sample that is not statistically representative (Aliapoulios et al. 2021a ). On the other hand, our experiments showed that #QAnon is more popular than #wwg1wga on Twitter. When a movement uses multiple hashtags, there is not a perfect choice of a hashtag to analyse. We choose #QAnon as we believe that a person who is not well-versed or already a believer of the conspiracy theory would not know what #wwg1wga is, and #QAnon is a natural and easy first choice to search for. In addition, we highlight that our study aims to establish how platforms compare to each other, rather than how #QAnonrelated posts compare to users' overall activity on each platform. We do not draw conclusions about how #QAnon compares to non-#QAnon content, or other hashtags. On comparability of platforms For all three platforms, we analyse re-posts made with additional comments as they provide additional information, while excluding verbatim re-posts from the analysis. We make this choice as the data retrieval methods for Parler and Gab do not allow us to know what posts have been re-posted without additional comment within the time period of interest. Parler and Gab allow comments on posts, while Twitter does not. We exclude comments from the analysis to ensure comparability, although we note that language in comments may differ from the language in posts, which warrants a separate study. Hashtags are a feature on all three platforms, and their intended use is the same -to allow users to add keywordlike metadata to their posts, signifying a topic. All three platforms enable their users to add any number of hashtags to their posts -but the hashtags count within the character limit. Character limits vary across platforms -the maximum length of posts on Twitter, Parler and Gab being 280, 1000, and 3000 characters respectively. As such, hashtagging behaviour is different across the platforms, and the ratio of hashtags to text across the three data sets varies. While this has implications for our results, cross-platform research that is necessary to understand social media usage in perspective inevitably has to deal with the differences in platform functionalities and/or usage and account for them, as we do and describe below. Further, as the intended function of hashtags is the same across the three platforms, the comparison remains relevant despite these differences. Language Twitter API returns the information on the language of the post. For Parler and Gab, we have used Perspective API's automatic language detection to determine the language (Jigsaw 2021). In our data, 96.9% of Parler, and 95.9% of Gab are posts in English, and while Twitter contains a larger proportion of non-English posts (19% in languages other than English), we limit our analysis to posts written in English, to ensure the comparability across the three platforms. Table 1 shows the number of posts that have been collected, and analysed, across the three platforms. Moderation Twitter has been moderating content related to QAnon since at least late July 2020 (TwitterSafety 2020). The implication is that there likely is a body of Tweets that Twitter has removed and are not available for analysis. It is not unreasonable to assume that these Tweets might contain a higher amount of anti-social language, although that can't be established without an analysis made impossible due to the lack of data. Due to this, we consider anti-social language results on Twitter a lower bound for the platform. We collected Parler data over the last few days of Parler being online in January, and Gab data on February 28. While Gab reportedly has little to no moderation (Lima et al. 2018 Table 2 : Means (µ) and standard deviations (σ) for number of characters, words and hashtags in posts had been posted. Still, given the largely unmoderated nature of both Gab and Parler, we argue that it is reasonable to assume that only a small number of posts would have been removed, if any, and that our Parler and Gab data truthfully represent the conversation on those platforms, as it happened. Time period Our study focuses on a very specific time period. While this is partially driven by the fact that Parler going offline interrupted our data collection, it also allowed us to achieve higher internal validity of the data. Additionally, the month of analysis includes events that could increase interest in QAnon, as well as lead to increased activity on social media platforms. Conversely, this means that the volume of data collected is limited, which could affect analysis of language undertaken as part of RQ3 and RQ4. In future work, it could be relevant to examine the development of cross-platform QAnon-related discourse over time and, for instance, compare the observations during the period after the US presidential election, shortly before it and in more routine periods. To understand how the volume of #QAnon activity, number of users participating in the discussion, and posting styles vary across the three platforms, we utilise metadata extracted from posts, which allow us to summarise temporal posting activity, number of users posting, and similar descriptive statistics. In addition to analysing the metadata, we extracted various features from the text, such as hashtags, and text length descriptive statistics, to capture differences in posting styles. Table 1 shows the volume of posts with #QAnon for the observation period on each platform. On Parler, the number of posts was much higher than on Gab and Twitter. Though at first glance that indicates a much higher prevalence of #QAnon-related discussions on Parler as compared to Gab and Twitter, the observation might be partially explained by the differences in the way hashtags are used across platforms. The discrepancies in the number of users who posted about #QAnon between platforms were smaller than the differences in the sheer volume of posts with this hashtag. Though Parler contained substantially more posts with #QAnon than Twitter, the number of unique users posting with the hashtag was higher on the latter (Table 1) . This difference might mean that Parler and Gab users include this hashtag in posts not explicitly about QAnon, or are more in- Average account ages on the three platforms differ as well. The average account age at the time of posting on Twitter was 2781 days, on Parler 131 days, and on Gab 675 days. To examine whether this discrepancy is attributed to the difference in the age of the platform alone (Twitter was founded in 2006 , Gab in 2016 , and Parler in 2018 or to external events, we further scrutinised the dates when users joined each platform. We found that 19.8% of #QAnon posts on Parler and 18.8% on Gab were created by users who have been on the platform for less than 1 month. On Twitter, the share of such users is only 1.7%. This implies that the average account age is not attributed only to the age of the platform but also external events such as the election and events that followed: of the users posting with the #QAnon, 1.9% have joined Twitter after the Election (November 3, 2020), compared to 56.1% on Parler, and 40.1% on Gab. However, the apparent growth in the number of users who joined Gab and Parler during the one-month observation period was not accompanied by the bursts of #QAnon posting activity. The number of posts with the hashtag was relatively stable across the three platforms up until early January, as depicted in Figure 1 . Then both Twitter and Parler but not Gab saw a rise in popularity of the hashtag around January 6, with the conversation on Parler further intensifying in volume in the run-up to the platform going offline. In Table 2 we report the means and standard deviations for character counts, word counts, and the number of hashtags. Unsurprisingly, posts on Gab and Parler are longer on average as the two platforms have higher character limits than Twitter. The higher limit leaves the users of the two "alt-tech" platforms more space for hashtags, with the average number of hashtags being much higher on Gab and Parler than on Twitter, and 75.5% of Parler posts having more hashtags than other words, compared to 51.7% for Gab, and 17.5% for Twitter. The extra character limit is employed for particularly active hashtagging by Parler users. The mean words to hashtag ratio on Parler is 1.44, compared to 6.94 on Gab and 10.15 on Twitter. This points to a very different way in which hashtags are used across the platforms. In fact, upon a qualitative inspection of a sample from all three, we noticed that many Parler posts used "hashtag walls" -blocks of many continuous hashtags, not necessarily related to the We used Perspective API to extract features that act as proxies of anti-social language (Jigsaw 2021). While it has exhibited bias in certain contexts (e.g. classifying language predominantly used by African-Americans as more toxic than the language used by white people (Sap et al. 2019) ), it was also reported to outperform other available tools on similar texts (Zannettou et al. 2020a) . For each feature of interest, when queried against text, Perspective provides a score between 0 and 1, denoting the probability that the post is as the feature describes. As it is a probability, a threshold that is set to determine if a post is, for example, severely toxic, is both arbitrary and set according to the application of interest. In our comparison, rather than choosing an arbitrary threshold, we look at the distribution of scores to capture the differences. We perform statistical testing using a twosample Kolmogorov-Smirnov test to test for the difference in score distributions. All results we report below are statistically significant at p <0.001. The features we have selected to analyse are presented in Table 3 . To examine the effect of hashtags on anti-social language measures, due to the difference in hashtagging behaviour across the platforms discussed in the previous section, we query Perspective three times for each post. Firstly, we query it on the post cleaned of URLs and mentions, secondly having additionally removed all hashtags, and finally on hashtags alone. The cross-platform differences in hashtagging have important implications for the results in relation to the prevalence of anti-social language. In Figure 2 , we show cumulative distributions of severe toxicity scores across the We further explored the relationship between the presence of hashtags and the posts' classifications. In Figure 3 , we show the distribution of differences in Severe Toxicity scores between full posts and posts with tags removed, by platform. Values smaller than 0 indicate that a post is assigned a higher severe toxicity score with hashtags. As a much longer left tail for Parler indicates, Parler posts with hashtags included are assigned much higher severe toxicity scores compared to the same posts without hashtags. Severe toxicity scores for different post components across all three platforms are in Figure 4 . Parler's hashtagless texts, compared with hashtags only or combined (texts and hashtags), are overwhelmingly not severely toxic, and compare to scores of tweets. Mean scores for severe toxicity across time on all three platforms are presented in Figure 5 . The addition of hashtags changes the overall picture in this case as well. For full posts (with hashtags), Parler consistently has a higher mean severe toxicity score assigned than Twitter and Gab, which have similar scores to each other. On the other hand, with the hashtags removed, Parler consistently has lower mean severe toxicity scores than both Twitter and Gab, except for a large uptick in toxicity seen after January 6. Cumulative distribution functions (CDFs) for Insult, Identity attack, and Threat features are presented in Figure 6 . The model identifying threatening language shows that posts on all three platforms do not have a high probability of being threatening, with platforms having a similar proportion of posts more likely than not to be threatening (with a probability over 0.5). The identity attack model shows Parler posts scoring the worse, with Twitter and Gab being very similar. Finally, the platforms only show minor differences when it comes to insults, although a lower proportion of Gab posts are likely to be insulting than posts on Twitter and Parler. While Perspective allows us to capture measures such as toxicity, and it takes into account words that are considered hate speech, we separately measure hate speech specifically, as the most extreme type of anti-social language. We perform keyword-based classification by using Hatebase lexicon of hateful terms (Hatebase Inc. 2021). Hatebase collects words and phrases considered hateful across multiple categories, such as ethnicity, nationality, religion, gender, class, etc. We matched cleaned text (without links, mentions, or hashtags) from all three platforms with words and phrases in the lexicon. Upon inspection, a few ambiguous words have resulted in a high number of false positives (e.g. "Apple" is in Hatebase lexicon as it can be used to signify "An American Indian who is 'red on the outside, white on the inside."', but all occurrences of the word in our corpus referred to the company Apple Inc., or the fruit apple). For this reason, we have omit- ted a number of words 2 from the analysis. We chose words to remove by manually going through each occurrence of a hate word, and removing it from a dictionary if all occurrences of it were used in a non-hateful context. While Gab scores similarly, or even better, on different anti-social language measures than Twitter and Parler, it takes the leading position in terms of the share of posts with hate words (see Table 4 for the exact percentages and corresponding numbers of posts). We conclude by noting that most indicators show that the language of most posts across all three platforms should not be considered anti-social. An exception is a higher probability of posts being insulting on Twitter and Parler. Language on Parler appears marginally worse in terms of severe toxicity and identity attacks than the other two platforms, although we have demonstrated that this result is greatly affected by the prolific hashtagging of Parler users. While most Gab posts have a lower probability of being anti-social than Parler, and even Twitter for some indicators, they also have the highest percentage of hate speech occurrence. To compare similarities and differences in mentioned themes (e.g. individuals, locations, phrases), we use the NLP library Stanza's named entity recognition feature (Qi et al. 2020) on clean text (after removing links, mentions, and hashtags). We discarded some categories of entities, such as numbers and percentages, as they were of no interest for the analysis. The resulting entity list required cleaning, as terms signifying the same entity are not always automatically recognised as the same (e.g., "January 6", "Jan 6", "6. January"). We calculate the frequency of occurrence of enti- Table 5 : Top 20 terms on the three platforms ties across platforms, and compare their relative popularity. Following the removal of irrelevant entities, we have obtained 12759 named entities from Twitter, 90769 from Parler, and 15937 from Gab. 50.8% of Twitter posts, 42.7% of Parler posts, and 54.7% of Gab posts contain at least one entity. The number of unique entities extracted from Twitter is 3937, from Parler is 17439, and from Gab is 5066, suggesting that the conversation on Twitter is more focused around a few subjects, while on Parler and Gab there is a higher diversity of entities discussed in connection to #QAnon. The top 20 entities from all three platforms are shown in Table 5 . Many entities are popular across all platforms (Donald Trump, United States, Georgia, Twitter, Joe Biden, American(s), Democrat). Even the terms which appear in the top 20 on only two of the three platforms (e.g. "Republican" and "GOP" on Twitter and Parler) are still just outside of the top 20 on the third (25th, and 33rd most popular on Gab, respectively). This suggests that the discourse in relation to #QAnon on all three platforms was centred largely around similar subjects. To infer the differences between popular entities, we first removed entities used by less than 10 users (on either platform). This is to filter out unusual and erroneous entities which are used by one user repeatedly (e.g. "q and the plan to save the world"). We present the top 30 entities that are relatively more popular on one platform (in the top 50 of the most popular entities), as compared to the other two, in Table 6 . If r(x) t , r(x) p , r(x) g represent the ranking of popularity of term (x) on Twitter, Parler and Gab, the maximum difference is calculated as The results are indicative of potential differences in the focus of the discussions on the three platforms. For instance, mentions of the Capitol, Jake Angeli ("QAnon Shaman"), Antifa, Nazis as well as Kelly Loeffler and Marjorie Taylor Greene were more prevalent on Twitter than on Parler or Gab, while the latter two platforms saw higher popularity of Pennsylvania (probably connected to the vote count in the state) and Dominion voting machines (according to conspiracy theories, a company which aided the "stealing" of the election), as well as Deep state (Gab), MSM ("mainstream media", Gab), 2nd Amendment (Parler) and defense (Gab). To analyse if there is a difference in the use of anti-social language when discussing selected groups of political figures, we examined entities occurring at least 20 times in the whole corpus manually, and classified them into groups of interest. We form four groups, with political figures divided by gender and party 3 (Table 7) . We consider either the party membership, or service in an administration, when dividing by party lines. As in RQ2, we ensure that we only report results stemming from different distributions (tested using a two-sample K-S test, at p<0.001, excluding posts that include both groups being compared to ensure the independence of samples). Differences are observed with regard to the groups of political figures and two presidential candidates (Table 8) . While Donald Trump and male Republican politicians were 3 excluding Trump and Biden, who are analysed separately mentioned consistently more (in terms of the share of posts mentioning them) than their Democratic counterparts across all platforms, there was a divergence in posts about female politicians. There was a higher share of posts about female Republicans than Democrats on Twitter, while on Gab and Parler, the situation was reversed. We observe significant differences in the mean Perspective scores of posts mentioning different groups of politicians (female vs male; Republican vs Democrat). On all three platforms, posts mentioning female (vs male) politicians, Democrats (vs Republicans) and Trump (vs Biden) scored higher on average for anti-social language features that exhibited significant cross-group differences (Table 9 ). In conclusion, the entity-based analysis shows that while most prevalent named entities across the platforms are similar, there are differences between the mainstream platform and the alt-tech platforms. Twitter saw more mentions of high profile individuals considered QAnon supporters (Jake Angeli, Marjorie Taylor Greene), while more Parler and Gab posts used conspiratorial terms (Dominion, Deep state, Mainstream media). On all three platforms, posts mentioning female politicians, Democrats and Donald Trump score higher on all anti-social language features than posts mentioning male politicians, Republicans, and Joe Biden. While NER-based analysis allows us to measure the prevalence of terms on each platform, as well as undertake anti- social language analysis related to groups of interest, it does not offer insight into the context in which those terms are mentioned. Simply knowing that something, for example QAnon, is mentioned a lot does not allow us to capture if one platform is overwhelmingly critical of QAnon, while the other is supportive. To deepen our understanding of content across the three platforms, we analyse narratives and interconnected terms, using an adjusted Relatio library, and the method presented alongside it (Ash, Gauthier, and Widmer 2021) . We use Ash et al. operationalisation of narratives as triples of words or phrases that take the form Agent -Verb -Patient (Figure 7) . The narrative analysis pipeline is presented in Figure 8 . In pre-processing, we split posts into sentences, as the method can not reliably connect narratives spanning multiple sentences. Sentences can contain incomplete narratives (which we omit from analysis), or one or more complete narratives, such as the example we give in Figure 7 . Special features of social media posts, such as hashtags, required adjustment of Ash et al. pre-processing. While some users use hashtags out of context, to boost visibility of their posts (in our case, especially on Parler as discussed previously), hashtags can be used in context, and simple removal would result in loss of information. To overcome this, we make best effort to remove hashtag blocks while preserving hashtags that are likely used in context. As some hashtags contain multiple words (e.g. #fakenews in Figure 7) , we manually establish how to split hashtags that occur more than five times on either platform. This maximises the information extracted from data, as we do not simply discard posts such as the one given in Figure 7 . Finally, we clean text by removing mentions, links, emoji, and stopwords from posts, and remove duplicate posts made by the same person to ensure that spamming does not affect our analysis. After pre-processing, we use semantic role labelling to We convert verbs to their root form, reducing the dimensionality of verbs to 3510. While dimensionality reduction results in some loss of information, it is necessary due to language diversity. A summary of the number of sentences and narratives is available in Table 10 . Of the 44395 complete narratives, 31180 are unique. The reason for this is a high number of unique verbs that connect Agents and Patients. While we reduced the dimensionality of verbs from 8213 to 3510, we could not reliably group them according to the similarity of action they represent. The overlap between unique narratives between the platforms is presented in Figure 10 . 96% of narratives appear on only one of the three platforms, with only 56 narratives appearing on all three. Parler shares more narratives with Gab than with Twitter, despite Gab having fewer unique narratives than Twitter. This indicates a higher similarity of the conversations happening on the Alt-Tech platforms compared to the mainstream platform. The low overlap in narratives suggests that themes frequently appearing together differ substantially on the three platforms. To understand what are frequently co-occurring agents and patients, and how platforms differ in this regard, we model Agents and Patients as nodes in an undirected network. This network is agnostic to what verb connects the two but allows us to look at the most popular Agents and Patients on platforms, and which terms tend to appear together in a narrative. We present giant connected components for the 30 most frequent Agent-Patient pairs in Figure 9 . The three platforms differ substantially when it comes to what is central to the conversation -conversations on Twitter (Fig. 9a ) are highly centred on QAnon, which is expected given how we collected the data. Terms connected to QAnon represent political entities -such as media, government, Donald Trump, and public personalities (mainly overlapping with those observed in RQ3). In addition, we note the mention of Capitol, and that Twitter users linked it to QAnon. In stark contrast, QAnon does not even appear in the top 30 connections on Parler (Fig. 9b) . This is yet another indication of different use of hashtags on Parler, where #QAnon was frequently used as part of hashtag blocks, rather than in the conversation context. Donald Trump and other public personalities take central roles of conversation on Parler, and the core is well-connected, except for "Location", "Book", and "One" (representing a person) making an appearance due to the high popularity of quote "the man who reads nothing at all is better educated than the man who reads nothing but newspapers" amongst Parler users. Gab network (Fig. 9c) is more similar to Parler's, although public personalities are more connected than Trump himself, with QAnon making an appearance in relation to lockdowns, public personalities, and popular phrase used by Trump supporters "Drain the swamp". Finally, we present the ten most prevalent narratives on all three platforms in Table 11 . These results provide some context in which Agent-Patient pairs presented in Figure 9 appear together, although we note that even the most popular narratives are not used many times (some of the narratives in the top 10 appear as few as 5 times), a result of the high diversity of verbs used. QAnon is central to the discussion on Twitter, appearing in 6 top narratives, which are indicative of republicans' and Trump's relationship with QAnon, and the role that Twitter users believe QAnon had in the storming of the Capitol. On the other hand, Parler mostly sees the popularity of quotes (such as the example above) and calls to action by other Trump supporters -asking them to email him or follow and re-post. Gab has the highest level of conspiratorial ideas breaking into top 10 narratives -CIA infiltrating local states' governments, suggesting that nobody can prove QAnon is fake, and calling on Trump to invoke the Insurrec- tion Act in response to the "election being stolen". While the three platforms proved to be vastly similar when tested for anti-social language, prevalent topics of conversation, and which groups of political figures attract more anti-social language, narrative analysis reveals considerable differences in the context of mentions of these figures and how central QAnon itself is to the discussion. Twitter centres the conversation on QAnon, often mentioning it in a negative context; Parler focuses on expressing support for Trump and making calls to action towards like-minded users, while Gab narratives hint at a higher prominence and more central role of conspiratorial ideas. We have compared discourse around #QAnon on three platforms -Twitter, Gab, and Parler -in a month preceding the storming of the US Capitol on January 6, 2021., using a variety of computational methods to capture different aspects of posts -from measures of anti-social language, to themes at the core of the discourse. Our findings show that the volumes of posting with this hashtag differ drastically across platforms, with Parler having the highest volume of data. While more unique users have posted using #QAnon on Twitter, users on Parler and Gab made considerably more posts per user on average. We found that the prevalence of anti-social language on the three platforms varies depending on the measure. While Twitter and Parler emerge as leaders in terms of the distribution of posts with anti-social language based on the analysed Perspective API features, Gab has the highest proportion of posts with hate words. The analysis of the most frequently mentioned named entities across the three platforms revealed important similarities and differences between them. On the one hand, there are obvious overlaps in the most popular named entities across the platforms, suggesting that #QAnon-related discourse mentioned largely similar themes during the observation period. Such overlapping entities include Donald Trump, the US, and Washington, D.C. Entities enjoying higher popularity on Twitter compared to the other two platforms include individuals salient in the US political context in connection to QAnon (e.g., Marjorie Taylor Greene or Jake Angeli) as well as "Antifa" and "Nazi", suggesting divisive stance in relation to #QAnon. On Gab and Parler, popular entities suggest a focus on discussions around the election results and related conspiracies (e.g., Dominion voting machines or Deep State) and the right to bear arms. We observe differences in anti-social language measures in posts mentioning different political groups or individuals. Though the prevalence of such posts differs across platforms, as well as the prevalence of anti-social language in posts about different groups of politicians, we find that on all platforms posts mentioning female politicians, Democrats, and Donald Trump score higher on anti-social features than those with mentions of their male or Republican counterparts, or Joe Biden. Since our analysis focused on a very specific period and topic, it is unclear whether this observation can be generalised. We suggest that it is worthwhile to examine the prevalence of anti-social language in posts about different political groups in further cross-platform research. Finally, our analysis of which terms appear together in posts, and the narratives they appear as part of, indicates that the core of discussions related to #QAnon differs substantially across the three platforms. While Twitter focuses on QAnon and the most prevalent narratives related to it are critical of it, Parler focuses on supporting Donald Trump, while Gab sees a bigger focus on other political figures, as well as Trump, and has a higher prevalence of conspiratorial content amongst its most popular narratives. Our study has limitations in addition to the ones listed in the Data section. Firstly, we focused on a single hashtag that, despite being one of the most popular in relation to QAnon conspiracy theory on all three platforms and one of the most popular hashtags in general on Parler, still reflects overall discourse surrounding QAnon only partially. Nonetheless, due to the prominence of this hashtag, we argue that the collected data reflects the most dominant aspects of QAnon-related discourse. Secondly, while the monthlong time frame our study encompasses is appropriate for analysis of events that see a heightened volume of activity, a further study examining the development of QAnon-related discourse over time could be relevant, comparing the observations during the periods of high activity, as well as in more routine periods. There are three implications of our findings that we wish to highlight in particular. Firstly, while the differences between the three platforms exist in our sample, they do not exactly align with what public opinion appears to be. Given the press Parler has received, and the consequences it has suffered in part related to the anti-social language, one might expect that Parler would exhibit a much higher prevalence of threatening, or toxic language than, for example, Twitter. Yet, our results show that platforms are largely comparable in this regard, with no clear "villain", and that platforms, including mainstream ones, can go further in limiting particularly harmful aspects of anti-social language. This also highlights the need for cross-platform comparative studies. The research community is uniquely equipped with methods, independence, and thoroughness required to make fair judgements to inform public perceptions, which might be currently largely based on preliminary analyses. Secondly, we show that more nuanced methods are necessary for comparative cross-platform studies, as single measures or simple metrics often do not reveal the differences, especially when observing multifaceted phenomena such as language and discourse. Finally, as we have seen with hashtags, seemingly identical functionalities can be used differently across different platforms, and pre-processing choices we make can substantially affect the conclusions of the analysis. This highlights the importance of the decisions we make at each step of our analyses, including pre-processing, as well as how transparently we communicate them in our studies, both in cross-platform research, and beyond. An Early Look at the Parler Online Social Network The gospel according to Q: Understanding the QAnon conspiracy from the perspective of canonical information Text Semantics Capture Political and Economic Narratives Sheryl Sandberg Downplayed Facebook's Role In The Capitol Hill Siege-Justice Department Files Tell A Very Different Story On Twitter Purge: A Retrospective Analysis of Suspended Users Repeat Offenders: Voting Misinformation on Twitter in the 2020 United States Election Hate lingo: A target-based linguistic analysis of hate speech in social media Characterizing social media manipulation in the 2020 U.S. presidential election The storming of Capitol Hill was organized on social media. The New York Times Perspective API Attributes and languages Inside the rightleaning echo chambers: Characterizing gab, an unmoderated social system From "welcome new gabbers" to the pittsburgh synagogue shooting: The evolution of gab The effect of extremist violence on hateful speech online The spread of COVID-19 conspiracy theories on social media and the effect of content moderation Toxicity Detection: Does Context Really Matter? Stanza: A Python Natural Language Processing Toolkit for Many Human Languages Politicians in the line of fire: Incivility and the treatment of women on social media The risk of racial bias in hate speech detection Trumping Hate on Twitter? Online Hate Speech in the 2016 US Election Campaign and its Aftermath Analyzing the targets of hate in online social media 2020. garc: Python and Command-Line Interface for Gab Go eat a bat, Chang!": On the Emergence of Sinophobic Behavior on Web Communities in the Face of COVID-19 Contours and Controversies of Parler Pro-Trump forums erupt with violent threats ahead of Wednesday's rally against the 2020 election Twitter Academic Research API US Department of Justice. 2021. Three Men Charged in Connection with Events at U.S. Capitol A Multi-Platform Analysis of Political News Discussion and Sharing on Web Communities Rightwingers flock to 'alt tech' networks as mainstream sites ban Trump. The Guardian What is gab: A bastion of free speech or an alt-right echo chamber Measuring and characterizing hate speech on news websites A quantitative approach to understanding online antisemitism Conceptualizing "Dark Platforms QAnon and the Emergence of the Unreal