key: cord-0655165-nqnpohu8 authors: Waissbluth, Elliott; Farid, Hany; Sehgal, Vibhor; Peshin, Ankit; Afroz, Sadia title: Domain-Level Detection and Disruption of Disinformation date: 2022-05-06 journal: nan DOI: nan sha: 88504ea9200eabd5deac5ccbee6c16adb31c791a doc_id: 655165 cord_uid: nqnpohu8 How, in 20 short years, did we go from the promise of the internet to democratize access to knowledge and make the world more understanding and enlightened, to the litany of daily horrors that is today's internet? We are awash in disinformation consisting of lies, conspiracies, and general nonsense, all with real-world implications ranging from horrific humans rights violations to threats to our democracy and global public health. Although the internet is vast, the peddlers of disinformation appear to be more localized. To this end, we describe a domain-level analysis for predicting if a domain is complicit in distributing or amplifying disinformation. This process analyzes the underlying domain content and the hyperlinking connectivity between domains to predict if a domain is peddling in disinformation. These basic insights extend to an analysis of disinformation on Telegram and Twitter. From these insights, we propose that search engines and social-media recommendation algorithms can systematically discover and demote the worst disinformation offenders, returning some trust and sanity to our online communities. In the midst of the global pandemic, 20% of the public believed Bill Gates was planning to use COVID-19 to implement a mandatory vaccine program with tracking microchips [28] . Around the same time in mid-2020, the QAnon conspiracy ripped through social media, contending that a cabal of Satan-worshipping cannibalistic pedophiles and child sex-traffickers plotted against Donald Trump during his term as US President. A recent poll found 37% of Americans are unsure whether the far-reaching QAnon is true or false, and 17% believe it to be true [27] . The common thread in Bill Gates' COVID-microchips, QAnon's Satan-worshipping cannibals, and the long litany of conspiracies, lies, and general nonsense polluting the internet is the ease with which billions of online users can create and distribute content, the favoring by social-media's recommendation algorithms of the most outrageous and salacious content that drives user engagement, and the seemingly endless appetite of the general public for this content. Today's disinformation campaigns are leading to real-world harms from vaccine hesitancy to the denial of the scientific consensus of the catastrophic effects of global climate change, denial of human-rights violations in Ukraine, and a lack of confidence in our electoral system. Tackling disinformation on a per-post/image/video basis (e.g., [1-3, 10, 11, 14, 16, 18, 21-25, 32, 33, 38-43, 47] ) is leading to a maddeningly massive game of online whack-a-mole. At the same time, recent studies have found that a relatively small number of users are responsible for the majority of online COVID [34] and climate change [30] disinformation, suggesting that tackling disinformation at a higher-level based on specific users, groups, or domains may be more effective. By way of nomenclature, we refer to "disinfo domains" as those consisting of a broad category of domains that traffic in conspiracies, distortions, lies, disinformation, and more generally do not follow accepted journalistic standards, whether they are maintained by a state-sponsored actor, a private or public entity, or an individual. All other domains will be referred to as "info domains. " We describe in more detail in Section 3.1 how domains are characterized as one or the other. In this work, we describe a domain-level predictor of disinformation peddlers that classifies an entire domain (e.g., www.rt.com) as an unreliable news source. This classifier relies on three signals: (1) hyperlinks (similar to [35] ); (2) meta tags specified by the domain owners and used primarily for search-engine optimization; and (3) content in the form of the visible text on the landing and internally linked pages. Our primary contributions include: (1) Analyzing a large set of 2500 domains identified as trafficking in disinformation. (2) Revealing a distinct and predictive patterns of hyperlinking, meta tag construction, and underlying content among disinformation peddlers in the form of accurate classifiers (F1=94.1%) able to distinguish between disinfo and info domains. (3) Building a hyperlink graph to reveal coordinated disinformation efforts. (4) Extending our analysis to Telegram to discover the most prolific channels and users trafficking in disinformation, and using these identified channels to expand our disinfo domain data set. (5) Extending our analysis to Twitter to discover the most prolific accounts trafficking in disinformation, and using these identified channels to expand our disinfo domain data set. According to a 2020 poll, slightly more than half of Americans rely on social media for at least some of their news [37] . Given the growing flood of misinformation, lies, and conspiracies found on social media, recent efforts have focused on understanding the promotion and spread of misinformation on social media. Vosoughi et al. [44] , for example, analyzed the spread of misinformation on Twitter and found that misinformation spreads faster than the truth, and that misinformation is more novel than the truth and is designed to inspire a strong response of fear, disgust, and surprise, and hence more engagement in terms of likes, share, and retweets. Because of this, social media's recommendation algorithm tend to favor misinformation. Faddoul et al. [9] and Tang et al. [42] , for example, showed that YouTube's recommendation algorithm contribute to the spreading of conspiracies and misinformation. And, automated tools, such as Hoaxy [36] , reveal in real time how misinformation spreads on Twitter. Related efforts have also analyzed the infrastructure support behind disinformation. Han et al. [12] , for example, studied the service providers that power hundreds of disinformation and hate sites and found that disinformation sites disproportionately rely on several popular ad networks and payment processors, including RevContent and Google DoubleClick. In related work, by analyzing ads placed on over a thousand disinformation domains, Zeng et al. [46] discovered that 42.5% of the ads on these domains are political ads with a significant portion (26.0%) being left-leaning. Designed to inspire a strong response, the distinct linguistic characteristics of text-based misinformation have been used for automatic detection. A plethora of techniques have been proposed to automatically classify individual posts as misinformation or not. These range from leveraging classic machine learning techniques (SVM, LR, Decision Tree, Naive Bayes, k-NN) to more modern machine learning (CNN, LSTM, Bi-LSTM, C-LSTM, HAN, Conv-HAN) models to automatically detect misinformation (see [19] for a comparative study of detection approaches). Having a large and accurately labeled list of misinformation, however, is difficult to obtain, which is why most of these studies use relatively small datasets. Retroactive fact checking, however, is unlikely to stem the flow of misinformation. It has been shown, for example, that the effect of misinformation may persist even after false claims have been debunked [5, 20] . In contrast, we focus our analysis at a higher level of determining if specific domains (e.g., www.rt.com), Telegram channels, or Twitter users, are complicit in the spreading of misinformation. Detection at this level lends itself to earlier interventions in the form of demotion, demonetizing, and -in the most extreme cases -de-platforming. We analyze the content of and the interconnection (through hyperlinking) of 2500 disinfo domains -identified by independent and nonpartisan sources -to build a domain-level classifier of disinformation peddlers. This type of hyperlink analysis has previously been examined. By analyzing 89 news outlets, for example, Pak et al. [29] found that partisan media outlets are more likely to link to nonpartisan media, but that liberal media link to liberal and neutral outlets whereas conservative media link more exclusively to conservative outlets. In analyzing hyperlinks between news media between 1999 to 2006, Weber et al. [45] found that establishing hyperlinks with other, younger news outlets strengthens the position of that organization in the network thus boosting traffic. Hanley et al. [13] observed a small-world phenomenon in QAnon domains in which only 22% of authentic news websites hyperlink to a QAnon sites, as compared to 40% of known disinformation domains. While these previous studies were not focused broadly on disinformation, the recent work of [35] analyzed patterns of hyperlinking to classify domains as trustworthy or not. One crucial observation from this work is that disinformation domains are heavily connected to one another and loosely connected to authentic news domains. We expand on this work by analyzing more than twice as many domains, collated -unlike in this earlier work -from objective sources using standard metrics of journalistic standards. We also, unlike this earlier work, look at both the underlying HTMLbased content as well as the hyperlinking connectivity. And, we expand our analysis of domain-level disinformation to Telegram and Twitter. We obtained two commercially available, manually annotated disinformation domain data sets. The first, from the Global Disinformation Index (GDI), is a monthly-updated list of global domains determined, using established journalistic standards, as being purveyors of disinformation. Our list, from Oct 2021, contains 1276 disinfo domains. The second, from NewsGuard, is also a monthly-updated list of global domains rated and reviewed by trained journalists. Our list, from Oct 2021, contains 3403 disinfo domains. Each site in this list is rated using nine basic, apolitical criteria of journalistic practice, from which each domain is scored on a scale of 0 (generally not trustworthy) to 100 (generally trustworthy). We labeled any site with NewsGuard's criteria that a score below 60 is classified as a disinfo domain. The top 10, 000 Alexa-rated sites -sans domains in our disinfo data set -constitute our informational domains. This info data set is, of course, a relatively crude comparison data set, but was selected because of their popularity and therefore overall online impact. For each domain in our dataset, OpenWPM 1 was used to scrape the contents of the top-level domain (level 1), and to scrape internal pages linked by the top-level domain (level 2), etc., up to a maximum of 100 pages. Any domain that returned a 404 error were excluded from our analysis, yielding 2499/3689 disinfo domains and 7888/10000 info domains. These domains were further narrowed because of scraping blocks or primarily non-English content, yielding a final domain count of 2435 disinfo domains and 5396 info domains. The HTML hyperlink tag () is used to link to an internal or external page. From the domains described above, we extracted all hyperlinks to external pages. An unweighted, directed graph of hyperlinks was then constructed in which the graph nodes are the top-level domains and a directed edge connects one domain that hyperlinked to another. The tldextract library 2 was used to extract the top-level domain name from the URL. For example, if rt.com hyperlinks to www.inforwars.com/posts/<...>, then the graph contains a directed edge from rt.com to infowars.com. Each domain is then featurized into a 3-D vector consisting of the following ratios: where # incoming disinfo is the sum of all unique disinfo domains linking to a specific domain (regardless of how many individual pages are linked), # outgoing disinfo is the sum of all unique disinfo domains a specific domain links to, and the total incoming count is only measured from our combined info and disinfo domains. Any feature for which the denominator is 0 is reassigned a value of −0.5. Meta tags are text-based HTML elements present in a web page's source code describing the page's content. These tags are mostly invisible to the user, as their purpose is to guide search engines. When a web architect creates a website, they take on the task of summarizing its content in meta tags to situate it among other relevant pages. Well crafted meta tags lead to better search engine optimization and user click through, motivating the architect to consider their meta tags carefully. Because of its importance in surfacing a web page, it is reasonable to assume that a peddler of disinformation will pay careful attention to this aspect of their page. From the the set of up to 100 pages per domain, we scraped the content associated with seven meta tags: (1) keywords, (2) description, (3) og:title, (4) og:keywords, (5) og:description, (6) twitter:description, and (7) twitter:title. These meta tags were selected because they tend to be descriptive of the underlying domain content, and because, unlike some other meta tags, were more likely to be specified in our domains (61.6% the domains in our dataset specified keywords, 92.9% specified description, 86.1% specified og:title, 0.3% specified og:keywords, 84.3% specified og:description, 59.1% specified twitter:description, 60.3% specified twitter:title, and 96.9% of domains specified at least one of these meta tags). Each scraped domain yields a single string of concatenated words pre-processed to remove capitalization, punctuation, numbers, and stop words (e.g., a, the, is, are, etc.), and then stemmed using the Porter Stemmer algorithm [31] , and lemmatized using the Word-Net Lemmatizer [26] (both of which are are implemented in the Python Natural Language Toolkit 3 ). In order to reduce featurevector dimensionality and focus on discriminatory features, any under-represented words (present in fewer than 10% of our domains) or over-represented words (present in more than 90% of our domains) are eliminated from consideration. From the pre-processed meta-tag words, a unigram bag of words (BOW) [15] is weighted with each word's inverse document frequency [17] to yield a 12244-D meta-tag feature vector for each domain. As described in Section 3.5, this feature vector is then pruned down to the top 500 most diagnostic words. We define content as any visible text on a web page. This includes headers, footers, article titles, article content, link texts, button texts, etc. A website's visible text is likely to contain the majority of the linguistic information a user might engage with upon visiting a site. From the the set of up to 100 pages per domain, the visible text is extracted from the raw HTML by parsing out all text wrapped in any HTML tag, except for head, meta, title, script, and style. The extracted content from all scraped pages within a domain is then concatenated into a single string and, as described in Section 3.3, pre-processed to remove capitalization, punctuation, numbers, and stop words, then stemmed, and lemmatized. As before, any underrepresented tokens or over-represented tokens are eliminated from consideration. From the pre-processed words, a unigram bag of words (BOW) is weighted with each token's inverse document frequency to yield the final 80386-D content-based feature of each domain. As described in Section 3.5, and as with the meta-tag features, this feature vector is then pruned down to the top 500 most diagnostic words. Our classification objective is to distinguish between info and disinfo domains based on the domain-specific meta tag, content, and hyperlink features described in the previous sections, as well as from a combination of all three features. For the meta tag and content-based classifiers, we begin by reducing the original feature dimensionality of 12244-D and 80386-D to 500-D. This is accomplished by training a logistic-regression classifier (using Python's scikit-learn library) and extracting the top 500 most predictive distinguishing words from, separately, the original meta tag and content bag-of-words. For each feature, a separate linear support vector machine (SVM) is then trained (using Python's scikit-learn library) on a random 90% of the info and disinfo domains, and tested on the remaining 10% of the domains. A single combined classifier is also trained on all three features. Each component of the 500-D meta-tag and content features are specified in the range [0, 1]. Each component of the 3-D hyperlink feature is normalized into the range [0, 1] (by subtracting the minimum possible value of −0.5 [see Section 3.2] and dividing by the maximum value of 10.5). The SVM is optimized using a 5-fold cross validation grid search across kernel functions, regularization parameter , and 1 and 2 penalty functions, yielding a classifier with a linear kernel, an 2 penalty function, and = 29.76 for the 500-D text-based features, and = 10000 for the 3-D hyperlinking features. Shown in Table 1 is the testing SVM (Section 3.5) performance averaged over 100 random 90%/10% dataset training/testing splits using only the 500-D representation of a domain's meta tags (Section 3.3). With an average accuracy of 94.4%, and a minimum and maximum accuracy of 93.9% and 94.9%, it would appear the meta tags are a reasonably good predictor of domain trustworthiness. With an average (min/max) precision -ratio of correctly classified disinfo domains to all classified disinfo domains -of 96.1% (95.3%/97.0%) and an average recall -ratio of correctly classified disinfo domains to the total number of disinfo domains -of 85.5% (84.0%/87.0%), the classifier is somewhat biased. We hypothesize that this bias is due to more diversity in the info domains, as compared to the disinfo domains. The top-ten most predictive (stemmed and lemmatized) words for classifying a domain as disinfo are: prank, archiv, conserv, gadget, ha, trump, freedom, polit, forb, russian where archiv is the prefix for such words as archive, archives, archiving, ha is a prefix for, among others, ha-ha, polit is the prefix for such words as politic, political, politician, and forb is a prefix for, among others, forbes. One benefit of the somewhat simplistic logistic regression classifier is it is interpretable allowing us to see that the classifier seems to be picking up on words that are intuitively related to disinformation. Shown in Table 1 is the testing SVM (Section 3.5) performance averaged over 100 random 90%/10% dataset training/testing splits using only the 500-D representation of a domain's content (Section 3.4). With an average accuracy of 95.7%, and a minimum and maximum accuracy of 95.3% and 96.1%, the content is a slightly better predictor of domain trustworthiness than the meta tags. It is a little surprising that this classifier is only slightly more accurate than the classifier using only the meta tags (95.7% vs. 94.4%). We posit this is because the meta tags -designed to drive search engine optimization -consist of the most descriptive parts of a webpage's content. With an average (min/max) precision of 96.1% (95.3%/96.9%) and an average recall of 89.8% (88.7%/90.9%), the classifier is slightly less biased than the meta tag classifier. The top-ten most predictive (stemmed and lemmatized) words for classifying a domain as disinfo are: prank, donat, petit, drudg, trunew, wordpress, immigr, post, freedom, would where donat is the prefix for such words as donation, donate, donating, petit is the prefix for such words as petition, petitioning, petitioned, immigr is the prefix for such words as immigration, immigrant, immigrating. The popular blogging site wordpress.com often embeds the text "Powered by WordPress" in individual posts, and this domain also allows users to create profiles on other sites, tagging any comments on those sites with "WordPress Profile." These are the likely sources of the predictive wordpress. Shown in Figure 1 is the hyperlinking adjacency matrix for the 7831-node graph consisting of 5396 info domains and 2435 disinfo domains (Section 3.1). There is a connection (blue dot) from one domain (horizontal axis) to another domain (vertical axis) if the from-domain contains at least one hyperlink to the to-domain. In this visualization, we can see that there are many hyperlinks from disinfo domains to other disinfo domains (upper right corner of the adjacency matrix). At the same time, info domains infrequently link to disinfo domains (upper left), and both info an disinfo domains link to info domains with similar frequency (lower left, and lower right corners of the adjacency matrix). The info domains on the horizontal axis are ranked -from bottom to top -in order of decreasing popularity (as determined by Alexa rankings), hence the gradient in hyperlinks with more links evident to the more popular domains. Shown in Table 1 is the testing SVM (Section 3.5) performance averaged over 100 random 90%/10% dataset training/testing splits using only the 3-D representation of a domain's hyperlinking patterns (Section 3.2). With an average accuracy of 95.1%, and a minimum and maximum accuracy of 94.7% and 95.5%, the hyperlinks are a comparable predictor of domain trustworthiness as compared to 4 Figure 2 : Shown is the subgraph induced by the domain (www.freebeacon.com) with the highest in-degree. The top-ranked domain is rendered with a yellow ring around the red node symbol. The small, fully-connected clusters correspond to imposter sites designed to look like local news sources (see Figure 3 for a description of three of these clusters labeled (a)-(c)). the meta tags (94.4%) and content (95.7%). The advantage, however, of this hyperlinking feature is that it is language agnostic and, as we will see below (Section 5.2 and 5.3) seems to generalize to other services like Telegram and Twitter. With an average (min/max) precision of 97.4% (96.8%/98.0%) and an average recall of 86.6% (85.3%/87.9%), the classifier, as with the meta tags and content, remains biased. Shown in Table 1 is the testing SVM performance averaged over 100 random 90%/10% dataset training/testing splits using the concatenated meta tag, content, and hyperlinking features as described in Section 3.5. With an average accuracy of 96.3%, and a minimum and maximum accuracy of 95.9% and 96.7%, the combined classifier slightly outperforms the individual classifiers. With an average (min/max) precision -ratio of correctly classified disinfo domains to all classified disinfo domains -of 94.5% (93.5%/95.4%) and an average recall -ratio of correctly classified disinfo domains to the total number of disinfo domains -of 93.7% (92.6%/94.8%), the classifier is, unlike the individual classifiers, unbiased. The disinfo domains with the highest in-degree (i.e., many domains link to these) are: (1) www.freebeacon.com (2) www.naturalnews.com "Moderately to strongly biased toward conservative causes through story selection and/or political affiliation. They may utilize strong loaded words (wording that attempts to influence an audience by using appeal to emotion or stereotypes), publish misleading reports, and omit information that may damage conservative causes. Some sources in this category may be untrustworthy. sources. Overall, we rate the Washington Free Beacon Right Biased based on story selection that favors the right and Mixed for factual reporting due to misleading and false claims. " A total of 735 disinfo domains hyperlink to this top-ranked domain (only 35 info domains, not shown in the subgraph, link to this top-ranked domain). This subgraph reveals a fascinating structure consisting of a large number of highly-connected sub-graphs, each of which hyperlinks to www.freebeacon.com (shown in the center with a yellow ring around the red node symbol). The small compact clusters each correspond to purportedly regional news sites. The seven, fully-connected domains in Figure 2 (a), for example, consist of the Connecticut-based sites 4 . Similarly, the seven domains in the fully-connected Figure 2 (b) consist of Oklahoma-based sites 5 , and the eight fully-connected domains in Figure 2 (c) correspond to Alabama-based sites 6 . Shown in Figure 3 (a)-(c) are thumbnails of these regional domains, revealing a similar layout, ads, and stories within and across the clusters. According to NewsGuard, all of these domains are controlled by Metric Media News which Media Bias/Fact check describes as: "Overall, we rate Metric Media LLC right-center biased and Questionable based on a lack of transparency, the publication of false information, and nondisclosure of over 1000 imposter websites that are designed to look like local news sources. " Except for the larger cluster at the bottom of the graph in Figure 2 , each of the other tightly-coupled clusters belong to a local news sources controlled by Metric Media News. Our analysis is capable not only of predicting if a domain is a likely peddler of disinformation, but also to reveal these types of sub-structures in the disinfo ecosystem. Of particular interest are these types of highly linked-to domains, but also the fully connected cliques revealing potentially coordinated disinformation efforts. The nodes are color-coded according to their general category, revealing a striking relationship between Russian-based sites, QAnon, and US-based far-right groups. Because Metric Media's imposter sites are so numerous, accounting for 36.1% of our disinfo domains, we worried that their dominance skewed our classifier (Table 1) . We, therefore, retrained our classifier using only one domain from Metric Media's vast network, and only one domain from any other identified (by NewsGuard or GDI) similar type of network. This yielded a total of 1061 disinfo domains and a classification accuracy of 94.3% (as compared to the previous 96.3% accuracy), but a significantly reduced F1 score of 81.0% (as compared to 94.5%). Although the precision remained relatively high at 89.1% (as compared to 93.7%), the recall reduced significantly to 74.4% (as compared to 94.1%). We hypothesize that this reduction in overall performance is due to a significantly smaller and more diverse dataset. We expect, however, that as we grow and diversify our disinfo data set, classification will further improve. Disinformation, of course, travels in many different forms and circles, from imposter news sites to social-media posts, public and private groups, and individual text messages. We wondered if our domain-level hyperlink analysis would generalize to some of these other mediums. We next describe a hyperlink analysis of disinformation networks on public Telegram channels. The telegram channel Donbass Insider has previously been implicated in advancing pro-Russian propaganda 7 Given its popularity and impact on the current disinformation landscape, we began our analysis with this channel and recursively scraped all channels discovered through the forwarded messages feature. The open-source Python library Telethon 8 was used to scrape these channels. As with the domain-level graph in Figure 2 , a graph was constructed in which each node is a telegram channel, and a directed edge connects node A to node B if channel A forwarded a message to channel B. This recursive scraping of 1, 802 channels yielded a total of 11, 127 channels. From these channels, a total of 1, 013, 034 messages were collected from which 147, 584 contained a shared domain URL, 834 of which were contained in our disinfo data set. To understand who shares misinformation on Telegram and their relationship to each other, the graph was partitioned -using the Louvain method [4] into communities based on channel-to-channel connectivity. This process yielded two broad communities shown in Figure 4 : (1) Russian, constituting 59.8% of all channels; and (2) the US Far Right, constituting 40.2% of all channels. The Russian community are further composed of two communities consisting 7 The perceived threat of this and related channels led to Telegram co-founder Nikolai Durov to issues the warning "Telegram channels are increasingly becoming a source of unverified information related to the Ukrainian events. We do not have the physical ability to check all the publications of the channels for reliability. I urge users from Russia and Ukraine to doubt any data that is distributed in Telegram at this time. We do not want Telegram to be used as a tool to exacerbate conflicts and incite ethnic hatred. In the event of an escalation of the situation, we will consider the possibility of partially or completely limiting the work of Telegram channels in the countries involved for the duration of the conflict. " Shortly after posting this message on Telegram, however, Durov, walked back his threat to limit access to channels. 8 https://pypi.org/project/Telethon of Russian media channels and Russian influencers. The US Far-Right community contains notable QAnon followers such as Sidney Powell (best known as an avid conspiracy theorist and for her part in trying to overturn the 2020 US presidential election) and QAnon enthusiast Rep. Marjorie Taylor Green. We find that the channel SidneyPowell is the most prolific sharer of content found in our disinfo data set. A total of 61 new domains were scraped from 1, 000 recent posts from this Telegram channel. Of these 61 domains, 21 domains not in our original data set are classified by our amalgamated classifier (Section 4.4) as disinfo, and manually verified by us to be clear disinfo domains. These domains include: independentside.com, dailynewsbreak.org, and freeworldnews.tv (Media Bias/Fact rates this cite as "Extreme Right Biased and a Tin-Foil Hat Conspiracy and Quackery level Pseudoscience source based on the routine publication of disinformation and outright fake news." One site, madison.com, was incorrectly flagged as disinfo. After the SidneyPowell channel, the next four most prolific disinfo channels are worlddoctorsalliance, RealGenFlynn, RealEMichaelJones, and DrLynnFynn. By searching public Telegram channels for identified disinfo domains, and then performing a connectivity analysis similar to our earlier hyperlink analysis, we can identify prolific pushers of disinformation (and their relationship to each other) not just at the domain level, but also at the Telegram channel level. This analysis also allows us to grow our disinfo domain list by collating and classifying previously unseen domains from these problematic channels. Private channels, of course, would not be amenable to this type of analysis. We next investigate the ability to generalize our methods to identify disinfo domains by tracking the hyperlinks shared by certain socialmedia users. Because of the relative ease of access, we focus on Twitter's publicly available user data using two of Twitter's APIs (using Python's Tweepy Library 9 ). First, the Search Tweets API 10 allows for curating a set of recent tweets (up to 9 days old), which can be filtered by the keywords, hashtags or the URLs they contain. Starting with 2500 NewsGuard disinfo domains (this is an updated version of the database as compared to that used in the main analysis), we collected up to a maximum of 1000 tweets per domain, surfacing which Twitter users were sharing these disinfo domains. Second, the Get Tweet Timelines API 11 allows for curating up to 3200 most recent tweets posted by a queried Twitter user. Using this API, we collect the recent domains shared by users surfaced in the first step. The data returned by both APIs contains additional user attributes such as geo-location and tweet attributes such as timestamp and replied-to, that could be leveraged in the future. We found 569 of the 2500 disinfo domains had been shared at least once on Twitter. An undirected domain-level graph was constructed ( Figure 5 ) in which each node is a disinfo domain, and an undirected edge connects node to node if atleast 3% of the users sharing either domain also share both domains (i.e., the Jaccard similarity ≥ 3%). The resulting graph contains 194 nodes and 211 edges. A community detection algorithm [6] was employed with the aim of finding if clusters of domains shared by the same users would reveal a common theme in the content being shared. This community analysis revealed some interesting clusters. Eight of the total 11 domains from Community A are part of the Natural News Network, included in the NewsGuard disinfo dataset, and whose articles have failed fact checks by fact-checking organizations and are described by Wikipedia as "fake news website" spreading pseudoscience and disinformation. Community B is dominated by health pseudoscience spreaders, of which naturalnewsblogs.com is also part of the Natural News Network. Community C contains some Chinese state-controlled media, which have been accused of spreading Russian disinformation in the Ukraine war. To find what other domains are shared by prolific disinformation spreaders, we scraped the Along with online video platforms like youtube.com and rumble.com, hyperlinks to particularly partisan news websites also feature prominently in their tweets. We should begin our discussion with an acknowledgement that the topic of disinformation can and has become highly political. This can make research into and proposals to mitigate disinformation highly volatile. The political right may accuse us of bias because right-leaning organizations may be more implicated in spreading election-related misinformation, and the political left may accuse us of bias because left-leaning organizations may be more implicated in spreading others forms of misinformation. We have, throughout, tried to remain politically neutral. From the very beginning of our research, we relied on established and independent organizations to seed our analysis in the form of the initial disinfo domains supplied by NewsGuard and GDI. Throughout our research and analysis, we then allowed the data to drive the results and conclusions. With some 500 hours of video uploaded to YouTube every minute, and over a billion posts to Facebook each day, the massive scale of social media makes tackling misinformation (the unintentional spreading of false or misleading information) an enormous challenge. Tackling disinformation (the intentional spreading of false or misleading information), however, may be more tractable. Previous studies, for example, have found that a relatively small number of users are responsible for the majority of COVID and climate-change misinformation [30, 34] . We propose that tackling disinformation at the domain-level similarly narrows the scope of problematic content to a more manageable number. In particular, we propose that search engines and recommendation algorithms can simply demote content that contains links to previously documented (and periodically reviewed) domains. Similarly, Telegram, Twitter, TikTok, etc. channels that predominantly share problematic content can be demoted, possibly demonetized, and only in extreme cases deplatformed. We understand and appreciate the need to balance an open and free internet, where ideas can be debated, with the need to protect individuals, societies, and democracies. Social media, however, cannot hide behind the facade they are creating a neutral marketplace of ideas where good and bad ideas compete equally. They do not. It is well established that social media's recommendation algorithms favor the divisive, outrageous, and conspiratorial because it increases engagement and profit [7] . As a result, Brandies' concept [8] that the best remedy for falsehoods is more speech, not less, simply doesn't apply in the era of algorithmic curation and amplification. By demoting disinformation peddlers, we can create a more fair and balanced marketplace of ideas where Brandeis' principle can, in fact, flourish. As with any inherently adversarial relationship, all approaches to addressing disinformation -including ours -will have to adapt to new and emerging threats. In our case, disinformation peddlers may add decoy hyperlinks to external trustworthy domains to escape being classified based on their hyperlinks to other disinformation domains. This, in turn, will require techniques to root out such decoy links. And so on, and on, and on. While such a cat and mouse game can be frustrating, the end game will be that it will become increasingly more difficult and time consuming to intentionally create and spread false and misleading information, with the eventual goal of discouraging most, leaving us to contend with the die-hard adversary. While this is not a complete success, it will mitigate the risk of disinformation and, hopefully, return some civility and trust to our online ecosystems. All of the code associated with this research is available in an anonymous repository: https://anonymous.4open.science/r/disinfo_detection-96BA/ README.md Protecting World Leaders Against Deep Fakes Detecting fake news in social media networks Detecting fake news with machine learning method Fast unfolding of communities in large networks Debunking: A meta-analysis of the psychological efficacy of messages countering misinformation Community detection via semi-synchronous label propagation algorithms Moral outrage in the digital age Free Speech and Its Discontents: The Rebellion Against General Propositions and the Danger of Discretion A Longitudinal Analysis of YouTube's Promotion of Conspiracy Videos Deep learning algorithms for detecting fake news in online text Deepfake video detection using recurrent neural networks On the Infrastructure Providers that Support Misinformation Websites No Calm in The Storm: Investigating QAnon Website Relationships A retrospective analysis of the fake news challenge stance detection task Distributional structure Identifying Disinformation Websites Using Infrastructure Features A statistical interpretation of term specificity and its application in retrieval Multisource multi-class fake news detection Gias Uddin, and Anindya Iqbal. 2021. A benchmark study of machine learning models for online fake news detection Beyond misinformation: Understanding and coping with the "post-truth" era Exposing deepfake videos by detecting face warping artifacts Celeb-df: A large-scale challenging dataset for deepfake forensics Early detection of fake news on social media through propagation path classification with recurrent and convolutional networks Detect Rumors in Microblog Posts Using Propagation Structure via Kernel Learning Exploiting visual artifacts to expose deepfakes and face manipulations WordNet: An electronic lexical database More Than 1 in 3 Americans believe a 'deep state' is working to undermine Trump Examining the global spread of COVID-19 misinformation Intermedia Reliance and Sustainability of Emergent Media: A Large-Scale Analysis of American News Outlets' External Linking Behaviors Climate misinformation on Facebook 'increasing substantially', study says An algorithm for suffix stripping Supervised learning for fake news detection CSI: A hybrid deep model for fake news detection Majority of Covid misinformation came from 12 people Sadia Afroz, and Hany Farid. 2021. Mutual Hyperlinking Among Misinformation Peddlers Anatomy of an online misinformation network Detecting fake news on social media Understanding user profiles on social media for fake news detection Beyond news contents: The role of social context for fake news detection The role of user profiles for fake news detection Down the Rabbit Hole" of Vaccine Misinformation on YouTube: Network Exposure Study Fake news detection in social networks via crowd signals The spread of true and false news online Newspapers and the long-term implications of hyperlinking Polls, clickbait, and commemorative $2 bills: problematic political advertising on news and media websites around the 2020 US elections Detecting fake news for reducing misinformation risks using analytics approaches The data, consisting of the disinfo domains, are licensed from NewsGuard and GDI and we therefore cannot make this data publicly available per the licensing agreement terms.