Skip to main content
Wiley Open Access Collection logoLink to Wiley Open Access Collection
. 2019 Sep 26;12(1):109–138. doi: 10.1002/poi3.223

Do Machines Replicate Humans? Toward a Unified Understanding of Radicalizing Content on the Open Social Web

Margeret Hall, Michael Logan, Gina S Ligon, Douglas C Derrick
PMCID: PMC10647225  PMID: 38023688

Abstract

The advent of the Internet inadvertently augmented the functioning and success of violent extremist organizations. Terrorist organizations like the Islamic State in Iraq and Syria (ISIS) use the Internet to project their message to a global audience. The majority of research and practice on web‐based terrorist propaganda uses human coders to classify content, raising serious concerns such as burnout, mental stress, and reliability of the coded data. More recently, technology platforms and researchers have started to examine the online content using automated classification procedures. However, there are questions about the robustness of automated procedures, given insufficient research comparing and contextualizing the difference between human and machine coding. This article compares output of three text analytics packages with that of human coders on a sample of one hundred nonindexed web pages associated with ISIS. We find that prevalent topics (e.g., holy war) are accurately detected by the three packages whereas nuanced concepts (Lone Wolf attacks) are generally missed. Our findings suggest that naïve approaches of standard applications do not approximate human understanding, and therefore consumption, of radicalizing content. Before radicalizing content can be automatically detected, we need a closer approximation to human understanding.

Keywords: social media, violent extremist organizations, LIWC, latent dirichlet allocation, n‐grams, counter terrorism

Introduction

The world is seemingly continually surprised by the advance of deviant content on social media and the open Internet (Ingram, 2016). This is particularly true in the case of violent extremist organizations (VEOs)—that is, coordinated efforts among individuals with a shared ideological framework and goal toward collective actions (Ligon, Simi, Harms, & Harris, 2013). The capacity for posting and hosting ever‐more disturbing, violent propaganda is matched only by the willingness of (vulnerable) individuals to consume it. Attempting to proactively detect (and remove) VEO content from open social platforms is a large‐scale social and research problem (UNCTED, 2016, Whine, 2010).

As VEO content appears online it swims in an ocean of other unrelated content, leaving the noise‐to‐signal ratio very high. To solve the detection problem, manual approaches like content coding, or labeling, have been standard. On one hand, manual approaches have proven to be inappropriate to the task of identifying and treating content at scale. Human understanding is the “gold standard” of classification, but with respect to VEO content it is a poor strategy due to aspects like coder burnout and potential errors and bias due to inexact classification definitions. Especially, with nuanced or latent VEO content there can be worldview conflicts (e.g., conflicts with personally held beliefs on free speech or religion) leading to stress in the coder, and few explicit, standardized definitions that apply to online material exist except in cases of depictions of crime (Facebook, 2019; Logan & Hall, 2019). Even with these potential downsides, manual coding is currently in use, or forms part of an artificial intelligence (AI) strategy at the major social content platforms and research labs (Alonso, 2019). On the other hand, even the most successful AI tools misclassify radicalizing content. For example, Facebook reported finding 99 percent of terror‐related content in Q1‐3 of 2018 (https://newsroom.fb.com/news/2018/11/staying‐ahead‐of‐terrorists/). This leaves an estimated one hundred forty‐three thousand terror‐related content items, which were unable to be classified. The length of time the identified items were available on the platform before being identified ranged from two minutes to seven hundred and seventy days in Q3 2018. Most misclassification comes from the advent of new modalities of content provision or from disagreement between human definitions and machine classification. These misclassifications suggest a potential systematic bias in the classifier in use.

Given its positive attributes, including increased efficiency and low researcher exposure, automated detection of VEO content online should be a broad research and societal goal (Correa & Sureka, 2013). However, a fundamental challenge is reducing the barriers to machine understanding (Alonso, 2019), given even in binary classification schemes some measure of disagreement exists (Hashemi & Hall, 2018, 2019). Approaches to classification tend to either be fully machine (e.g., Alghamdi & Selamat, 2012; Sabbah, Selamat, & Selamat, 2014; Scrivens, Davies, Frank, & Mei, 2016) or human (e.g., Droogan & Peattie, 2016; Ligon, Hall, & Braun, 2018; Torres, Jordán, & Horsburgh, 2006; Torres‐Soriano, 2016). Very few studies to date compare performance of humans and machines when classifying the underlying themes of radicalizing content. Even those with mixed modalities rarely show empirically the degree to which human coders and machines classifying the same radicalizing content have similar results (Cohen, Kruglanski, Gelfand, Webber, & Gunaratna, 2016; Correa & Sureka, 2013, Prentice, Taylor, Rayson, Hoskins, & O’Loughlin, 2011)—implying that we have an incomplete understanding of the degree of overlap of human and machine understandings of content and its classification. This uncertainty has an unknown impact on the way stakeholders manage their detection/classification regimes. Each of these differences contributes to/against external and internal validity and overall reliability of the analyses (Ruths & Pfeffer, 2014). Applications and algorithms to be deployed in the future are dependent upon some agreement of which content is radical: this does not currently exist.

We propose to help close that gap by considering the domain of text analysis. This study applies two probabilistic topic modeling algorithms and one off‐the‐shelf sentiment analysis package to an extant data set, namely 100 transient websites associated with the Islamic State (ISIS) (see the Methodology section). The coding themes in use are created by human coders evaluating ISIS content from these one hundred open websites (Derrick, Sporer, Church, & Ligon, 2017b). We introduce the themes established in (Derrick et al., 2017b) and refer to this throughout as the “benchmark study.” The current study identified the probabilistic topic models n‐grams (Jurafsky & Martin, 2014) and latent dirichlet allocation (LDA) (Blei, Ng, & Jordan, 2003) as the most appropriate, albeit simplistic, methodologies to assess the unlabeled corpus for a first attempt at comparing the performance of human and machines when coding VEO content. In addition, we assess the results of the Linguistic Inquiry and Word Count (LIWC) sentiment analysis program (Pennebaker, 2011). We chose these three packages due to their prevalence in related studies, their relatively strong benchmarks, and the ease of use for noncomputational scholars. The aim of this evaluation is assessing the degree to which computer‐mediated coding and human coders agree on the underlying themes of a radical corpus. The thematic view in use reflects the reality that different propaganda types exist, and that these differently impact individuals’ intent to act offline (Crosset, Tanner, & Campana, 2019).

It is important to identify the degree to which a gap exists between humans and machines due to the nature of the content. The imperative from a platform perspective reflects that VEO content generally violates their Terms of Service. More important is the human dimension: individuals consuming VEO content on the open social web may be vulnerable to its messaging. Pathways to radicalizing are many (Ligon et al., 2018); appropriately diagnosing radicalizing content before it can impact an individual is therefore critical. Finding that machines and humans are reasonably similar in their understanding of the themes that exist in radicalizing content would support a stronger content management regime overall.

In the next section we present and classify the relevant literature, showing that the question of performance of automated algorithms compared with human‐coded benchmarks in VEO content is under‐addressed in the literature. We present our research question, and then introduce the experimental design, data, and methodology. We then evaluate the results, and discuss their implications. We conclude with suggestions for expanding the research, and note some limitations to our approach.

Usage Patterns of VEOs Online

The promulgation of open‐ and free‐Internet architectures requires less technical infrastructure for smaller or resource‐poor organizations to communicate and conduct operations (Derrick, Ligon, Harms, & Mahoney, 2017a). Asymmetric organizations such as VEOs and criminal groups likewise expanded their reach and capacity for operations. VEOs have narrowed the infrastructure gap with larger, more affluent adversaries who have the means to develop expensive infrastructure for internal communication (e.g., the US Government invests in internal communication portals for secure communication). VEOs leverage existing (Internet) technologies to conduct their operations, recruit members, solicit financing, and facilitate strategic objectives during conflict (Denning, 2010; Weimann, 2004). The lowering of technological barriers has changed organizations’ behaviors on the short‐end of the power asymmetry and enabled the historically disadvantaged VEOs to launch social media campaigns akin to those of large governments and corporations (Balmer, 2001; Bartlett and Reynolds, 2015; Fenstermacher & Leventhal, 2011).

Given the web’s inherent anonymity, limited regulation, rapid flow of information, and large audience (Weimann, 2004), the Internet has prompted a shift in criminal and deviant behavior. In a review of how the Internet may amplify deviant communities online (Recupero, 2008), defined and examined the online disinhibition effect of fifty‐three groups including propedophilia, hate groups, proeating disorders groups, and terrorist groups. She found that problematic internet use by individuals interested in these groups increased dramatically. Relatedly, clandestine digital forums allow participants to express deviant opinions with little fear of negative social consequences, such as being discredited or stigmatized by mainstream elements (Steinfeldt et al., 2010).

Much research has been conducted on online behaviors of deviant groups, such as the far right and racially‐motivated groups (Caren, Jowers, & Gaby, 2012; Holt & Bolden, 2014; Steinfeldt et al., 2010), animal‐rights extremists (Braddock, 2015), gangs and gang members (Chen et al., 2004; Pyrooz, Decker, & Moule, 2013), proanorexia support groups (Boero & Pascoe, 2012; Fox, Ward, & O’Rourke, 2005), sex crimes and human trafficking (Dubrawski, Miller, Barnes, Boecking, & Kennedy, 2015; Elliott & Beech, 2009; Latonero, 2011; Nagpal, Miller, Boecking, & Dubrawski, 2017), and the drug trade (Fox, Ward, & O’Rourke, 2006; Martin, 2013). VEOs’ Internet usage is similar to other types of online criminal activities (Recupero, 2008) in that it is used to disseminate information and acquire value resources such as recruits, funds, and weapons. For example, Moule, Pyrooz, and Decker (2014) show that street gangs with a high degree of organization are more likely to utilize the Internet for recruitment and branding relative to other street gangs. It must be noted, however, that while some studies on deviant groups work in the automated detection space (e.g., Dubrawski et al., 2015), the characteristics of the signals from each type of deviant group are quite distinct—signals and detection benchmarks of child pornography, gangs, or supremacists do not necessarily transfer to jihadi content. The importance of our ability to quickly and reliably detect and classify radicalizing content became apparent in the aftermath of the 2019 Christchurch mosque shootings: a live feed of the attack from the assailant’s bodycam was taken down from all large social media platforms where it was hosted. The platforms, however, did not take down the feed before it was copied, sometimes recut, and reposted hundreds of thousands of times across platforms like Facebook, Reddit, and Twitter (https://www.washingtonpost.com/technology/2019/03/18/inside‐youtubes‐struggles‐shut‐down‐video‐new‐zealand‐shooting‐humans‐who‐outsmarted‐its‐systems/?utm_term = .bb581c44ea23). YouTube was forced to restrict search functionality as the reposting of the (sometimes remastered) video overwhelmed their algorithmic approach.

Recruiting

One of the most interesting aspects of VEOs is their use of information and communication technologies (ICTs) to recruit members and acquire skills. Recruiting is a central component of VEO leader decision making, resulting in a focus on branding, organizational legitimacy, and creating a compelling narrative.1 Decisions are both made and framed in relation to the brand, such as what alliances to endorse, what media to use in recruiting, and what statements to make by key figures. In the late 1990s and early 2000s, for example, members of the Chechen Mujahidin would film and distribute their armed attacks on the Russians on the Internet.2 While the Chechen militants were aware of the marginal effects of their attacks, they were also cognizant that hosting these attacks online would have greater effects in terms of mobilizing sympathizers of their cause and demoralizing the opposition (Torres et al., 2006). Additionally, VEOs utilize the Internet to establish online or virtual communities of sympathizers and supporters. Individuals who feel otherwise disconnected may gravitate toward virtual communities where they feel accepted. As such, virtual communities are a way for VEOs to foster a sense of in‐group membership and reach potential recruits in a way that could not be reached through traditional face‐to‐face recruitment (Radlauer, 2007; Simi & Futrell, 2006).

As previously noted, VEOs are strategic about what media they publish online. VEOs, like any organization, are concerned about establishing and maintaining their brand given its influence on their target audiences’ (friend and foe) perceptions about the group (Pelletier, Lundmark, Gardner, Ligon, & Kilinc, 2016). In turn, VEOs engage in innovative marketing strategies and craft specific types of media to draw potential recruits toward the movement. For instance, Al Qaeda and its regional affiliates (e.g., AQIM and AQAP) frequently craft messages that highlight the ongoing “battle of civilizations” and the struggle against the far enemy (United States). Al Qaeda deliberately used this narrative to project themselves as a global force, appealing to bored and alienated young males with the promise of excitement and adventure (Baines & O’Shaughnessy, 2014; Baines et al., 2010). Alongside violent narratives, VEOs such as ISIS have also capitalized on using non‐violent themes as a recruitment tool. During a month‐long period in 2015, Winter (2015) discovered that 52.57 percent of all online media outputs from ISIS emphasized utopian themes (also see Derrick et al., 2017b). These utopian themes focused on everyday life in the caliphate such as religious and social gatherings, economic activity, nature and wildlife, and governance. These themes are intended to promote ISIS and its caliphate as a legitimate state (Winter, 2015).

Command and Control and Decision Making

VEOs use digital communications for command and control (Whine, 2010). This is especially true for deterritorialized and/or decentralized organizations where the Internet works as a command and control center to coordinate and dispersed membership or attacks. For example, ISIS has been particularly adept at expanding its organizational reach through the use of encrypted messaging applications. Recent evidence indicates that ISIS remotely organized and instructed a cell in India to carry out multiple attacks. Shortly after arresting five members of the cell, Indian authorities discovered that ISIS had provided the cell with weapons, ammunition, chemicals, and operational and technical assistance (Callimachi, 2017).

Previous work found that decision making occurs in private/deep web Internet technologies and/or in face‐to‐face interactions (Derrick et al., 2017b), with the open internet used to disseminate commands and to get reports on outcomes. The dissemination of the decisions and command and control instructions may be delivered in several ways. For example, Al Qaeda relays messages to the public or specific target audiences, and once an effective message is released, the public and news sources tend to spread the message with little effort from the group (Thomas, 2003).

Transferring Ideological and Pragmatic Knowledge

Another way VEOs use ICT is to educate their followers. Essentially, ICT allows for the widespread transferring of knowledge among all types of followers. Knowledge transfer involves transmitting or sending knowledge to a potential recipient and the absorption or use of the transferred knowledge (Davenport, Delong, & Beers, 1998). VEOs have leveraged different ICTs such as knowledge repositories to support the transfer of knowledge. These repositories are especially relevant as they enable the rapid sharing of “know‐how” in an asynchronous manner. This know‐how is represented in two dominant themes, ideology, and pragmatism. First, VEOs use ICTs to transmit knowledge around “ideological purity” and vision. In other words, organizations use technology to communicate sacred values. A sacred value is more than just a shared value or moral value. It is a value that is prescribed by deity and by which adherents judge and evaluate their actions and ultimately their lives. Ideological leadership represents vision‐based leadership, where the vision is a compelling, emotionally evocative view that appeals to virtues of a past ideal state (e.g., a caliphate). Ideological leaders frame this vision around values and standards that must be maintained to rebuild this “pure” society (Mumford, Strange, & Bedell‐Avers, 2006). For example, VEOs use ICTs to identify and highlight violations of standards, morals, and codes of conduct that underlie the overarching ideology. Extreme ideological leaders tend to view violations of such standards and values in dichotomous terms, drawing distinct differences between adherents and nonadherents. The ability to identify and enumerate violations gives a sense of moral superiority and “justness” to the leaders of the cause. Their perceived and perpetuated rightness is necessary to build ideological legitimacy. The strict adherence removes dissonance that brings follower attitudes more in‐line with the leaders’ desires. Similarly, ideological leaders tend to hearken back to times of past greatness, drawing from key figures in the ideology and lessons learned from them. Ideological leadership compares present and past conditions. It can motivate and give a sense of time and purpose (Freeman, 2014; Hofmann, 2017; Hofmann & Dawson, 2014; Mumford et al., 2007).

Second, VEOs use ICTs for the transmission and application of both declarative and procedural knowledge through different presentation modes. These knowledge repositories are likely to be found in the deep web, but some are also freely available. There are sites that offer information about systems being used at locations of interest. Many of these knowledge repositories allow a user to enter system information for the desired target and find specific vulnerabilities that have been cited about that system or software. VEOs can use internet technologies to disseminate and train individuals virtually (Derrick et al., 2017b). This requires fewer experts and overcomes concern about colocation. This distance learning will be in private venues, possibly taking the form of synchronous or asynchronous packages.

Existing Approaches

As illustrated above, VEOs across ideological types (e.g., religious, ethno‐nationalist, and left‐wing) use ICTs to further their ideological and organization goals. The remainder of this section, however, focuses on prior research on jihadist VEOs for two reasons. First, data for this study comes from ISIS—a VEO with jihadi motivations. Second, the past two decades have seen more and more interest in researching digital jihadist narratives. For example, Table 1 summarizes some of the recent works in this area. Although Table 1 is not an exhaustive list, it provides context for differentiation of this work to other approaches. Immediately noticeable is the preponderance of qualitative works using human raters in the VEO space. These “expert‐based” and/or “content‐analytic” studies rely on one or more human raters to identify and analyze different forms of ideological propaganda (Cohen et al., 2016, pp. 2–3). The common criticism of this approach is that it relies on subjective ratings and is open to biases. However, others suggest this approach is better at contextualizing information compared to automated techniques (Cohen et al., 2016). It is very clear that there are many potential negative downstream effects like burnout or post‐traumatic stress for those who repetitively view and process this content.

Table 1.

Sample of Recent Works on VEO Use of Internet Technologies

Article VEO(s) Modality Sample Evaluation
Abrahms, Beauchamp, and Mroszczyk (2016) Various Jihadi Groups (i.e., Al Qaeda in the Arabian Peninsula, Al Qaeda in the Islamic Maghreb, al‐Shabaab, Al Qaeda in Iraq, and Afghan Taliban) Video 473 Videos IntelCenter database (Originally 900 videos) Human
Baines and O’Shaughnessy (2014) Al Qaeda Central Video 37 Videos (from 97 videos) IntelCenter database Human
Cohen et al. (2016) Al Qaeda Central & Al Qaeda in Iraq Text 18 Transcripts (from Site Intel & START organization) Automated and Human
Droogan and Peattie (2016) Al Qaeda in the Arabian Peninsula Text 14 Issues of AQAP’s Inspire magazine Human
Hafez (2007) Various Jihadi Groups Text, Image, and Video 29 Video clips, 5 audio clips, and over 20 online magazines Human
Ingram (2016) Al Qaeda in the Arabian Peninsula & Islamic State in Iraq and Syria (ISIS) Text 14 issues of AQAP’s Inspire and 13 issues of ISIS’s Dabiq Magazine Human
Klausen (2015) Various Jihadi Groups Text 563 Tweets from 59 social media accounts Human
O’Halloran et al. (2016) ISIS Text and Image Dabiq magazine Automated
Pennebaker (2011) Al Qaeda Central & Al Qaeda in the Arabian Peninsula Text 296 Speeches, interviews, and articles Automated
Prentice et al. (2011) Various Jihadi Groups Text 50 Online texts Human and Automated
Qin, Zhou, Reid, Lai, and Chen (2007) Various Jihadi Groups Text, Image, and Video 222,000 Web documents Automated
Smith, Suedfeld, Conway, and Winter (2008) Al Qaeda Central & Al Qaeda in the Arabian Peninsula Text 53 Speeches, interviews, and articles from original and current Al Qaeda Central leaders; Human
49 Documents signed by or issued by Al Qaeda in the Arabian Peninsula members
Torres Soriano (2010) Various Jihadi Groups Text and Video 199 Documents Human
Torres et al. (2006) Various Jihadi Groups Text, Image, and Video 2,878 Documents Human
Torres‐Soriano (2016) Al Qaeda in the Islamic Maghreb Text 482 communiqués issued by AQIM between 1998‐2015 Human
Winter (2015) ISIS Text, Image, and Video 896 (1,146 total gathered) propaganda events Human

Note: We did not “audio” as a separate modality because this generally goes hand‐in‐hand with the “video” classification.

A second interesting takeaway from Table 1 is the diverse number of modalities used to examine radicalizing propaganda. For example, several studies have included more than one modality to examine propaganda‐related narratives. This may be due to an increased awareness in the research community regarding the number and types of social media websites currently available to terrorist organizations. As Correa and Sureka (2013) suggest, deviant groups now have access to various communication modalities such as weblogs, online forums, image sharing websites (e.g., Flickr), and video sharing websites (e.g., YouTube). Thus, researchers’ ability to triangulate information sources is more important now than ever before. The ability to analyze multimodel content is further supported by the domain’s general reliance on qualitative approaches. Very few studies tie multiple platforms together (Hall, Mazarakis, Chorley, & Caton, 2018), and even fewer analyze multiple contents types and/or formats (Wendlandt, Mihalcea, Boyd, & Pennebaker, 2017).

Finally, Table 1 indicates that most of the recent works on Jihadi propaganda have focused on Al Qaeda and their affiliated movements (AQAM) and, to a lesser extent, ISIS. This is not surprising for two (related) reasons. First, AQAM and ISIS have dominated the Jihadi extremist market for several years now (Ligon et al., 2017). In turn, AQAM and ISIS are two of the most recognizable extremist “brands” with the capacity and legitimacy to produce extended propaganda campaigns. In other words, AQAM and ISIS can afford to publish extensive amounts of propaganda, while smaller and lesser‐known groups either completely lack resources or must funnel them to more pertinent organizational processes to survive.

Limitations of Existing Approaches

As alluded to above, the primary limitation of earlier attempts to categorize types of messaging in extremist content is the overreliance on human coders. This is problematic for three reasons. First, the scalability of such approaches is minimal. The use of human coders is highly labor intensive. Extremist propaganda can also be horrifically violent and likewise disturbing to human coders or platforms’ content reviewers—for example, the footage of the 2015 immolation of a Jordanian pilot by ISIS. Burnout of the human coders is a real concern. Second, human coding strategies are susceptible to pre‐existing assumptions or biases held by the raters. In other words, the replicability and reliability of human‐coding strategies are limited. Third, although the use of computerized approaches has grown in recent years, few studies have examined the degree to computational methods compare with the benchmark solutions found by human coders when categorizing extremist content. In radicalization studies, often either human coders or computational pattern assessments are presented as holistic results. However, to our knowledge no work addresses if and to what extent a gap between human and machine understanding of radicalizing content exists. On the basis of these limitations, we present the guiding research question:

RQ: To what degree do computerized text‐analysis approximate benchmark solutions of human coders taken from extremist content on the open web?

Methodology

Data and Its Collection

The data used in this exploratory study are one hundred English language nonindexed, transient web pages posted by ISIS activists or supporters to open, web‐based architectures. These data were harvested using a custom program extracting transient websites introduced and used in Derrick et al. (2017b) (Figure 1).

Figure 1.

Figure 1

Flow Diagram for Monitoring and Scraping Transient Web Pages.

A faction of the hacktivist group Anonymous code‐named Controlling Section (@CtrlSec) was formed in early 2015 to help find and remove ISIS accounts from Twitter (Macri, 2015). During much of this collection, CtrlSec posted ISIS members’ Twitter handles at a rate of approximately one every two minutes. Using the Twitter API, the custom program (Figure 1) follows and logs the individual tweets dispatched by CtrlSec accounts. The program first started collecting accounts released by CtrlSec in August 2015 and has been running since. Using the handles posted by CtrlSec, the handles of suspected ISIS accounts are identified and followed using the Twitter API.

From that list, the system utilizes the Twitter API to download a sample of the latest tweets from the ISIS‐affiliated accounts. After being logged in the database, the tweets are sorted into metadata (e.g., number of web addresses and links, hashtags, mentions) and content (tweet text, links). Links which are posted in the tweets of ISIS accounts are used as launching points to the open architectures where richer content is housed. The software searches for links within tweets referencing anonymous posting services for open content‐publishing transient web pages (e.g., JustPaste.it, dump.to). Next, the software automatically crawls to the referenced webpage, captures PDF and HTML versions of the pages, and stores them to the database. From these pages, the program identifies any links to other transient web pages/open architectures, recursively downloading and analyzing the content until all possible transient links have been found and captured. To date, this process has resulted in the capturing and storing of over eight million one hundred thousand tweets; one million three hundred thousand tweets URLs; one million two hundred thousand images; and forty‐eight thousand transient web pages. From the tweet metadata it can be determined that 69 percent of the total tweets are written in Arabic, 13 percent in English, and 18 percent are other or undefined. In a secondary step a Naive Bayes algorithm was applied to the transient web pages to determine the primary language and determined that 54 percent of the web pages are written in Arabic, 38 percent in English, and 8 percent are other or undefined.

Analysis

A high‐level overview of this article’s approach is as follows, and Figure 2 illustrates the workflow of the current study:

  • (1)

    Data Collection (see previous section, and Table 2).

  • (2)

    Analyze text samples via standard text analytics toolkits.

  • (3)

    Name concepts found in machine results in order to compare with qualitative results (Table 2).

  • (4)

    Repeat steps four and five to produce range of accuracies and cross‐validate results.

  • (5)

    Compare results of automated processes with qualitative results.

Figure 2.

Figure 2

A High‐Level Workflow of the Current Comparative Study.

Table 2.

Manually Coded Results of VEO Content Scraped From Open Architectures and Social Media (see Derrick et al., 2017b)

Quran 64 Help Locals 19
Legitimacy 43 Destroy property 19
Caliphate 38 Bayat 18
Education 37 Destroy by enemy 17
Violent items 35 Motivate 16
Mujahideen 34 Manuals 13
Apostate 34 Baghdad 12
Anti‐west 34 Shame 11
Jihad 31 Apocbattle 9
Sharia 30 Cyber 9
Destroy by ISIS 30 Hijrah 8
Atrocities 29 Hisbah 8
Mohammed 28 Repent 7
Media 28 Diyya 5
Justify 26 Ribat 5
Leader 25 Training 5
Territory 24 Village ldr 2
Military ops 24 Lone wolf 2
Shahid (Martyr) 20

Note: The terms reflect the feature identified, and the numbers reflect the frequency the coder identified the item in the corpus.

ISIS, Islamic State in Iraq and Syria; VEO, violent extremist organization.

Seven researchers trained in qualitative measurements of ISIS leadership and VEO communications evaluated the machine‐created features (explanatory characteristics) to measure goodness of fit to the human‐coded baseline in an open coding format. Intercoder agreement was assessed in a jury format where coders independently assessed the theme presented by the algorithm. Results were compared between coders, and disagreements were discussed with two or more coders and three faculty members with seven or more years of experience in qualitative research until one single concept was agreed upon (see Results). A significant problem in grounding truth between qualitative and quantitative results is in its assessment. Human and the text analytics results produced different numbers of themes (referred to as vectors where k ∈ {37, 10, 20, 18}). Due to the uneven vectors produced by the humans and algorithms (see discussion below) our study concentrated on qualitatively measuring goodness of fit. Finally, the themes were operationalized in four states: the theme was present, half present, partially present, or not present (see Tables 6 and 7).

Table 7.

Full Comparison of n‐grams, LDA, and LIWC Against Manually Coded Radicalizing Discourse Found in Transient Web Pages. Fully Darkened Circles Indicate Fully Present; Half Indicates Mostly Present; A Quarter Circle Indicates Partially Present; An Empty Space Indicates Not Present in the Results

graphic file with name POI3-12-109-g002.jpg

The 2017 authors established thirty‐seven categories (themes) of activities and outreach taking place online, where the reported the Cohen’s κ interrater agreement as 0.64. These are thirty‐seven categories are represented in Table 1. By randomly selecting the one hundred web pages, the 2017 authors sampled a wide variety of types of web content produced by ISIS supporters as disseminated via transient open architecture, including pictures or photographs with short captions or headlines and text or essays. At the time of the data collection and analysis, English‐only web pages warranted deep empirical attention for three primary reasons. First, ISIS has directly engaged English speakers, particularly in the United States. For example, in 2014 ISIS established a “#AmessagefromISIStoUS” marketing campaign on Twitter and posted a YouTube video called “A message to the American People”—all in English. Second, given the threat of foreign terrorist fighters returning to the West, it is imperative to examine the messaging content that the majority of western citizens can read and understand. As such, despite the majority of ISIS propaganda being in Arabic or other non‐English languages, a deep understanding of ISIS’s messaging in English warrants attention. Finally, this is an exploratory study that can be used to benchmark and examine larger data sets with more languages.

The one hundred English‐only transient web pages provided a corpus comprising two hundred forty‐five thousand and seventy‐five words after removing transliterations (6areeq, shaam) and misspellings (th for the; viel for veil). Transliterations and misspellings have an outsized impact on result quality in the n‐grams analysis as they make the term matrix sparser, and are not recognized (and thus excluded) by the dictionaries in the LIWC analysis. As such, they were excluded for all three analyses. Hence, two hundred fifteen thousand four hundred and ninty‐nine words were recognized by English dictionaries (~88 percent). On average, 67 percent of the words from each transient website were detected as English. Words were transformed to lower case to avoid errors due to capitalization. Non‐alphanumeric characters were removed. Once a primary language of the webpage was determined, characters outside of the particular language were removed via regular expressions. URLs and other information specific to the hosting service were also stripped from the webpage content before the analyses.

The current study employed simple, off‐the‐shelf packages. This has several purposes. By employing standard packages, replicability is aided. Good practice demands employing the computationally simplest packages in terms of eventually benchmarking performance of more complex algorithms. Due to these reasons, we employ n‐grams and LDA (Blei et al., 2003; Jurafsky & Martin, 2014). Both algorithms are appropriate for creating feature vectors (vectors of explanatory characteristics) from unstructured text. While similar in concept, the two approaches are distinct.

The n‐grams are commonly used to summarize text given their basic co‐occurrence algorithmic basis. n‐grams can be trained and cross‐validated in a single corpus. This has the desirable property of being insensitive to—and thus comparably flexible with—unknown words. It has also been used in similar approaches in previous work (see e.g., Burnap & Williams, 2015, considering online hate). In the n‐gram model, each word is assumed to be drawn from the same term distribution. A topic is drawn if a predetermined number comprised of a string of words occurs. These strings are counted if the sequence occurs in the text 1+ times.

So, n‐grams are limited in contextualizing corpora as the results are context insensitive (Jurafsky & Martin, 2014). Therefore LDA is also implemented in this study (McCallum, 2002). A generative model, or a model with a joint probability distribution, LDA is even more recognized as fitting the task of modeling themes in text‐based corpora (Blei, 2012). A major draw, LDA’s bag‐of‐words assumption allows the context and not only the topics to be identified. It has been applied in similar detection scenarios, including detection of online pornography (Tang et al., 2009) and hate speech (Burnap & Williams, 2015). We follow the standard process of LDA with Gibbs sampling (Blei, 2012; Boecking, Hall, & Schneider, 2015; McCallum, 2002):

  • 1.
    For each topic k
    • i.
      Draw a distribution over words βk~DirV(η).
  • 2.
    For each document d
    • i.
      Draw a vector of topic proportions θd~DirK(α)
    • ii.
      For each word n,
      • 1.
        Draw a topic assignment Zd,n~Mult(θd), Zd,n{1,,K}
      • 2.
        Draw a word Wd,n~Mult(βZd,n), Wd,n{1,,V}

The hidden variables in this model are the topics β, the proportions of topics per document θ, and the topic assignments per word Z. α controls the concentrations of topics per document, while η controls the topic‐word concentrations. The multinomial parameters β, the topics are smoothed by being drawn from a symmetric Dirichlet conditioned on the data.

Finally, the sentiment analysis package LIWC was employed to assess the primary structure of the content and latent meaning of the corpus. Meaning is fundamental in the case of radicalizing content, and Natural Language Processing testing and validation on VEO content (and other propaganda generally) is still lacking (Table 1). Sentiment analysis is generally more appropriate for tackling latent meaning (Chen, 2008; Liu, 2010; Mirani & Sasi, 2016). LIWC is a limitedly content‐sensitive sentiment analysis package (Pennebaker, Chung, Ireland, Gonzales, & Booth, 2007). This approach enables us to look not only at usage or context but also the latent meaning of the corpus (Pennebaker, Mehl, & Niederhoffer, 2003). The focus on latent meaning rather than individual wording has the secondary useful attribute of allowing assessment of context even in the case that the language/slogan changes, as the focus for LIWC is patterns in word usage, and not the term itself. While drawbacks and criticisms of LIWC’s approach exist (see e.g., Beasley & Mason, 2015; Panger, 2016), LIWC has been found to cope admirably well with relatively sparse social media data (Lindner, Hall, Niemeyer, & Caton, 2015), and has been successfully applied to suicide ideation scenarios (Handelman & Lester, 2007) and analyses of terrorist discourse (Cohen et al., 2016; Logan & Hall, 2019; Pennebaker, 2011). A large basis of literature employs LIWC scoring either directly or in combination with machine learning on social media data, making it the strongest benchmark sentiment analysis package to date (Coviello, Fowler, & Franceschetti, 2014; Hughes, Rowe, Batey, & Lee, 2012; Kramer, 2012; Lin & Qiu, 2013; Thelwall, Buckley, & Paltoglou, 2012; Thelwall, Buckley, Paltoglou, Cai, & Kappas, 2010).

Evaluation

n‐grams

Given the sparse corpus, we ran three experiments. The experiments looked at strings of three, four, and five words (or, n‐grams). Of these, 4grams produced the most compact results, reasonably preserving the grammar and structure of the documents. Instances of repetition or broken structure were consolidated when they were obviously related (i.e., United States Department of, States Department of State). This resulted in eight hundred and ninty‐seven unique observations producing five thousand three hundred and seventy‐one 4grams. Figure 3 shows that the distribution approximates Zipf’s Law. Twitter handles, tags, and IDs are removed.

Figure 3.

Figure 3

4gram Distribution by Term Volume Demonstrating Zipf’s Law.

In the manual assessment, the jury (a team of four researchers) found that ten topical themes emerged (Table 3). This is significantly lower than the thirty‐seven themes listed in Table 2. Approximately a third of the corpus involves self‐references to ISIS, and another quarter represents Quranic references to Islam and jihad. Fifty‐one instances of 4grams are unclassified as they are either internet jargon or otherwise defy meaningful interpretation in the 4gram form (e.g., “before right and after”; “singing” n a n a in the text).

Table 3.

Results of the 4gram Analysis With Topic Categories

Topic Comparison to Derrick et al. (2017b) Example Text f
Self‐reference Leadership; motivational speeches compassionate islamic state wilayat: islamic state of iraq 1,806, 33.6%
Religion Quran verses/teachings/prayers; Mohammed may allah be pleased; the cause of allah; al fatawa volume 13 1,399, 26%
US Government/Military Defense or justifications for violence; Military operations; cyberwarfare defense information systems agency; state.gov department of state 1,320, 24%
Delegitimizing Rival Leaders Diminishing the legitimacy of others; Apostate regimes abu bakr al baghdadi; bahira omar yaqoob said; by jabhat al joulani 262, 4.8%
Delegitimizing Rival Orgs Diminishing the legitimacy of others; destruction gunfire exchange with rafida; the free syrian army 156, 2.9%
Transliteration of ahrar al shaam; in deir az zour 105, 1.9%
Unity Apostate regimes it saddens me how; brothers in the west 103, 1.9%
Public Institutions national institutes of health; army corps of engineers 81, 1.5%
Society Media centers may contain sensitive material; islamic states reports src 88, 1.6%
Misc n a n a 51, .09%

Note: Column 1 represents the contained topics; column 2 compares to Derrick et al. (2017b); column 3 contains sample 4grams; column 4 indicates the frequency of occurrence and percent of corpus represented. The long tail is not included, thus the total does not equal one hundred percent.

In comparison to the original study, the 4gram results are less granular than the human coders. What is compelling about these findings is that while n‐grams provided fewer categories than human coders, the observations within these categories resulted in higher‐level themes when compared with human coders—in other words, 4grams produced more compact results compared with human coders. Table 7 demonstrates that thirteen of the themes found in (Derrick et al., 2017b) were subsumed by the seven n‐grams themes identified in Table 3. With one thousand three hundred and twenty unique 4grams, our results indicate that the presence of discourse about the US government and military is much more dominant than the qualitative coders initially assessed in Derrick et al. (2017b). This may be due to the fact that the qualitative coders added references to the US government and military into adversarial themes. Interesting themes also emerged from the n‐gram results that were not identified by human coders, especially transliterated aspects and broad references to western society. Unity, a theme identified 88 times by n‐gram, is particularly of interest as it theoretically should be tied to reasons for joining terrorist organizations (i.e., Horgan’s work on “opportunities for camaraderie” indicates that terrorist propaganda is replete with messages about kinship and unity [Derrick et al., 2017b]). As social media are one of the major recruitment arms of ISIS and VEO groups broadly speaking, this particular theme would have been expected to be much more present in the themes identified qualitatively.

Overall, naïve approaches like n‐grams deliver visibility into the corpus, but poorly approximate qualitative benchmarks, particularly community‐oriented topics and the least‐common results found manually (i.e., the long tail). This is likely due to the word‐counting approach used. With its higher granularity, n‐gram features miss the nuance and detail of qualitative results across all categories, but do detect subtleties that human coders missed in other cases.

LDA

The MAchine Learning for LanguagE Toolkit’s (MALLET; McCallum 2002) implementation of LDA was used to create feature vectors in the style of Boecking et al. (2015). Several text‐preprocessing steps were conducted. Unwanted characters and items in the messages are removed such as punctuation, emoticons, user‐mentions, and URLs. Common stop words are also removed. A single input document of the transient pages is created (n = 245,075) and passed on to LDA.

We used a symmetric Dirichlet distribution, initializing α to 0.5; β corresponds to the word distribution for each topic (here, 15). K represents the number of distinct concepts. We estimated the number of concepts in the corpus by examining if the resulting topic models of different sizes resulted in reasonable, cohesive topics that are not included as training data. All reported results are subject to a 10‐fold cross‐validation. Thus, 2 × 6 experiments were conducted: one treatment where the corpus contained twitter handles, and one where the corpus did not. No significant differences were observed. We conducted two hundred runs apiece for each K ∈ {10, 15, 20, 25, 30, 37} on the training set (90 percent of documents), and measured the perplexity on the validation set (the remaining 10 percent). Of these, K = 20 yielded the best results. This is likely due to performance degradation from the sparse text. Figure 4 represents the one hundred most frequent terms where the word size is proportionate to its probability in the topic, but not proportionate to its frequency in the corpus.

Figure 4.

Figure 4

Word Cloud Representing the One Hundred Most Frequently Returned Terms in Twenty Feature Vectors of Jihadi Discourse.

The twenty features that emerged converge more closely to the qualitative results in terms of the focus on religious violence and apocalyptic viewpoints (Table 4). This lines up with the two primary themes of ISIS content identified in Derrick et al. (2017b), namely, Ideological Rationality and Violence (Table 5). Particularly interesting is that only two Twitter handles, @almoohajeermos and @souriethipeace, were allocated in the experimental results with Twitter data. This indicates the intensity of these two handles’ posts on specific topics. Surprisingly, the LDA features highlight the hotly fought‐over cities Homs, Aleppo, and Mosul, but neither Raqqa (the then‐capital of ISIS) nor Baghdad (as indicated in Table 2) appear. Also broadly absent are features discussing the community around ISIS, something quite present in the qualitative results. In agreement with the 4gram results, religious outreach is hugely present. However, the US military, civil society, and public media are nonexistent. Overall, however, LDA outperformed n‐gram in that it identified seventeen of the constructs found by human coders (see Table 7). Also, LDA was more effective at identifying violent themes. Given its relatively low computational complexity and the fact that it is not strictly keyword or phrase‐based, LDA may be useful for quickly identifying text‐based content that violates Terms of Service, in terms of violent content.

Table 4.

Results of the Latent Dirichlet Allocation (LDA) Analysis With Topic Categories. Column 1 Represents the Named Topics; Column 2 Contains Sample Topics

Comparison to Results of Manual Coding in Benchmark Article (Derrick et al., 2017b) Text in Feature Topics, α = 0.5, β = 15, K = 20
Caliphate/Unity Fight dawlah syria shariah war mentioned back path tawheed called banner allaah evidence council benefit
Sharia Law/Motivational Speeches Group muslim fighting clear center men matter support dont disbelief reality areas true sharia worship
Military Operations/Caliphate Muslims left person making killed caliphate believers apostate proof makes mariahajer face fighter haqq homs
Delegitimizing Rivals/Media People time good ali god taliban permissible end amir wal didnt twitter giving pray deeds
Military Operations/Atrocities Make leave kill sham asked regime law wa put khawarij khilafah video judgement battalion fact
Anti‐West Islam systems information ibn agency takfeer found world great defense agreed prison means abdul albani
Delegitimizing Rivals/Leadership People army groups disbelievers fought quran evil hanisibu free syrian hadith imam based allegiance stop
Violence State islamic shaam iraq ahrar gov join attacked photo attacks difficult munafiqeen defected hour turn
Jihad/Mujahedeen Jihad mujahideen religion leaders soldiers west part attack wealth taking kuffar followers doesnt control find
Mujahedeen/Bayat Killing kufr air base today battle ziadzd accept women states lot muhajireen camp sake bayah
Delegitimizing Rivals Abu al command nusra told death signal lol sin shaykh mujahid members jn muhajir operations
Quran Verses, Teachings, Prayers/Apostate Regimes Allah prophet messenger peace sunnah blessings pleased heard merciful judge revealed family barracks surely christian
Military Operations Khawrij fsa blood day scholars knowledge brother claim city military mosul message amir till commander
Delegitimizing Rivals/Apostate Regimes Ummah weapons muhammad call rule land apostates years dead aleppo mullah knew omar long mariahajer
Military Operations/Motivational Speeches Made killed souriethipeace abo words allies hasan almoohajeermos fighters back happened night jabalalaiza actions intelligence
Quran Verses, Teachings, Prayers Allah truth give lies book high life volume enemies lying mercy fatawa lack speech hand
Anti‐West/Shaming Force brothers media ruling reason place order office son lie sisters hell town issue witness
Self‐Reference Man due allah laws fear isis victory news leadership emirate false matters country live praise
Delegitimizing Rivals Al jabhat nusrah front joulani statement deir zour az started qaidah jawlani central approach joined
Leadership Al sheikh bin ah leader umar bakr shari baghdadi ayman abdullah zawahiri liwa abi speaking

Table 5.

Component Transformation Matrix Results of the Exploratory Factor Analysis (EFA) on Linguistic Inquiry and Word Count (LIWC) Sentiment Analysis Data

Inline graphic Inline graphic

LIWC

Finally, we assessed the latent structure of the sentiment behind the text with LIWC (Tausczik & Pennebaker, 2010). Considering composite variables, LIWC 2015 identified that 90.41 percent of the text could be classified as “Analytical,” indicating the text’s propensity to portray reasoning: formal, logical, and hierarchical thinking patterns. Analytical texts might be perceived as intellectual, rational, systematic, emotionless, or impersonal. On the other hand, the analysis found that only one percent of the corpus could be described as having an “Authentic” tone. Authenticity in the case of LIWC 2015 is closely related to deception. Both values are largely outside of their expected ranges as established in Pennebaker, Boyd, Jordan, and Blackburn (2015). These aspects suggest that the text is pitching a vision of the Caliphate that aims to recruit individuals, albeit in a way that doesn’t reflect the on‐the‐ground reality.

The two hundred fifteen thousand four hundred and ninty‐nine words recognized by English dictionaries from the one hundred web pages are scored for sentiment over sixty‐four psycholinguistic categories. With this numeric data, we assess the latent structure of the discourse with an exploratory factor analysis (EFA) run in SPSS Statistics 24. EFA allows us to understand how the concepts and the emotions behind the corpus combine, creating a more granular assessment of online jihadists. The EFA employed a varimax rotation, which converged in twenty‐one iterations, suppressing coefficients below 0.3. The EFA produced eighteen latent constructs (Table 5) based on the sentiment of the text, which were named by the content experts as discussed above. Table 5 displays the component transformation matrix (CTM). The CTM transforms rotated loadings by postmultiplying the matrix of the original loadings by the transformation matrix. The values in the CTM are functions of the angle(s) of rotation of the components. As they are measures of rotation degree, the intersection of two constructs is not expected to equal 1 as in a correlation matrix (For more information on EFA and CTM, please see Harman, 1976). For the 18 × 64 rotated component matrix listing the loading factors, please see the Supplementary Materials online.

Similar to LDA, LIWC directly matched human coders on 16 constructs, and LIWC’s results lined up with those most frequently identified by humans (e.g., violent content; see Tables 6 and 7 for more detail). LIWC also identified groupings that the human coders did not (e.g., ideological adherence, as discussed in Derrick et al. 2017b). Moreover, the two primary themes of ISIS content identified in the original article are strongly present in the LIWC results (see Supplementary Materials; Table 7). The alignment of the principle component analysis (PCA) results with the previous findings suggest that the LIWC constructs demonstrate construct validity for VEO research in as far as it concerns social data. Additionally, LIWC identified aspects of communal living not present in the other feature vectors, similar to the Unity theme found by the 4grams analysis. Thus, while LIWC was not a perfect match to the human coder, it was a close approximate on the most important constructs. LIWC particularly identified subtler groupings that human coders also found, which were however lost in the 4gram and LDA results (e.g., impact uncertainty).

Table 6.

n‐grams, LDA, and LIWC Results Compared With the Two Major Themes of Ideological Rationality and Violence Reported in Derrick et al. (2017b)

graphic file with name POI3-12-109-g001.jpg

Note: Fully darkened circles indicate fully present; half indicates mostly present; a quarter circle indicates partially present; an empty space indicates not present in the results.

Discussion

Overall, all three automated coding procedures performed similarly to the human coder in terms of detecting the most frequently occurring content (e.g., anti‐Western sentiment). Their performance varied wildly from the 2017 results at the long tail. Table 7 compares the efficacy of each computational approach as compared to the human coders. While no one procedure matched human coder findings perfectly, human‐identified themes such as Qur’anic Verses, Diminishing the Legitimacy of Rivals, and Anti‐West rhetoric were identified by n‐gram, LDA, and LIWC. However, less‐frequently occurring themes identified by human coders (e.g., Hisbah,3 Training Camps) were not detected by any of the automated techniques. This has several implications. It is reasonable to assume that automated processes can be powered by off‐the‐shelf text analytics packages to detect and classify major themes; however, we cannot expect a high overall accuracy rate due to these approaches’ lack of nuanced or latent theme detection. This split between effectiveness when classifying common themes and latent or nuanced themes has been shown in practice by those platforms that publish their results (e.g., Facebook).

While it is encouraging to find that humans and machines recognize and understand major concepts similarly, it is disappointing that the machines did not detect themes like Lone Wolf/Call for Attacks on Foreign Lands, as these are the type(s) of propaganda which can be acted on with very little interaction between the group and the individual (Ligon et al., 2018), and thus are harder to practically manage. The reasons for the lack of signal at the long tail could be many, in addition to high signal/noise problems. Some of the text may be accompanied by images, which are out of scope for this study. With the original study’s α of 0.64, it is possible that the human classifiers artificially compartmentalized the structure of the text compared to computational classification (e.g., Hijrah could belong to Qu’ranic Verses/Teachings).

Conclusion

Much like pornography, “you know VEO content when you see it”—describing it however is not straight‐forward. Where the qualitative efforts are valuable, their practical usefulness is limited due to the nature and scale of the content, as well as with inherent bias added by human processing. Automated classification based on latent patterns in radical content promises to be more efficient. This work finds however that the computational approach has not yet fully approximated manual coders’ understanding of ground truth. Reasons for this are varied, but are likely due to the differences in humans’ holistic consumption of VEO content compared with the reliance of computational approaches on naïve pattern mining of the data. While neither approach is perfect, a computational approach that misses highly emotive VEO content (e.g., Lone Wolves) is not sufficient to implement in a real‐world scenario. This has been practically experienced by open social platforms as they reactively manage and remove VEO content. This work is a first effort toward empirically outlining possible reasons that misclassifications can (and do) occur. On the basis of these results, we argue that before fully automated approaches can be implemented, a more careful blending of human content understanding and machine abilities is required.

In response to the original research question of how well the processes approximate one another, it is compelling and promising that many high‐level concepts transfer well, especially themes surrounding violence and delegitimizing rivals. However, the granularity found by human coders is absent, for example, the machine approaches missed themes of education, instructional materials, and some theological topics like Hisbah. And indeed, some concepts like Unity (found by the n‐gram analysis) were not present in the 2017 results. This suggests that while there is more work to be done, a computationally efficient and theoretically‐grounded set of features can be developed to aid in automatically detecting radical content on the open internet. The missing puzzle pieces are described in the coming section.

Limitations and Future Work

Despite the global outlook of jihadists, the lingua franca of groups like ISIS is Arabic, which is poorly covered in the literature. This is important to note for any language‐based automated detection task, as it has been shown that prediction accuracy is best when the locally dominant language is also employed in conjunction with English (Boecking et al., 2015). Thus, a serious limitation to our approach is the lack of availability of validated Arabic NLP packages. Currently available packages like CORE NLP and Weka only offer partial functionality; LIWC’s Arabic dictionaries are also still in the validation process. Our approach also lacks several more expansive or more granular assessments due to the lack of reliable Arabic language algorithms. This also partially explains the relatively simplistic, off‐the‐shelf approach selected. Our approach relies on character recognition to split multilingual texts; this unsophisticated approach would be better treated with language detection capabilities and could have possibly caused misallocation of words. Other approaches for text processing including td‐idf, entity recognition algorithms, or lemmatization do not fit the current scope, although they are being considered for future work. tf‐idf, a process to establish how important a word is in a given document, was not included in the initial work as it may inadvertently misallocate words that appear frequently (e.g., self‐references, references to religious figures), which are pertinent to ISIS’s branding and recruitment strategies (Derrick et al., 2017b; Ingram, 2016).

Future works may also consider evaluating the accuracy of the assignment of documents to specific topics, and not merely the creation of topics as demonstrated here. Finally and possibly crucially, some web pages are accompanied by images; this could explain the granularity of the human‐coder results, particularly at the long tail, which is not replicated by text analysis algorithms employed in the current analysis. More work needs to be directed at the analysis of images (pictures and videos) alone and concurrently with text‐based feature vectors.

Possible extensions of this work are many. The foremost goal is developing a flexible, detection algorithm to support the classification of radicalizing content. This can include language‐based features, metadata‐based features (views, shares, likes), or image properties. Such content requires that any feature‐based classifier be linguistically localized, and extensive work must be done on Arabic NLP capabilities to realize this. Another extension is understanding precisely what content radicalizes. This requires a multi‐method approach, including content analysis, topic modeling, and sentiment analysis. To support triangulation of the above results, an eye‐tracking study evaluating what content is consumed by reviewers when they view VEO content would be useful. At this point, we can expect to find an approximation of the “ground truth,” which allows us to begin working on predication scenarios.

Validating our approach across different open and social media platforms would help understand the context of radicalization, and pathways of radicalization of (vulnerable) individuals. Finally, from both the perspectives of content generation and research, the next frontier of social media research is in visual media. Computer vision has been employed recently to combat human trafficking (Dubrawski et al., 2015) and child pornography (Ulges & Stahl, 2011) but has yet to be trained and tested on jihadi content. Research is required to adapt the concepts to the VEO scenario, and expand the capabilities of the computer vision domain by training algorithms on radicalizing content scraped from outlets like YouTube, Instagram, and visual content from sites like Twitter and Facebook.

Supporting information

Additional Supporting Information may be found in the online version of this article at the publisher’s website.

Supporting information

Biographies

Margeret Hall, School of Interdisciplinary Informatics, 1110 S 67th St PKI 275 C, Omaha, Nebraska 68182, United States [mahall@unomaha.edu]

Michael Logan, School of Criminology and Criminal Justice, Omaha, Nebraska, USA

Gina S. Ligon, College of Business Administration, Omaha, Nebraska, USA

Douglas C. Derrick, School of Interdisciplinary Informatics, Omaha, Nebraska, USA

Notes

1

“Leadership decision making” refers to the strategic, operational, and tactical decisions made by violent extremist leaders to ensure the survival of their organization. For instance, leadership decisions on how to brand their organization play a large role in the specific types of people they can recruit.

2

At the time, the Chechen Mujahidin would qualify as a VEO since the organization showed markers of operational collaboration, specialization, and a shared ideological mission against Russian occupation (Ligon et al., 2013).

3

Hisbah refers to the individual and collective duty to intervene and coercively “enjoy good and forbid wrong” in accordance with Sharia law.

References

References

  1. Abrahms, M. , Beauchamp N., and Mroszczyk J.. 2016. “What Terrorist Leaders Want: A Content Analysis of Terrorist Propaganda Videos.” Studies in Conflict & Terrorism 40: 1–18. [Google Scholar]
  2. Alghamdi, H.M. , and Selamat A.. 2012. “Topic Detections in Arabic Dark Websites Using Improved Vector Space Model.” In Conference on Data Mining and Optimization, 6–12.
  3. Alonso, O. 2019. “The Practice of Labeling: Everything You Always Wanted to Know About Labeling (But Were Afraid to Ask).” In Companion Proceedings of the 2019 World Wide Web Conference on—WWW ’19, 1293.
  4. Baines, P.R. , O’Shaughnessy N.J., Moloney K., Richards B., Butler S., and Gill M.. 2010. “The Dark Side of Political Marketing: Islamist Propaganda, Reversal Theory and British Muslims.” European Journal of Marketing 44 (3/4): 478–95. [Google Scholar]
  5. Baines, P.R. , and O’Shaughnessy N.J.. 2014. “Al‐Qaeda Messaging Evolution and Positioning, 1998–2008: Propaganda Analysis Revisited.” Public Relations Inquiry 3 (2): 163–91. [Google Scholar]
  6. Balmer, J.M.T. 2001. “Corporate Identity, Corporate Branding and Corporate Marketing‐Seeing Through the Fog.” European Journal of Marketing 35 (3/4): 248–91. [Google Scholar]
  7. Bartlett, J. , and Reynolds L.. 2015. State of the Art 2015: A Literature Review of Social Media Intelligence Capabilities for Counter‐Terrorism. London: Demos. [Google Scholar]
  8. Beasley, A. , and Mason, W. 2015. “Emotional States vs. Emotional Words in Social Media.” In Proceedings of ACM WebSci’15. Oxford, England: ACM Press.
  9. Blei, D.M. 2012. “Probabilistic Topic Models.” Communications of the ACM 55 (4): 77–84. [Google Scholar]
  10. Blei, D.M. , Ng A.Y., and Jordan M.I.. 2003. “Latent Dirichlet Allocation.” Journal of Machine Learning Research 3: 993–1022. [Google Scholar]
  11. Boecking, B. , Hall M., and Schneider J.. 2015. “Event Prediction With Learning Algorithms—A Study of Events Surrounding the Egyptian Revolution of 2011 on the Basis of Micro Blog Data.” Policy & Internet 7 (2): 159–84. [Google Scholar]
  12. Boero, N. , and Pascoe C.J.. 2012. “Pro‐Anorexia Communities and Online Interaction: Bringing the Pro‐ana Body Online.” Body & Society 18 (2): 27–57. [Google Scholar]
  13. Braddock, K. 2015. “The Utility of Narratives for Promoting Radicalization: The Case of the Animal Liberation Front.” Dynamics of Asymmetric Conflict 8 (1): 38–59. [Google Scholar]
  14. Burnap, P. , and Williams M.. 2015. “Cyber Hate Speech on Twitter: An Application of Machine Classification and Statistical Modeling for Policy and Decision Making.” Policy & Internet 7 (2): 223–42. [Google Scholar]
  15. Callimachi, R. 2017. “Not ‘Lone Wolves’ After All: How ISIS Guides World’s Terror Plots From Afar.” The New York Times: 1–13.
  16. Caren, N. , Jowers K., and Gaby S.. 2012. “Research in Social Movements, Conflicts and Change! Media, Movements, and Political Change! A Social Movement Online Community: Stormfront and the White Nationalist Movement.” Research in Social Movements, Conflicts and Change 33: 163–93. [Google Scholar]
  17. Chen, H. 2008. “Sentiment and Affect Analysis of Dark Web Forums: Measuring Radicalization on the Internet.” In IEEE International Conference on Intelligence and Security Informatics, 2008, IEEE ISI 2008, 104–09.
  18. Chen, H. , Chung W., Xu J.J., Wang G., Qin Y., and Chau M.. 2004. “Crime Data Mining: A General Framework and Some Examples.” Computer 37 (4): 50–6. [Google Scholar]
  19. Cohen, S.J. , Kruglanski A., Gelfand M.J., Webber D., and Gunaratna R.. 2016. “Al‐Qaeda’s Propaganda Decoded: A Psycholinguistic System for Detecting Variations in Terrorism Ideology.” Terrorism and Political Violence 30 (1): 1–30. [Google Scholar]
  20. Correa, D. , and Sureka A.. 2013. “Solutions to Detect and Analyze Online Radicalization: A Survey.” arXiv:1301.4916, 1–30. [Google Scholar]
  21. Coviello, L. , Fowler J.H., and Franceschetti M.. 2014. “Words on the Web: Noninvasive Detection of Emotional Contagion in Online Social Networks.” Proceedings of the IEEE 102 (12): 1911–21. [Google Scholar]
  22. Crosset, V. , Tanner S., and Campana A.. 2019. “Researching Far Right Groups on Twitter: Methodological Challenges 2.0.” New Media & Society 21 (4): 939–61. [Google Scholar]
  23. Davenport, T.H. , Delong D.W., and Beers M.D.. 1998. “Successful Knowledge Management Projects.” Sloan Management Review 39 (2): 43–57. [Google Scholar]
  24. Denning, D.E. 2010. “Terror’s Web: How the Internet is Transforming Terrorism.” In Handbook of Internet Crime, eds. Jewkes Y., and Yar M.. New York, NY: Routledge, 194–213. [Google Scholar]
  25. Derrick, D. C. , Ligon G. S., Harms M., and Mahoney W.. 2017a. “Cyber‐Sophistication Assessment Methodology for Public‐Facing Terrorist Web Sites.” Journal of Information Warfare 16 (1): 13–30. [Google Scholar]
  26. Derrick, D. , Sporer K., Church S., and Ligon G.. 2017b. “Ideological Rationality and Violence: An Exploratory Study of ISIL’s Cyber Profile.” Dynamics of Asymmetric Conflict 9 (1–3): 57–81. [Google Scholar]
  27. Droogan, J. , and Peattie S.. 2016. “Reading Jihad: Mapping The Shifting Themes of Inspire Magazine.” Terrorism and Political Violence 30 (4): 1–34. [Google Scholar]
  28. Dubrawski, A. , Miller K., Barnes M., Boecking B., and Kennedy E.. 2015. “Leveraging Publicly Available Data to Discern Patterns of Human‐Trafficking Activity.” Journal of Human Trafficking 1 (1): 65–85. [Google Scholar]
  29. Elliott, I.A. , and Beech A.R.. 2009. “Understanding Online Child Pornography Use: Applying Sexual Offense Theory to Internet Offenders.” Aggression and Violent Behavior 14 (3): 180–93. [Google Scholar]
  30. Facebook . 2019. Community Standards Enforcement Report. Menlo Park, USA.
  31. Fenstermacher, L. , and Leventhal T.. 2011. Countering Violent Extremism Scientific Methods & Strategies. Washington, DC: NSI, Inc. [Google Scholar]
  32. Fox, N. , Ward K., and O’Rourke A.. 2005. “Pro‐Anorexia, Weight‐Loss Drugs and the Internet: An ‘anti‐recovery’ Explanatory Model of Anorexia.” Sociology of Health and Illness 27 (7): 944–71. [DOI] [PubMed] [Google Scholar]
  33. Fox, N. , Ward K., and O’Rourke A.. 2006. “A Sociology of Technology Governance for the Information Age: The Case of Pharmaceuticals, Consumer Advertising and the Internet.” Sociology 40 (2): 315–34. [Google Scholar]
  34. Freeman, M. 2014. “A Theory of Terrorist Leadership (and its Sonsequences for Leadership Targeting).” Terrorism and Political Violence 26 (4): 666–687. [Google Scholar]
  35. Hafez, M.M. 2007. “Martyrdom Mythology in Iraq: How Jihadists Frame Suicide Terrorism in Videos and Biographies.” Terrorism and Political Violence 19 (1): 95–115. [Google Scholar]
  36. Hall, M. , Mazarakis A., Chorley M., and Caton S.. 2018. “Editorial of the Special Issue on Following User Pathways: Key Contributions and Future Directions in Cross‐Platform Social Media Research.” International Journal of Human–Computer Interaction 34 (10): 895–912. [Google Scholar]
  37. Handelman, L.D. , and Lester D.. 2007. “The Content of Suicide Notes From Attempters and Completers.” Crisis 28: 102–104. [DOI] [PubMed] [Google Scholar]
  38. Harman, H.H. 1976. Modern Factor Analysis, 3rd ed. Chicago, IL: University of Chicago Press. [Google Scholar]
  39. Hashemi, M. , and Hall M.. 2018. “Visualization, Feature Selection, Machine Learning: Identifying The Responsible Group for Extreme Acts of Violence.” IEEE Access 6: 70164–71. [Google Scholar]
  40. Hashemi, M. , and Hall M.. 2019. “Detecting and Classifying Online Dark Visual Propaganda.” Image and Vision Computing 89: 95–105. [Google Scholar]
  41. Holt, T.J. , and Bolden M.. 2014. “Technological Skills of White Supremacists in an Online Forum: A Qualitative Examination.” International Journal of Cyber Criminology 8 (2): 79–93. [Google Scholar]
  42. Hofmann, D. C. 2017. “The study of terrorist leadership: where do we go from here?” Journal of Criminological Research, Policy and Practice 3 (3): 208–221. [Google Scholar]
  43. Hofmann, D. C. , and Dawson L. L.. 2014. “The Neglected Role of Charismatic Authority in the Study of Terrorist Groups and Radicalization.” Studies in Conflict & Terrorism 37 (4): 348–368. [Google Scholar]
  44. Hughes, D.J. , Rowe M., Batey M., and Lee A.. 2012. “A Tale of Two Sites: Twitter vs. Facebook and the Personality Predictors of Social Media Usage.” Computers in Human Behavior 28 (2): 561–69. [Google Scholar]
  45. Ingram, H. 2016. “A Brief History of Propaganda During Conflict: Lessons for Counterterrorism Strategic Communications.” The International Centre for Counter‐Terrorism—The Hague 7: 35–6. [Google Scholar]
  46. Jurafsky, D. , and Martin J.H.. 2014. n‐Grams, Speech and Language Processing. Upper Saddle River, NJ: Prentice‐Hall, Inc. [Google Scholar]
  47. Klausen, J. 2015. “Tweeting the Jihad: Social Media Networks of Western Foreign Fighters in Syria and Iraq.” Studies in Conflict & Terrorism 38 (1): 1–22. [Google Scholar]
  48. Kramer, A. 2012. “The Spread of Emotion Via Facebook.” In Proceedings of the 2012 ACM annual conference on Human Factors in Computing Systems—CHI ’12. New York, NY: ACM Press, 767–70.
  49. Latonero, M. 2011. Human Trafficking Online. The Role of Social Networking Sites and Online Classifieds. Los Angeles: Center on Communication Leadership & Policy Research Series. [Google Scholar]
  50. Ligon, G. , Hall M., and Braun C.. 2018. “Digital Participation Roles of the Global Jihad: Social Media’s Role in Bringing Together Vulnerable Individuals and VEO Content Introduction: The Internet’s Facilitation of Violent.” In International Conference on HCI in Business, Government, and Organizations, 485–495.
  51. Ligon, G.S. , Logan M., Hall M., Derrick D.C., Fuller J., and Church S.. 2017. The Jihadi Industry: Assessing the Jihadi Industry. College Park, MD: START. [Google Scholar]
  52. Ligon, G.S. , Simi P., Harms M., and Harris D.J.. 2013. “Putting the “O” in VEOs: What Makes an Organization? Dynamics of Asymmetric Conflict 6 (1–3): 110–34. [Google Scholar]
  53. Lin, H. , and Qiu L., 2013. “Two Sites, Two Voices: Linguistic Differences Between Facebook Status Updates and Tweets.” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 8024 LNCS, 432–40.
  54. Lindner, A. , Hall, M. , Niemeyer, C. , and Caton, S. 2015. “BeWell: A Sentiment Aggregator for Proactive Community Management.” In CHI EA '15 Proceedings of the 33rd Annual ACM Conference Extended Abstracts on Human Factors in Computing Systems. Seoul, Korea: ACM Press, 1055–60.
  55. Liu, B. 2010. “Sentiment Analysis and Subjectivity.” In Handbook of Natural Language Processing, eds. Indurkhya N., and Damerau F.J.. Boca Raton, FL: CRC Press, 1–38. [Google Scholar]
  56. Logan, M. , and Hall M.. 2019. “Comparing Crime Types: A Linguistic Analysis of Communiqués Associated With the Animal and Earth Liberation Movement.” Dynamics of Asymmetric Conflict 12: 161–81. [Google Scholar]
  57. Macri, G. 2015. “Hacker Group ‘Anonymous’ Releases 9,200 ISIS‐Supporting Twitter Accounts.” The Daily Caller (March).
  58. Martin, J. 2013. “Lost on the Silk Road: Online Drug Distribution and the ‘Cryptomarket’.” Criminology & Criminal Justice 14 (3): 351–67. [Google Scholar]
  59. McCallum, A.K. , 2002. MALLET: A Machine Learning for Language Toolkit.
  60. Mirani, T.B. , and Sasi S., 2016. “Sentiment Analysis of ISIS Related Tweets Using Absolute Location.” In 2016 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, 1140–45.
  61. Moule, R. K., Jr , Pyrooz D. C., and Decker S. H.. 2014. “Internet adoption and online behaviour among American street gangs: Integrating gangs and organizational theory.” British Journal of Criminology 54 (6): 1186–1206. [Google Scholar]
  62. Mumford, M.D. , Strange J.M., and Bedell‐Avers K.E.. 2006. “Introduction—Charismatic, Ideological, and Pragmatic Leaders: Are They Really Different? In Pathways to Outstanding Leadership: A Comparative Analysis of Charismatic, Ideological, and Pragmatic Leaders, ed. Mumford M.D.. Mahwah, NJ: Erlbaum Press, 3–24. [Google Scholar]
  63. Mumford, M. D. , Espejo J.,  Hunter S. T., Bedell‐Avers K. E., Eubanks D. L., and Connelly S.. 2007. “The sources of leader violence: A comparison of ideological and non‐ideological leaders.” The Leadership Quarterly 18 (3): 217–235. [Google Scholar]
  64. Nagpal, C. , Miller, K. , Boecking, B. , and Dubrawski, A. 2017. “An Entity Resolution Approach to Isolate Instances of Human Trafficking Online.” arXiv. 1509.06659 [cs. SI].
  65. O’Halloran, K.L. , Tan S., Wignell P., Bateman J.A., Pham D.‐S., Grossman M., and Moere A.Vande. 2016. “Interpreting Text And Image Relations In Violent Extremist Discourse: A Mixed Methods Approach For Big Data Analytics.” Terrorism and Political Violence 6553: 1–21. [Google Scholar]
  66. Panger, G. 2016. “Reassessing the Facebook Experiment: Critical Thinking About the Validity of Big Data Research.” Information, Communication & Society 19 (8): 1108–26. [Google Scholar]
  67. Pelletier, I.R. , Lundmark L., Gardner R., Ligon G.S., and Kilinc R.. 2016. “Why ISIS’ Message Resonates: Leveraging Islam, Socio‐Political Catalysts and Adaptive Messaging.” Studies in Conflict & Terrorism 0731: 1–66. [Google Scholar]
  68. Pennebaker, J.W. 2011. “Using Computer Analyses to Identify Language Style and Aggressive Intent: The Secret Life of Function Words.” Dynamics of Asymmetric Conflict 4 (2): 92–102. [Google Scholar]
  69. Pennebaker, J. , Boyd R.L., Jordan K., and Blackburn K.. 2015. The Development and Psychometric Properties of LIWC2015. Austin, TX: University of Texas at Austin. [Google Scholar]
  70. Pennebaker, J. , Chung C.K., Ireland M., Gonzales A., and Booth R.J.. 2007. The Development and Psychometric Properties of LIWC2. Austin, TX: University of Texas, Austin. [Google Scholar]
  71. Pennebaker, J.W. , Mehl M.R., and Niederhoffer K.G.. 2003. “Psychological Aspects of Natural Language Use: Our Words, Our Selves.” Annual Review of Psychology 54: 547–77. [DOI] [PubMed] [Google Scholar]
  72. Prentice, S. , Taylor P.J., Rayson P., Hoskins A., and O’Loughlin B.. 2011. “Analyzing the Semantic Content and Persuasive Composition of Extremist Media: A Case Study of Texts Produced During the Gaza Conflict.” Information Systems Frontiers 13 (1): 61–73. [Google Scholar]
  73. Pyrooz, D.C. , Decker S.H., and R.K. Moule, Jr. 2013. “Criminal and Routine Activities in Online Settings: Gangs, Offenders, and the Internet.” Justice Quarterly 8825: 1–29. [Google Scholar]
  74. Qin, J. , Zhou Y., Reid E., Lai G., and Chen H.. 2007. “Analyzing Terror Campaigns on the Internet: Technical Sophistication, Content Richness, and Web Interactivity.” International Journal of Human‐Computer Studies 65 (1): 71–84. [Google Scholar]
  75. Radlauer, D. 2007. “Virtual Communities as Pathways to Extremism.” In Hypermedia Seduction for Terrorist Recruiting, eds. Ganor B., Von Knop K., and Duarte C.. Amsterdam: IOS Press, 65–75. [Google Scholar]
  76. Recupero, P.R. 2008. “Forensic Evaluation of Problematic Internet Use.” The Journal of the American Academy of Psychiatry and the Law 36 (4): 505–14. [PubMed] [Google Scholar]
  77. Ruths, D. , and Pfeffer J.. 2014. “Social Media for Large Studies of Behavior.” Science 346 (6213): 1063–64. [DOI] [PubMed] [Google Scholar]
  78. Sabbah, T. , Selamat A., and Selamat H.. 2014. Revealing Terrorism Contents Form Web Page Using Frequency Weighting. Malaysia: Universiti Teknologi Malaysia. [Google Scholar]
  79. Scrivens, R. , Davies G., Frank R., and Mei J., 2016. “Sentiment‐Based Identification of Radical Authors (SIRA).” In Proceedings—5th IEEE International Conference on Data Mining Workshop, ICDMW 2015. IEEE, 979–86.
  80. Simi, P. , and Futrell R.. 2006. “Cyberculture and the Endurance of White Power Activism.” Journal of Political & Military Sociology 34 (1): 115–42. [Google Scholar]
  81. Smith, A.G. , Suedfeld P., Conway L.G., and Winter D.G.. 2008. “The Language of Violence: Distinguishing Terrorist From Nonterrorist Groups by Thematic Content Analysis.” Dynamics of Asymmetric Conflict 1: 142–63. [Google Scholar]
  82. Steinfeldt, J.A , Foltz B.D., Kaladow J.K., Carlson T.N., Pagano L.a, Benton E., and Steinfeldt M.C.. 2010. “Racism in the Electronic Age: Role of Online Forums in Expressing Racial Attitudes About American Indians.” Cultural Diversity & Ethnic Minority Psychology 16 (3): 362–71. [DOI] [PubMed] [Google Scholar]
  83. Tang, S. , Chua T.‐S., Li J., Zhang Y., Xie C., Li M., Liu Y., Hua X., Zheng Y.‐T., and Tang J., 2009. “PornProbe: an LDA‐SVM Based Pornography Detection System.” In Proceedings of the Seventeen ACM International Conference on Multimedia—MM ’09, 1003–4.
  84. Tausczik, Y.R. , and Pennebaker J.W.. 2010. “The Psychological Meaning of Words: LIWC and Computerized Text Analysis Methods.” Journal of Language and Social Psychology 29 (1): 24–54. [Google Scholar]
  85. Thelwall, M. , Buckley K., and Paltoglou G.. 2012. “Sentiment Strength Detection for the Social Web.” Journal of the American Society for Information Science and Technology 63 (1): 163–73. [Google Scholar]
  86. Thelwall, M. , Buckley K., Paltoglou G., Cai D., and Kappas A.. 2010. “Sentiment Strength Detection in Short Informal Text.” Journal of the American Society for Information Science and Technology 61 (12): 2544–58. [Google Scholar]
  87. Thomas, T.L. 2003. “Al Qaeda and the Internet: The Danger of ‘Cyberplanning’.” Parameters: US Army War College 33 (1): 112. [Google Scholar]
  88. Torres‐Soriano, M.R. 2016. “The Caliphate Is Not A Tweet Away: The Social Media Experience of Al Qaeda in The Islamic Maghreb.” Studies In Conflict & Terrorism 39 (11): 968–81. [Google Scholar]
  89. Torres, M.R. , Jordán J., and Horsburgh N.. 2006. “Analysis and Evolution of the Global Jihadist Movement Propaganda.” Terrorism and Political Violence 18: 399–21. [Google Scholar]
  90. Torres Soriano, M.R. 2010. “The Road to Media Jihad: The Propaganda Actions of Al Qaeda in the Islamic Maghreb.” Terrorism and Political Violence 23 (1): 72–88. [Google Scholar]
  91. Ulges, A. , and Stahl A., 2011. “Automatic Detection of Child Pornography Using Color Visual Words.” In 2011 IEEE International Conference on Multimedia and Expo, 1–6.
  92. UNCTED . 2016. Private Sector Engagement in Responding to the Use of the Internet and ICT for Terrorist Purposes.
  93. Weimann, G. 2004. www.terror.net: How Modern Terrorism Uses the Internet. Washington DC: United States Institute of Peace. [Google Scholar]
  94. Wendlandt, L. , Mihalcea, R. , Boyd, R.L. , and Pennebaker, J.W. 2017. “Multimodal Analysis and Prediction of Latent User Dimensions.” In International Conference on Social Informatics. Cham: Springer, 323–40.
  95. Whine, M. 2010. “Cyberspace: A New Medium for Communication, Command, and Control by Extremists.” Studies in Conflict & Terrorism 22 (3): 231–45. [Google Scholar]
  96. Winter, C. 2015. Documenting the Virtual ‘Caliphate’. London: Quilliam, 52. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Additional Supporting Information may be found in the online version of this article at the publisher’s website.

Supporting information


Articles from Policy & Internet are provided here courtesy of Wiley

RESOURCES