Skip to main content
Frontiers in Psychology logoLink to Frontiers in Psychology
. 2020 Sep 30;11:1755. doi: 10.3389/fpsyg.2020.01755

Human Cognition Through the Lens of Social Engineering Cyberattacks

Rosana Montañez 1, Edward Golob 2, Shouhuai Xu 1,*
PMCID: PMC7554349  PMID: 33101096

Abstract

Social engineering cyberattacks are a major threat because they often prelude sophisticated and devastating cyberattacks. Social engineering cyberattacks are a kind of psychological attack that exploits weaknesses in human cognitive functions. Adequate defense against social engineering cyberattacks requires a deeper understanding of what aspects of human cognition are exploited by these cyberattacks, why humans are susceptible to these cyberattacks, and how we can minimize or at least mitigate their damage. These questions have received some amount of attention, but the state-of-the-art understanding is superficial and scattered in the literature. In this paper, we review human cognition through the lens of social engineering cyberattacks. Then, we propose an extended framework of human cognitive functions to accommodate social engineering cyberattacks. We cast existing studies on various aspects of social engineering cyberattacks into the extended framework, while drawing a number of insights that represent the current understanding and shed light on future research directions. The extended framework might inspire future research endeavor toward a new sub-field that can be called Cybersecurity Cognitive Psychology, which tailors or adapts principles of Cognitive Psychology to the cybersecurity domain while embracing new notions and concepts that are unique to the cybersecurity domain.

Keywords: social engineering cyberattacks, cyberattacks, cyberdefenses, human cognition, cognitive psychology, cybersecurity

1. Introduction

Social engineering cyberattacks are a kind of psychological attack that attempts to persuade an individual (i.e., victim) to act as intended by an attacker (Mitnick and Simon, 2003; Anderson, 2008). These attacks exploit weaknesses in human interactions and behavioral/cultural constructs (Indrajit, 2017) and occur in many forms, including phishing, scam, frauds, spams, spear phishing, and social media sock puppets (Stajano and Wilson, 2009; Linvill et al., 2019). For example, in the 2016 U.S. election, attackers used so-called social media sock puppets (also known as Russian Trolls) or fictional identities to influence others' opinions (Linvill et al., 2019). The effectiveness of current security technologies has made social engineering attacks the gateway to exploiting cyber systems. Most sophisticated and devastating cyberattacks often start with social engineering cyberattacks, such as spear phishing, where the attacker gains access into an enterprise network (Hutchins et al., 2011). Indeed, Mitnick and Simon (2003) describe many ways to gain access to secure systems using social engineering cyberattacks. Research in social engineering has mostly focused on understanding and/or detecting the attacks from a technological perspective (e.g., detecting phishing emails by analyzing email contents). However, there is no systematic understanding of the psychological components of these attacks, which perhaps explains why these attacks are so prevalent and successful . This motivates the present study, which aims to systematize human cognition through the lens of social engineering cyberattacks. To the best of our knowledge, this is the first of its kind in filling this void.

1.1. Our Contributions

In this paper, we make the following contributions. First, we advocate treating social engineering cyberattacks as a particular kind of psychological attack. This new perspective may be of independent value, even from a psychological point of view, because it lays a foundation for a field that may be called Cybersecurity Cognitive Psychology, which extends and adapts principles of cognitive psychology to satisfy cybersecurity's needs while embracing new notions and concepts that may be unique to the cybersecurity domain. This approach would pave the way for designing effective defenses against social engineering cyberattacks and assuring that they are built based on psychologically valid assumptions. For example, it may be convenient to assume that individuals are willing to participate in defenses against social engineering cyberattacks or that victims are simply reckless. However, these assumptions are questionable because most social engineering cyberattacks are crafted to trigger subconscious, automatic responses from victims while disguising these attacks as legitimate requests.

Second, as a first step toward the ultimate Cybersecurity Cognitive Psychology, we propose extending the standard framework of human cognition to accommodate social engineering cyberattacks. This framework can accommodate the literature studying various aspects of social engineering cyberattacks. In particular, the framework leads to a quantitative representation for mathematically characterizing persuasion, which is a core concept in the emerging Cybersecurity Cognitive Psychology and is key for understanding behavior in the traditional framework of human cognition. Some of our findings are highlighted as follows: (i) a high cognitive workload, a high degree of stress, a low degree of attentional vigilance, a lack of domain knowledge, and/or a lack of past experience makes one more susceptible to social engineering cyberattacks; (ii) awareness or gender alone does not necessarily reduce one's susceptibility to social engineering cyberattacks; (iii) cultural background does affect one's susceptibility to social engineering cyberattacks; (iv) the more infrequent the social engineering cyberattacks, the higher susceptibility to these attacks; (v) for training to be effective, it should capitalize on high-capacity unconscious processing with the goal of creating a warning system that operates in parallel with the user's conscious focus of attention; (vi) it is currently not clear how personality affects one's susceptibility to social engineering cyberattacks; and (vii) more studies, especially quantitative studies, need to be conducted to draw better and consistent results. In addition to these findings, we propose a range of future research directions, with emphasis on quantifying the effect of model parameters (i.e., victim's short-term cognition factors, long-term cognition factors, long-memory, and attacker effort) on the amount of persuasion experienced by the human target.

1.2. Related Work

To the best of our knowledge, we are the first to systematically explore the psychological foundation of social engineering cyberattacks. As discussed in the main body of the present paper, most prior studies focus on social engineering cyberattack or cyberdefense techniques. For example, Gupta et al. (2016) investigate defenses against phishing attacks; Abass (2018) discusses social engineering cyberattacks and non-technical defenses against them. Few prior studies have an aim that is similar to ours. Salahdine and Kaabouch (2019) review social engineering cyberattacks and mitigation strategies, but they do not discuss factors such as human cognition. Darwish et al. (2012) discuss at a high-level the relationship between human factors, social engineering cyberattacks, and cyberdefenses, but they neither examine what makes an individual susceptible to social engineering cyberattacks nor do they discuss the effect of a victim's psychological and situational conditions (e.g., culture and short-term factors) on the outcome. Pfleeger and Caputo (2012) take a multidisciplinary approach to examine cybersecurity from a Behavioral Science perspective, but they do not offer any systematic framework of looking at human cognition in the context of of social engineering cyberattacks. The lack of studies in social engineering cyberattacks might be associated with these studies involving human subjects. In an academic setting, approval for deceptive studies on human subjects requires consent from all entities involved, including ethics board and IT department (Vishwanath et al., 2011). The nature of the topic might also raise sensitivities among those involved (Jagatic et al., 2007), which can lengthen the process. This can be discouraging for most researchers.

1.3. Paper Outline

Section 2 reviews a basic framework for human cognition (without the presence of social engineering cyberattacks). Section 3 extends the basic framework to accommodate social engineering cyberattacks and systematizes victim's cognition through the lens of social engineering cyberattacks, with future research directions. Section 4 concludes the present paper.

2. Overview of Human Cognition

In this section, we review human cognition functions prior to the presence of social engineering cyberattacks. This framework of human cognition serves as a basis for exploring how victims' cognition functions are exploited to wage social engineering cyberattacks.

2.1. Human Cognitive Functions

The term “cognition” can have radically different meanings in different contexts. Here, we use the term “cognition” in the broadest sense as a descriptive term for the software counterpart to the brain as hardware. That is, cognition is the abstract information processing that is being implemented by neurons in the brain (Pinker, 2009). From this perspective, cognition can also include information processing that computes emotions as well as the vast majority of neural information processing that is not reflected in our conscious awareness (Baars, 1997).

Cognitive psychologists often consider information processing to be the basic function of the brain, in the same way that the liver functions as a complex filter and arteries and veins are essentially pipes. Correlates of information processing in the brain can be directly observed using various methods to record electrical and chemical activity (Kandel et al., 2000). Information processing is evident at multiple spatial, from compartments within individual neurons to tightly organized networks having members in different parts of the brain. These concrete, physically measurable, neurophysiological activities are analogous to the hardware of a computer. Indeed, neurons have been profitably studied in terms of functioning as Boolean logic gates, and action potentials, perhaps the most characteristic property of neurons, is convey all-or-none binary states (Shepherd, 2004).

Figure 1 presents a very basic, and selective, schematic of human cognition functions, which are centered at four information processing components analogous to software components in an information processing system. These four components are called perception, working memory, decision making, and action. These four components are elaborated below as follows. Perception converts information in the word, sampled from the senses, into neural codes that can be used for intelligent behavior and conscious experience (Mesulam, 1998). Working memory consists of attention and short-term memory, and coordinates processing information by prioritizing certain information for short periods of time, often to accomplish a goal (Miyake and Shah, 1999). Decision making further prioritizes information from working memory and other unconscious sources and is a gateway to behavior (Kahneman, 2011). Action is the implementation of computations from decision making, as well as other influences, and also organizes the physical activity of muscles and glands that are measurable as behavior (Franklin and Wolpert, 2011). Perception, working memory, decision making, and action are often considered to be roughly sequential, as when trying to hit a baseball, but can mutually influence each other in many ways. All of these cognitive processes operate on a foundation of accumulated knowledge in memory, which informs these processes, such as when perceiving a familiar face.

Figure 1.

Figure 1

A basic, selective schema of human cognition, where the blue background within the oval indicates long-term memory that is accessible by the four components of human cognition functions.

Memory is intrinsic to cognition, because information processing occurs over time and thus requires some information to be retained over time. The basic processes of perception, working memory, decision making, and action that are engaged “in the moment” use information that is preserved from earlier moments in time. Memory consists of distinct systems (Tulving and Craik, 2000), in the same way that our domain of “perception” includes the visual, auditory, somatosensory, olfactory, gustatory, and vestibular systems. One important distinction among systems is whether the information is retained over short periods of time, typically seconds to minutes, or longer periods of time. In our overview shown in Figure 1, short-term memory is a component of working memory. Long-term memory contributes to cognition in general, and for this reason, we have situated all of four domains supporting cognition in the moment within long-term memory (indicated by the blue background). As with the other cognitive domains, memory systems can work in parallel. For example, the memory of the previous sentence is supported by short-term memory, yet the memory for what each word means resides in long-term memory.

Above, we presented several basic types of information processing that together generate behavior. We now consider how these basic cognitive processes can be influenced, for better or worse, by a few important factors that are demonstrably relevant to cybersecurity. The “short-term” factors, reflecting the immediate situation, and “long-term” factors are ultimately coded in some form by the brain and exert an influence on the basic cognitive processes that drive behavior. The short-term and long-term factors are elaborated in the next two subsections.

2.2. Short-Term Cognitive Factors

We focus on three short-term factors: workload, stress, and vigilance. These factors operate on relatively short timescales (minutes to hours) that have been intensively studied because they impair human performance. We will consider how these factors may relate to social engineering and point out the extant literature and promising future directions.

2.2.1. Workload

Human cognition is affected by cognitive workload, which depends on task demand and the operator in question. Depending on the details, two tasks can be done at the same time with little or no performance costs (a manageable workload) or be nearly impossible to do well together (a very high workload). A nice example comes from Shaffer (1975), who found that typists could very accurately read and type while they also verbally repeated a spoken message. Performance, however, plummeted on both tasks if they tried to take dictation (typing a spoken message) while also trying to read a written message out loud. The differences are thought to reflect the use of phonological (sound-based) and orthographic (visual letter-based) cognitive codes. In the first example one code is used per task (phonological: listen to speech-talk; orthographic: read-type), while in the second each code is used for both tasks (speech-type; read-talk). To account for these complexities, psychologists have developed theories that consider different types of cognitive codes (Navon and Gopher, 1979), such as auditory or visual sensory input, higher-level verbal or spatial codes, and output codes for driving speech or manual behaviors (Wickens, 2008). Measures have also been developed to quantify the subjective sense of how much “cognitive work” is being done in a given task. Perhaps the most common instrument to measure subjective workload is the NASA-TLX, which has six dimensions that are clearly explained to the subject, such as “mental demand” or temporal demand (time pressure), and are rated on a scale from low to high. Lastly, neurophysiological measures are often used to provide objective, convergent measures of workload as well as suggest potential neural mechanisms. Neurophysiological measures such as transcranial Doppler measures of blood flow velocity in the brain, EEG measures of brain electrical potentials, autonomic nervous system activity such as skin conductance and heart rate and its variability, and functional magnetic resonance imaging (MRI) to quantify changes in blood flow that are secondary to neural activity are commonly used (Parasuraman and Rizzo, 2008).

2.2.2. Stress

Acute stress may also influence cognition and behaviors that are relevant to cybersecurity. We distinguish acute from chronic stress, with chronic stress beginning after a duration on the order of months, as their impact on cognition can differ and chronic stress is better classified here as a long-term factor. The neurobiological and hormonal responses to a stressful event have been well-studied, as have their impact on behavior (Lupien et al., 2009). Acute stress can influence attention, a vital component to working memory, in ways that are beneficial as well as detrimental (Al'Absi et al., 2002). Attentional tunneling is one such effect of acute stress where attention is hyper-focused on aspects relevant to the cause of the stress but is less sensitive to other information. The term tunneling derives from the use of spatial attention tasks, where arousal due to stress leads to subjects ignoring things that are more distant from the focus of attention (Mather and Sutherland, 2011). In the realm of cyber security, attention tunneling from an emotion-charged phishing message could lead one to hyper-focusing on the email text but ignore a suspicious address or warnings at the periphery. Working memory is also vulnerable to acute stress (Schwabe and Wolf, 2013), particularly by way of interfering with prefrontal cortex function (Elzinga and Roelofs, 2005; Arnsten, 2009). Decision making can be driven in two fundamentally different ways (Evans, 2008). The first is by relatively automatic processes that are fast but may not be the optimal choice in some instances (termed “heuristics” and “biases”) (Tversky and Kahneman, 1974; Gigerenzer, 2008). The second approach is by using conscious, controlled processing reasoning, which is slower but can be more sensitive to the particulars of a given situation. Acute stress has a variety of effects on decision making and many subtleties (Starcke and Brand, 2012), but it can, in general, impair rational decision making, and one way is by reducing the likelihood of controlled decision making and increasing the use of automatic processing.

2.2.3. Vigilance

Vigilance and sustained attention are two closely related, sometimes synonymous, terms for the concept that cognitive performance will systematically change the longer you perform a given task. Here we will use the term vigilance, which in the laboratory is studied in sessions that typically last 30–60 min. In a classic work by Mackworth (1948), subject watched an analog clock and responded to intermittent jumps in the clock hand. Much work since then has showed that performance in a wide range of tasks declines substantially over these relatively short periods of time (termed the “vigilance decrement”) (Parasuraman and Rizzo, 2008). In our view, the potential impact of the vigilance decrement on behavior is an important factor to explore, because the probability of user error may covary with time on task. For example, the likelihood of downloading malware may increase as users go through their email inbox, particularly if they have limited time. Lastly, we note that although the situational categories of workload, stress, and vigilance are individually important to examine in the realm of cybersecurity, they are also known to interact with each other. For example, a high workload and prolonged vigilance are stressful (Parasuraman and Rizzo, 2008). Another distinction to keep in mind is that many laboratory vigilance tasks are boring and have a low workload. The extent that the vigilance literature generalizes to other settings such as an office, where workers may have high workloads and stress from complex job demands, is an empirical question worth considering in future cybersecurity studies.

2.3. Long Term Cognitive Factors

In contrast to short-term factors that reflect the current situation and can change rapidly, our second grouping of “long-term factors” covers more stable attributes of a person and their experiences that only gradually change. We consider factors of personality, expertise, age and gender, and culture. We include personality as a long-term factor, even though it can be situation dependent as well (as with short-term factors) (Kenrick and Funder, 1988). These factors offer some predictability of individual behavior in a given situation. In the context of cybersecurity, long-term psychological factors can impact how an individual responds to social engineering attacks.

2.3.1. Personality

To Psychologists, “personality” is a technical term that differs somewhat from ordinary usage. It refers to individual differences in thoughts, feelings, and behaviors that are relatively consistent over time and situations. We say “relatively” because, as noted above, thoughts, feelings, and behaviors are highly dependent on the situation, and lifespan approaches have defined notable changes in personality with age (Donnellan and Robins, 2009). Personality research is dominated by the Big 5 framework of personality domains, which was developed over much of the twentieth century in various forms (Digman, 1997). The Big 5 framework is based on statistical methods (factor analysis) that identify abstract dimensions that can economically account for much of the variance in personality measures. The factors are labeled conscientiousness, agreeableness, neuroticism, openness to experience, and extraversion. For present purposes, the labels of the factors are adequate descriptions of the underlying constructs. Many studies on the relationship between social engineering and personality focus on openness, conscientiousness, and neuroticism which are thought to have the most impact on susceptibility to social engineering. The factors that comprise the Big 5 framework are the following:

  1. Openness: the willingness to experience new things.

  2. Conscientiousness: favors-norms, exhibiting self-control and self-discipline, and competence.

  3. Extraversion: being more friendly, outgoing, and interactive with more people.

  4. Agreeableness: being cooperative, eager to help others, and believing in reciprocity.

  5. Neuroticism: tendency to experience negative feelings, guilt, disgust, anger, fear, and sadness.

2.3.2. Expertise

Expertise is typically limited to relatively narrow domain and does not transfer to other areas as much as we tend to believe (termed the “transfer problem”) (Kimball and Holyoak, 2000). Limited transfer of expertise can be compounded by cognitive illusions such as the Dunning-Kruger effect. The Dunning-Kruger effect empirically shows that individuals often overestimate their competence relative to their objective performance (Kruger and Dunning, 1999). Similarly, the “illusion of knowledge” shows that people generally know far less about a topic than they believe, as revealed by questioning (Keil, 2003). In the realm of cybersecurity, these and other empirical phenomena underpin user over confidence. As will be detailed below, narrow expertise about cybersecurity can be beneficial, but computer expertise more generally may not confer security benefits.

2.3.3. Individual Differences

There are many kinds of individual differences and we focus on two kinds: age and sex/gender; others would include role in companies and seniority. In terms of age, there are dramatic changes in cognitive function and behavioral capacities of children as they develop (Damon et al., 2006). Considering how youths can safely use computers is a major parenting, education, public policy, and law enforcement challenge. Social engineering attacks can readily take advantage of the cognitive and emotional vulnerabilities of children, and countermeasures are often quite different than with adults (see below). Cognition changes throughout the adult lifespan at a less frenetic pace vs. in children, but longer-term changes are similarly dramatic (Park and Reuter-Lorenz, 2009; Salthouse, 2012). Declines in fluid intelligence, essentially one's ability to “think on your feet,” are particularly dramatic and have wide implications for everyday life (Horn and Cattell, 1967). Overall, there are many changes, some declining with age (fluid intelligence) and others not (Schaie, 2005). Another angle is that age is positively associated with the risk for many neurological disorders that can impair cognition, such as stroke and Alzheimer's disease (Hof and Mobbs, 2001). Age-related neurological disorders are not considered “normal aging,” but the potential vulnerability of many elders due to brain disease has been well-known to criminals for a long time. As expected, social engineering attacks are a major problem for this vulnerable population.

Psychology has a long history of studying sex differences, defined by biology (i.e., the presence of two X or one X and one Y chromosome) and gender, which is a social, rather than biological, construct. In terms of basic cognitive functions such as working memory and decision making, which are typically studied in a neutral laboratory context (such as remembering strings of letters, judging categories of pictures, etc.) there are generally little or no differences between sexes and genders. There are a few well-documented exceptions, such as males having an advantage for mental spatial rotations (Voyer et al., 1995). The situation is quite different when examining cognition in the context of social and emotional factors (Cahill, 2006). For our purposes, sex and gender are basic considerations for social engineering attacks, particularly spear phishing, which is tailored to an individual. Our list could include many other types of individual differences that are useful for social engineering attacks, such as socio-economic class, education, personal interests, job position. We chose to focus on age and sex/gender because they are prominent topics in the cognition literature and important considerations for cyber security challenges such as spear phishing.

2.3.4. Culture

In mainstream cognitive psychology, culture is not a prominent variable, as much of the basic literature studies participants in countries that have predominantly western cultures (Arnett, 2008). Nonetheless, a wide variety of studies have shown that cultural differences are evident in many aspects of cognition, such as basic perception, language and thought, attention, and reasoning (Grandstrand, 2013). Culture is an important variable to consider for any social engineering attack. A phishing email, for example, is unlikely to be effective if the message violates norms of the target's culture. We also consider the more specific case of organizational culture in the workplace because it is highly relevant to employee behavior as it applies to cyber security (Bullee et al., 2017). As with all of the other short-and long-term variables that we consider, culture is assumed to interact with the other variables, with particularly large interactions with age, gender, and perhaps personality.

3. Victim Cognition Through the Lens of Social Engineering Cyberattacks

Social engineering cyberattacks are a type of psychological attack that exploits human cognition functions to persuade an individual (i.e., victim) to comply with an attacker's request (Anderson, 2008). These attacks are centered around a social engineering message crafted by an attacker with the intent of persuading a victim to act as desired by the attacker. These attacks often leverage behavioral and cultural constructs to manipulate a victim into making a decision based on satisfaction (gratification), rather than based on the best result (optimization) (Kahneman, 2011; Indrajit, 2017). For example, one behavioral construct is that most individuals would trade privacy for convenience, or bargain release of information for a reward (Acquisti and Grossklags, 2005).

To establish a systematic understanding of the victim's cognition through the lens of social engineering cyberattacks, we propose extending the framework presented in Figure 1 to accommodate social engineering cyberattacks against human victims' cognition functions, leading to the framework highlighted in Figure 2. This implies that the resulting behavior of a victim will also depend on the attacker's effort. In what follows, we will cast the social engineering cyberattacks literature into this framework, by first discussing the literature related to short-term and long-term cognition factors, and then the literature related to cognition functions.

Figure 2.

Figure 2

Extending the basic schema of human cognition presented in Figure 1 to accommodate social engineering cyberattacks. The extension is to incorporate an attacker that wages a social engineering cyberattack against a victim's human cognition functions (i.e., the oval). The resulting behavior is associated to persuasion (i.e., an attack succeeds when a victim is persuaded to act as intended by the attacker).

3.1. Short-Term Cognition Factors Through the Lens of Social Engineering Cyberattacks

3.1.1. Workload

In computer-mediated communications, cognitive workload can affect an individual's ability to process socially engineered messages. Pfleeger and Caputo (2012) observe that cognitive workload could make individuals overlook elements that are not associated with the primary task. This effect, called inattentional blindness, affects an individual's ability to notice unexpected events when focusing on the primary task (Simons, 2000). In most cases, security is a secondary task. For example, when an employee attempts to manage several tasks simultaneously (e.g., reply to hundreds of emails in the email inbox while answering calls and an occasional request from the boss), the employee is more likely to overlook cues in phishing messages that might indicate deception. A study that examined actual phishing behavior by sending employees an innocuous phishing email, found that self-perceived work overload was positively associated with the likelihood of clicking on the phishing link (Jalali et al., 2020). Vishwanath et al. (2011) investigate the effect of information processing and user's vulnerability to phishing. Leveraging two phishing attacks that target a university, they survey undergraduate students on their recollection and response to the phishing emails. They find that in the presence of a perceived relevant email, individuals focus more on urgency cues, while overlooking deception cues in the message, such as sender's email address or email grammar/spelling. They also find that individuals that regularly manage large volumes of emails have a high inattentiveness when evaluating emails, making them more vulnerable to phishing attacks. They also find that a high email load triggers an automatic response, meaning that workload significantly increases a victim's vulnerability to phishing attacks.

Summarizing the preceding discussion, we draw:

Insight 1. Cognitive workload, via mechanisms such as inattentional blindness, can increase vulnerability to social engineering cyberattacks.

3.1.2. Stress

The particular kind of stress, namely acute stress mentioned above, has only been indirectly investigated in the context of social engineering cyberattacks. Stajano and Wilson (2009) examine how principles of scams apply to systems security. Scams are a form of social engineering cyberattack that usually involves a physical interaction between attacker and victim. One scamming technique is the principle of distraction, by which the attacker can take advantage of a victim that is in a state of mind that prevents them from evaluating deceptive cues. For example, when an unemployed individual pays a job recruiting company for job hunting assistance, the individual does not realize that it is a scam. Catphishing is a social engineering cyberattack by which the attacker creates a fictional online persona to lure a victim into a romantic relationship for financial gains. In this case, an individual who is searching for a romantic partner and is experiencing some personal stress might find a catphishing message appealing and, therefore, may be unable to detect the deception cues in the catphishing messages. In summary, we draw the following:

Insight 2. Stress may reduce one's ability to detect deception cues in social engineering cyberattack messages but the direct effects of acute stress on cybersecurity social engineering have not been examined.

3.1.3. Vigilance

Purkait et al. (2014) conduct a study to examine cognitive and behavioral factors that affect user's capabilities in detecting phishing messages. They study attentional vigilance and short-term memory by surveying 621 participants' ability to identify phishing sites, Internet skills, usage and safe practices, and demographics. The measure of “vigilance” was a brief visual search task in six photographs, which did not evaluate vigilance as we conventionally defined it above. Individual differences in these visual search scores were significant predictors of performance distinguishing spam from phishing websites, which likely reflects the ability to detect visual cues on the website that distinguish spam from phish sites.

Insight 3. Attentional vigilance, particularly the vigilance decrement, may be an important influence on susceptibility to social engineering attacks, but more research is needed.

3.2. Long-Term Cognition Factors Through the Lens of Social Engineering Cyberattacks

3.2.1. Personality

Personality has been extensively studied in the context of phishing. Studies show that Big 5 personality traits are related to individuals' susceptibility to social engineering cyberattacks. Pattinson et al. (2012) study how personality traits and impulsiveness affect behavioral responses to phishing messages. They find that individuals that score high on extraversion and openness manage phishing emails better than individuals with other personality types. Halevi et al. (2013) find that high neuroticism increases responses to prize phishing messages and that individuals with a high openness have low security setting on social media account, increasing their exposures to privacy attacks. Halevi et al. (2016) find that personality traits affect security attitudes and behaviors as follows: high conscientiousness is associated to highly secure behaviors but does not affect self-efficacy (i.e., one's ability in independently resolving computer security issues); high openness is associated to high self-efficacy; high neuroticism is associated to low self-efficacy; and high emotional stability (inverse of neuroticism) is associated to high self-efficacy. Cho et al. (2016) contradict some of the findings presented in Halevi et al. (2013), by finding that high neuroticism decreases trust and increases risk perception, which makes one more likely to misclassify benign emails as phishing ones. They also find that higher agreeableness increases trust and lowers risk perception (i.e., more likely classifying phishing messages as benign). Consciousness is commonly associated with self-control, which diminishes impulsive behavior (Cho et al., 2016). Pattinson et al. (2012) find that less impulsive individuals manage phishing messages better. Halevi et al. (2015) show that individuals with high consciousness and lower risk perception are more likely to fall victims to social engineering cyberattack messages. Lawson et al. (2018) find that extroversion decreases phishing detection accuracy while high consciousness increases detection accuracy, and that openness is associated with higher accuracy in detecting legitimate messages. Darwish et al. (2012) find that individuals high in extraversion and agreeableness pose a higher security risk. McBride et al. (2012) find that consciousness is associated with low self-efficacy and threat severity. Workman (2008) and Lawson et al. (2018) that personality traits are related to the degree of persuasion by social engineering cyberattacks. Summarizing the preceding discussion, we draw the following:

Insight 4. Literature results are not conclusive on how personality may influence one's susceptibility to social engineering cyberattacks.

3.2.2. Expertise

Related to expertise, domain knowledge, awareness, and experience have been studied in the literature on their impact on reducing one's susceptibility to social engineering cyberattacks.

Impact of domain knowledge. An individual's knowledge related to cyberattacks increases their capability to resist social engineering cyberattacks. For example, the knowledge can be about web browsers, including how to view site information and evaluate certificates. Kumaraguru et al. (2006) find (i) non-expert individuals consider fewer security indicators (e.g., meaningful signals) than experts; (ii) non-expert individuals used simple rules to determine the legitimacy of a request, while experts also consider other useful information (e.g., context) that may reveal security concerns with the request; (iii) non-expert individuals make decisions based on their emotions, while experts make their decisions based on reasoning; and (iv) non-expert individuals rely more on (spoofable) visual elements to make decisions because they lack the knowledge that security indicators can be compromised, while experts are more efficient at identifying suspicious elements in a message. For example, corresponding to (iii), they observe that a non-expert individual might decide to download a software program based on how much they want it and if the downloading website is recognizable; whereas an expert might consider how much they need it and if the downloading website is a reputable source. These findings resonate with what is found by Klein and Calderwood (1991), namely, that experts make decisions based on pattern recognition, rather than purely analyzing the available options. Byrne et al. (2016) find that risk perception for non-expert individuals is influenced by the benefit that can be obtained for an activity, meaning that actions that an individual considers beneficial are performed more often and are perceived as less risky.

Insight 5. Domain knowledge helps reduce vulnerability to social engineering cyberattacks.

Impact of awareness. As a rule of thumb, training on non-expert individuals often emphasize on awareness. In a study of victims in frauds involving phishing and malware incidents, Jansen and Leukfeldt (2016) find that most participants express that they have knowledge of cybersecurity, but it turns out only a few of them indeed have the claimed knowledge. Downs et al. (2006) find that awareness of security cues in phishing messages does not translate into secure behaviors because most participants are unable to tie their actions to detrimental consequences. On the other hand, it may be intuitive that individuals that have received formal computer education would be less vulnerable to social engineering cyberattacks. To the contrary, Ovelgönne et al. (2017) find that software developers are involved in more cyberattack incidents when compared to others. Purkait et al. (2014) find that there is no relationship between one's ability to detect phishing sites and one's education and technical backgrounds, Internet skills, and hours spent online. Halevi et al. (2013), Junger et al. (2017), and Sheng et al. (2010) find that knowledge acquired through priming and warning does not affect ones' susceptibility to phishing attacks.

Insight 6. Awareness and general technical knowledge do not necessarily reduce one's susceptibility to social engineering cyberattacks, perhaps because human cognition functions have not been taken into consideration.

Impact of experience. Harrison et al. (2016) find that knowledge about phishing attacks increases one's attention and elaboration when combined with subjective knowledge and experience, and therefore lowers one's susceptibility to fall victim to social engineering cyberattack messages. Wang et al. (2012) find that knowledge about phishing attacks increases one's attention to detect indicators. Pattinson et al. (2012) find that the higher familiarity with computers, the higher capability in coping with phishing messages. Wright and Marett (2010) find (i) a combination of knowledge and training is effective against phishing attacks; (ii) individuals with a lower self-efficacy (i.e., one's ability to manage unexpected events) and web experience are more likely to fall victims to social engineering cyberattacks; and (iii) individuals with high self-efficacy are less likely to comply with information requests presented in phishing attacks. Halevi et al. (2016) find that a high self-efficacy correlates a better capability to respond to security incidents. Arachchilage and Love (2014) find that self-efficacy, when combined with knowledge about phishing attacks, can lead to effective strategies for coping with phishing attacks. Wright and Marett (2010) find that experiential factors (e.g., self-efficacy, knowledge, and web experience) have a bigger effect on individuals' response to phishing attacks than dispositional factors (e.g., the disposition to trust and risk perception). Van Schaik et al. (2017) find that a higher risk perception of online threats is associated with exposure to the knowledge that is specific to the threat. Downs et al. (2006) find that users can detect social engineering cyberattacks that are similar to the ones they have been exposed to. Redmiles et al. (2018) find that the more time an individual spends online, the more skilled they are at identifying spams, and the less likely they will click on the links in the spam messages. Gavett et al. (2017) find that education and previous experience with phishing attacks increased suspicion on phishing sites. Cain et al. (2018) find that past security incidents do not significantly affect secure behaviors. Abbasi et al. (2016) find (i) older, educated females and males fell victim to phishing attacks in the past are less likely to fall victim to phishing attacks again; (ii) young females with low phishing awareness and previous experience in suffering from small losses caused by phishing attacks do not necessarily have a lower susceptibility to phishing attacks in the future; and (iii) young males with high self-efficacy and phishing awareness and previous experiences in phishing attacks also do not necessarily have a lower susceptibility to phishing attacks in the future.

Insight 7. Self-efficacy, knowledge, and previous encounter of social engineering cyberattacks collectively reduce one's susceptibility to social engineering cyberattacks. In particular, costly phishing experiences would greatly reduce one's susceptibility to social engineering cyberattacks, while non-costly experiences do not.

3.2.3. Individual Differences

Two kinds of individual differences have been investigated in the context of social engineering cyberattacks: gender and age.

Impact of gender. Initial studies suggest a relationship between gender and phishing susceptibility. Hong et al. (2013) finds that individual differences (e.g., dispositional trust, personality, and gender) are associated with the ability to detect phishing emails. Halevi et al. (2015) find that for women, there is a positive correlation between conscientiousness and clicking on links and downloading files associated with phishing attacks. Halevi et al. (2013) find that women exhibit a strong correlation between neurotic personality traits and susceptibility to phishing attacks, but no correlation to any personality trait is found for men. Halevi et al. (2016) reports that women exhibit lower self-efficacy than men. Sheng et al. (2010) find that women with low technical knowledge are more likely to fall victim to phishing attacks. Sheng et al. (2010) find that women are more likely to fall victim to phishing attacks.

However, later studies provide a different view. Sawyer and Hancock (2018) finds that there is no relationship between gender and phishing detection accuracy. Similarly, Purkait et al. (2014) find that there is no relationship between gender and the ability to detect phishing sites. Byrne et al. (2016) finds that there is no relationship between gender and risk perception. Rocha Flores et al. (2014) finds that there is no significant correlation between phishing resiliency and gender. Bullee et al. (2017) finds that gender does not contribute to phishing message responses. Abbasi et al. (2016) finds (i) women with a high self-efficacy have a low susceptibility to social engineering cyberattacks, and that women without awareness of the social engineering cyberattack threat have a high susceptibility to these attacks; and (ii) men with previous costly experiences with phishing attacks have a low susceptibility to these attacks, while overconfidence increases the susceptibility to these attacks. Cain et al. (2018) find that although men may have more knowledge about cybersecurity than women, there is no difference in terms of insecure behaviors by gender. Redmiles et al. (2018) show that, in the context of social media spam, gender affects message appeal but not susceptibility to social engineering cyberattacks, and that women are more likely to click on sales-oriented spams while men are more likely to click on media spams that feature pornography and violence. Goel et al. (2017) find that women open more messages on prize reward and course registration than men. Rocha Flores et al. (2014) find that gender affects the type of phishing message an individual would respond to and that women are less susceptible than men to generic phishing messages.

Insight 8. Gender does not have a big impact on the susceptibility to social engineering cyberattacks.

Impact of age. Most studies focus on age groups in young people (18–24) and old (45+) ones. In general, youth is related to inexperience, high emotional volatility (Zhou et al., 2016), less education, less exposure to information on social engineering, and fewer years of experience with the Internet. These factors are often accompanied by a low aversion to risk and therefore can increase the chances of falling victim to social engineering cyberattacks (Sheng et al., 2010). In an experiment involving 53 undergraduate students in the age group of 18–27, Hong et al. (2013) find that the students' confidence in their ability to detect phishing messages does not correlate to their detection rate. Sheng et al. (2010) investigate the relationship between demographics and susceptibility to phishing attacks and find that individuals at the age group of 18–25 are more susceptible to phishing attacks than other groups 25+. Lin et al. (2019) report a similar result but for an old group. Howe et al. (2012) find that age also affects risk perception: individuals in the age groups of 18–24 and 50–64 perceive themselves at lower security risks compared to other groups and therefore are more susceptible to social engineering cyberattacks. Purkait et al. (2014) find that the detection of phishing messages decreases with the age and frequency of online transactions. Bullee et al. (2017) find that age has no effect on spear-phishing responses and that Years of Service (YoS) is a better indicator of victimization (i.e., a greater YoS means less likely susceptible to social engineering cyberattacks). Gavett et al. (2017) examine the effect of age on phishing susceptibility and showed that processing speed and planning executive functions affect phishing suspicion, hinting a relationship between phishing susceptibility and cognitive degradation from aging.

Insight 9. Old people with higher education, higher awareness and higher exposure to social engineering cyberattacks are less susceptible to these attacks.

3.2.4. Culture

Culture affects individuals' online activities (Sample et al., 2018), decision making process and uncertainty assessment (Chu et al., 1999), development of biases and risk perception (Pfleeger and Caputo, 2012; da Veiga and Martins, 2017), reactions to events (Hofstede et al., 2010; Rocha Flores et al., 2014), and self-efficiency (Sheng et al., 2010; Halevi et al., 2016). Redmiles et al. (2018) suggest that country/communal norms might affect spam consumption as follows: in countries where spam is more prevalent, users are 59% less likely to click on spam when compared to countries where spam is less prevalent. Halevi et al. (2016) find that individuals with higher risk perception have higher privacy attitudes, which reduce the susceptibility to social engineering cyberattacks. Al-Hamar et al. (2010) perform experimental spear-phishing attacks against two groups from Qatar, where one group consists of 129 employees of a company (dubbed employees) and the other of 30 personal acquaintances (dubbed friends); they find that find that 44% of the individuals in the employees group are successfully phished while 57% of the friends groups are successfully phished. Tembe et al. (2014) report that participants from India exhibit a higher susceptibility to phishing attacks when compared with participants from the USA and China. (Bullee et al., 2017) report that participants from China and India might not be aware of the harms and consequences of phishing attacks, while participants from the USA exhibit more awareness of privacy and online security features (i.e., SSL padlocks) and are more active in safeguarding their personal information. Halevi et al. (2016) find that although culture is a significant predictor of privacy attitude, it does not predict security behavior and self-efficacy. Bohm (2011) finds that culturally sound messages do not raise suspicion. Farhat (2017) and Hofstede et al. (2010) show that scams with culture-specific shame appeal are more likely to be effective in a certain culture. Bullee et al. (2017) find that participants from countries with a higher Power Distance Index (PDI), which means that individuals are more likely to comply with hierarchy, are more vulnerable to phishing than those individuals from countries with a lower PDI. Sharevski et al. (2019) show how to leverage cultural factors to tailor message appeal.

Insight 10. Culture affects privacy and trust attitudes, which indirectly affect one's susceptibility to social engineering cyberattacks.

3.3. Victim Cognition Functions Through the Lens of Social Engineering Cyberattacks

3.3.1. Long-Term Memory

As reviewed in section 2.1, long-term memory is a very broad field, of which the following aspects have been studied through the lens of social engineering cyberattacks. The first aspect is the frequency of attacks. The environment in which a victim operates provides a context that may be exploited by an attacker. For example, the attacker may leverage an ongoing societal incident or personal information to craft messages to make the victim trust these messages. The attacker attempts to build trust with a victim while noting that a suspicion thwarts the attack (Vishwanath et al., 2018). Both trust and suspicion are affected by the environment, such as the frequency of the social engineering events exploited by the social engineered messages. For example, in a situation where social engineering cyberattacks are expected, the attacker is at a disadvantage (Redmiles et al., 2018); in a situation where social engineering cyberattacks are infrequent, the attacker has the advantage. Sawyer and Hancock (2018) investigate how infrequent occurrence of phishing events (i.e., the prevalence of phishing) affects individuals' abilities to detect cyberattacks delivered over emails. In their experiment, they ask three groups to identify malicious and legitimate email messages. The three groups, respectively, contain 20, 5, and 1% malicious emails. They find that the accuracy of the detection of malicious emails is lower for the group dealing with emails that contain 1% malicious ones. Similarly, Kaivanto (Kaivanto, 2014) show that a lower probability of phishing occurrence increases victim's susceptibility to phishing cyberattacks.

Insight 11. The success of social engineering cyberattacks is inversely related to their prevalence.

Insight 11 causes a dilemma: when automated defenses are effective at detecting and filtering most social engineering cyberattacks, the remaining attacks that make it through to users are more likely to succeed. One approach to dealing with this dilemma is to resort to principles in Cognitive Psychology. It is known that most of the brain's information processing is sealed-off from conscious awareness (Nisbett and Wilson, 1977), some permanently while other information could be consciously appreciated, but may not be conscious at a given moment. Our visual system, for example, computes 3-D depth from 2-D retinal inputs (DeValois and DeValois, 1990). We do not consciously experience the calculations needed to transform the 2-D input into a 3-D percept. Instead, we are aware of the product (i.e., seeing a 3-D world) but not the process that led to the product. The influences of subconscious processing are well-known to impact behavior (Kahneman, 2011; Nosek et al., 2011). This fact leads to the following insight:

Insight 12. Training methods that ask people to consciously think about social engineering cyberattacks are unlikely to be very successful unless the learning reaches the point where it is a habit that, largely unconsciously, guides safer computer use behavior.

Insight 12 would avoid the dilemma mentioned above because when the training/learning effort reaches the point that users can deal with social engineering cyberattacks subconsciously, users can effectively defend these attacks. This coincides with findings of Halevi et al. (2015), Rocha Flores et al. (2014), Halevi et al. (2016), Howe et al. (2012), and Sheng et al. (2010), namely that users with higher risk perception can reduce the chance they fall victim to social engineering cyberattacks.

3.3.2. Victim Cognition Functions: A Preliminary Mathematical Representation

The framework described in Figure 2 formulates a way of thinking in modeling how the behavior of a victim is influenced by the victim's short-term cognition factors (or short_term factors), long-term cognition factors (or long_term factors) and long-memory (or long_memory) as well as the attacker's effort (or attacker_effort). This formulation is applicable to phishing, spear phishing, whaling, water holing, scams, angler phishing, and other kinds of social engineering attacks, where the resulting behavior is whether a victim is persuaded by the attacker to act as intended. For example, spear-phishing is a special case of the model because the attacker often makes a big effort at enhancing message appeal by exploiting personalization; scam is another special case of the model because the attacker often makes a big effort at enhancing message appeal by exploiting situational setting and possibly stress. In principle, the behavior (behavior) of social engineering cyberattacks can be described as some mathematical function f (mathematically speaking, more likely it will be a family of functions):

behavior=f(short_term_factors,         long_term_factors,long_memory,         attacker_effort). (1)

Note that f mathematically abstracts and represents the interactions between the four cognitive domains operating in long-memory (i.e., perception, working memory, decision making, and action), while also taking short-term and long-term factors into account. Moreover, f accommodates attacker's effort as input. It is an outstanding research problem to identify the appropriate abstractions for representing these model parameters and what kinds of mathematical functions f are appropriate to what kinds of social engineering attacks. These questions need to be answered using experimental studies. Note that the framework can be expanded to include measures of brain activity, either direct measures such as electroencephalography, or indirectly using peripheral measures such as eye tracking and autonomic nervous system activity (Valecha et al., 2020).

Specific to the cybersecurity domain, we propose considering persuasion-related behavior, as shown in Figure 2 and the corresponding mathematical representation of Equation (1), meaning that the outcome in Equation (1) can be replaced by the probability that a user is persuaded by the attacker to act as the attacker intended. Intuitively, persuasion is the act of causing someone to change their attitudes, beliefs, or values based on reasoning or argument. Wright et al. (2014) defines Cialdini's Principles of Persuasion, which have been extensively used to study the response to social engineering messages (but not social engineering cyberattacks). Table 1 presents a brief summary of Cialdini's Principles of Persuasion.

Table 1.

Summary of Cialdini Principles of Persuasion.

Principle Description
Liking The act of saying yes to something you know and like; for example, a social engineer presenting himself as helpful and empathetic toward the victim in a password reset process.
Reciprocity Repaying an earlier action in kind; for example, conveying to a victim that they have detected suspicious activities in the victim's credit card account while encouraging the victim to reset the password with their assistance.
Social Proof The use of endorsement; for example, stating that, due to recent suspicious activities, new security requirements are issued and must be complied by all account holders.
Consistency Leveraging the desire of individuals to be consistent with their words, belief, and actions; for example, reminding users that they have to comply with a password reset policy as they have previously done.
Authority Responding to others with more experience, knowledge, or power; for example, an email signed by a Senior Vice President of a bank requesting customers to reset their account passwords.
Scarcity Something being valuable when it is perceived to be rare or available for a limited time; for example, giving a user 24-h notice before they deactivate the user's account.
Unity Shared identity between the influencer and the influenced

The principle of Unity was introduced in Cialdini (2016) but has not been studied in social engineering research; it is presented here for the purpose of completeness.

Intuitively, the mathematical function f in Equation (1) should accommodate or reflect Cialdini's Principles of Persuasion. Although the state-of-the-art does not allow us to draw insights into how these Principles would quantitatively affect the form of f, we can still draw some insights from existing studies, as discussed below.

van der Heijden and Allodi (2019) study the relation between phishing attack success and Cialdini Principles of Persuasion using enterprise emails from a financial institution. They find that phishing emails that received the most responses (“clicks”) are those who use consistency and scarcity principles mentioned above. They also find that emails with more cognitive elements (e.g., proper grammar, personalization, and persuasion elements) receive most responses. In a related study, Lin et al. (2019) find that younger individuals are more susceptible to phishing messages that use the scarcity and authority principles, while older individuals are more susceptible to phishing emails that use the reciprocity and liking principles. Rajivan and Gonzalez (2018) find that the most successful phishing message strategies are notifications messages, authoritative messages, friend request messages, shared interest messages, and assistance with a failure. These strategies map to Cialdini's Principles of liking, authority, and unity. Lawson et al. (2018) find that socially engineered messages that use authority and scarcity principles are considered more suspicious than those that use the liking principle.

There have been proposals to augment Cialdini's Principles to better represent the psychological vulnerabilities that have been exploited by social engineering cyberattacks. Ferreira and colleagues (Ferreira and Lenzini, 2015; Ferreira et al., 2015) present five Principles of Persuasion by combining (i) Cialdini Principles of Persuasion; (ii) Stajano's study on scams and how distraction, social compliance, herd, dishonesty, kindness, need and greed, and time affect the persuasive power of scam messages (Stajano and Wilson, 2009); and (iii) Gragg's psychological triggers on how strong affect or emotion, overloading, reciprocation, deceptive relationships, diffusion of responsibility and moral duty, authority, and integrity and consistency can influence an individual's response to social engineered messages (Gragg, 2003). Table 2 presents a summary of these newly proposed five principles.

Table 2.

Summary of the five newly proposed principles of persuasion that would better represent the psychological vulnerabilities exploited by social engineering cyberattacks (Ferreira et al., 2015).

Principle Description
Authority Obeying pretense of authority or performing a favor for an authority.
Social Proof Mimicking behavior of the majority of people.
Liking, Similarity, and Deception (LSD) Obeying someone a victim knows/likes, or someone similar to the victim, or someone a victim finds attractive.
Commitment, Reciprocity, and Consistency (CRC) Making a victim act as committed, assuring consistency between saying and doing, or returning a favor.
Distraction Focusing on what a victim can gain, need, or lose/miss out.

Guided by the newly proposed principles, Ferreira and Lenzini (2015) conduct experiments, using phishing emails, to show that distraction (e.g., fear of missing out, scarcity, strong affection, overloading, and time) is the most prevalent phishing tactic, followed by authority and LSD.

Summarizing the preceding discussion, we draw:

Insight 13. The representation of mathematical function f in Equation (1) should adequately reflect the Principles of Persuasion.

Since quantitatively describing the mathematical function f, as demanded in Insight 13, is beyond the scope of the state-of-the-art, in what follows we explore some qualitative properties of the these mathematical functions, showing how an increase (decrease) in a model parameter would affect the outcome behavior (more precisely, persuasion) of a victim. These qualitative observations also need to be quantitatively verified by future experimental studies.

3.3.3. Impact of Attacker Effort on Victim Behavior

As shown in Equation (1), the attacker can affect victim behavior through the attack effort variable, which can be reflected by attacker's message quality and message appeal with respect to the victim in question. In terms of the impact of message quality, Downs et al. (2006) find that most individuals rely on superficial elements when determining if a message is legitimate, without knowing that most of those elements can be spoofed. Jansen and Leukfeldt (2016) find that in online banking frauds involving phishing or malware, most victims report that fraudulent stories in phishing emails or phone calls appear to be trustworthy. Message quality appears increase social engineering cyberattack success. Wang et al. (2012) find that visceral triggers and deception indicators affect phishing responses. Visceral triggers increases phishing responses, whereas deception indicators have the opposite effect, reducing phishing response. Similarly, Vishwanath et al. (2011) find that individuals use superficial message cues to determine their response to phishing messages. The study reports that urgency cues make it less likely for an individual to detect deception cues. One common method of trustworthiness is through the use of visual deception, which is effective because most individuals associate professional appearance and logos with a website or message been legitimate. Visual deception involves the use of high-quality superficial attributes (e.g., legitimate logos, professional design, and name spoofing). Hirsh et al. (2012) find that phishing messages that use visual deception have a higher victim response rate. Jakobsson (2007) observes that for most individuals, the decision to trust a website or not is based on-site content and not on the status of a site security indicators (e.g., the use security certificates, HTTPS). He also notes that most users could not detect subtle changes in URLs (e.g., a malicious www.IUCU.com vs. a benign www.IUCU.org). Dhamija (Dhamija et al., 2006) conducts a study on malicious website identification. Using malicious and legitimate websites with professional appearance, with the difference that malicious sites display security indicators (e.g., missing padlock icon on the browser, warning on site's digital certificate), they find that 23% of their participants fail to identify fraudulent sites for 40% of the time. This group of participants are asked to assess a website's legitimacy based on its appearance (e.g., website design and logos), and 90.9% of participants fail to identify a high-quality malicious website that uses visual deception (i.e., URL spoofing replacing letters in a legitimate URL, for example, “W” for “vv”).

On the other hand, message appeal is associated with the benefit an individual derives from complying with a request. Halevi et al. (2013, 2015) find that many individuals that fall to social engineering cyberattacks ignore the risk of their actions because they focus on the potential benefit that the phishing email offers. Message appeal has the most weight on social engineering susceptibility. An example of this is the Nigerian scam, also known as the “419” scam. Herley (2012) notes that although the scam is well-known and information on the scam is readily available online, individuals still fall victim to it because the message is designed to appeal the most gullible. Two techniques that are commonly used to increase message appeal are contextualization, also known as pretexting, and personalization.

  • Contextualization is a variation of message framing where the sender provides details or discusses topics relevant to the group to vouch for the victim's membership in the group. Luo et al. (2013) conducts a study on phishing victimization with contextualization in the message. In the experiment, they use work benefits and compensation as a pretext in an email to University staff. They find that individuals interpret high-quality content and argument messages (e.g., well written, persuasive messages) as originating from a credible sender. Basing the message argument on a topic that is common within a community (i.e., contextualization) gives the message the appearance of originating from a known person within a group. Using this technique, they are able to achieve 15.24% victimization in 22 min by combining pretexting and message quality. Similarly, Goel et al. (2017) examine the effect of messages contextualization (i.e., pretexting) on phishing message opening and compliance rates. They find that highly contextualized messages that target issues relevant to the victim are more successful. They also find that messages with higher perceived loss have higher success rates than those with high perceived gains. Rajivan and Gonzalez (2018) also find that phishing messages on work-related or social communication topics (e.g., friend requests) have a higher success rate than messages requesting a password change or offering a deal.

  • Personalization is another framing technique in which a message is tailored to the preference of the victim (Hirsh et al., 2012), such as friendship appeals, expressing similar interests. In a phishing experiment (Jagatic et al., 2007), find that adding personal data found in social networks to phishing emails increased the response rate from 16 to 72%. Rocha Flores et al. (2014) find that targeted, personalized phishing messages are more effective than generic messages. They find that phishing emails that receive the most responses are those perceived to come from a known source. Bullee et al. (2017) also find that emails using personalized greeting line were responded 1.7 times more likely when compared with emails with generic greeting lines.

Summarizing the preceding discuss, we draw:

Insight 14. Message quality and message appeal, which reflect attacker effort (e.g., using contextualization and personalization), have a significant impact on the attacker's success.

3.3.4. Countermeasures Against Social Engineering Cyberattacks

There have been some studies on defending against social engineering cyberattacks. First, it is intuitive that effective training should heighten a victim's sense of threat because individuals are more cautious and sensitive to detecting elements that might indicate deception. Along this line, Wright and Marett (2010) find that suspicion of humanity was a dispositional factor that increases the detection of deception in phishing messages, more so than risk beliefs and trust. Pattinson et al. (2012) also find that when individuals participating in experiments are aware that a phishing attack is involved, they perform better on detecting phishing emails. Second, Tembe et al. (2014) find that individuals from the U.S. having higher suspicion and caution attitudes on online communications, when compared to individuals from China and India. Third, Vishwanath et al. (2018) find that habitual patterns of email habits and deficient self-regulation reduce viewers' suspicion. Moreover, detecting deception cues decrease social engineering susceptibility. Kirmani and Zhu (2007) find that detecting persuasion cues in a message activates suspicion and generates a negative response to the message. Fourth, Canfield et al. (2016) find that individuals' inability to detect phishing messages increases their susceptibility, regardless of their cautionary behavior. For suspicions to be effective in reducing social engineering susceptibility, the risk must out weight the benefit of complying with the message (e.g., message appeal) without affecting decision performance (i.e., accuracy, precision and negative prediction) (Cho et al., 2016). Goel et al. (2017) find that suspicion alone does not prevent phishing victimization because they report that individuals that are suspicious about email messages can still fall victim to social engineering cyberattacks. Summarizing the preceding discussion, we draw the following:

Insight 15. An individual's capabilities in detection of social engineering cyberattacks are affected by their awareness of the threats, their cultural backgrounds, and their individual differences in trust/suspicion.

Insight 15 highlights that there is no silver-bullet solution to countering social engineering cyberattacks. On the contrary, effective defense must take into consideration the differences between individuals because they are susceptible to social engineering cyberattacks at different degrees.

For more effective defenses against social engineering cyberattacks, the following aspects need to be systematically investigated in the future.

  • Achieving effective human-machine interactions in defending against social engineering cyberattacks. Effective defense would require to (i) detecting message elements that attempt to increase recipients' trust and (ii) increasing recipients' suspicion on messages. In either approach, we would need human-machine teaming in detecting social-engineering cyberattacks, highlighting the importance of effective defenses against social engineering cyberattacks.

  • Improving users' immunity to social engineering cyberattacks. To improve users' immunity to social engineering cyberattacks, we first need to investigate how to quantify their immunity. In order to improve users' immunity to social engineering cyberattacks, we need to enhance users' protection motivation and capabilities in detecting deceptive cues. One approach is to enhance the user interface design to highlight the security alerts/indicators in email systems and web browsers because their current designs are not effective (Downs et al., 2006; Kumaraguru et al., 2006; Schechter et al., 2007; Abbasi et al., 2016). Along this direction, the user interface must highlight security alerts/indicators with specific and quantified severity of threats to the user. The current user-interface design appears to mainly focus on usability and user experience while assuming the presence of (social engineering) cyberattacks as a default. This design premise needs to be changed to treating the presence of (social engineering) cyberattacks as the default. In order to enhance users' capabilities in detecting deceptive cues, we need to design new techniques to enhance our capabilities in automatically detecting or assisting users to detect social engineering cyberattacks. Automatic detection has been pursued by previous studies, which however only focus on examining certain message elements that are known to be associated with previous social engineering cyberattacks (i.e., signature-based detectors); these signatures can be easily avoided by attackers. In order to possibly detect new social engineering cyberattacks, future detectors should incorporate cognitive psychology elements to detect social engineering cyberattacks, such as the quantification of messages' persuasiveness and deceptiveness. In order to design automated techniques to assist users in detecting social engineering cyberattacks, human-machine interaction is an important issue.

  • Achieving human-centric systems design with quantifiable cybersecurity gain. Modern systems design, including security systems design, often focuses on optimizing performance without considering how humans would introduce vulnerabilities while interacting with the system (i.e., assuming away that humans are often the weakest link). One approach to addressing this problem is to change the designers' mindset to treat users of these systems as the weakest link. This can be achieved by, for example, using security designs that are simpler and less error-prone, while considering the worst-case scenario that the users may have a high cognitive load when using these systems. How to quantify users' vulnerability to social engineering attacks is an outstanding problem because it paves the way to quantify the cybersecurity gain of a better design when compared with a worse design.

  • Designing effective training to enhance users' self-efficacy. Training is an important mechanism for defending against social engineering cyberattacks. However, it is a challenging task to design effective training. One approach to addressing the problem is to routinely expose users to specific socially engineering cyberattacks (e.g., messages that have been used by attackers). Another approach is to insert sanitized social engineering attacks (e.g., phishing emails without malicious payload) into users' routine activities to trigger users' response to social engineering cyberattacks (e.g., the feedback will point to the user in question whether the user correctly processed the message). This also effectively increases users' perception of social engineering cyberattacks, making them appear more frequent than it is.

  • Understanding and quantifying the impact of short-term factors in social engineering cyberattacks. We observe that as discussed above, few studies have examined the effect of short-term factors in social engineering cyberattacks. Short-term factors are known to affect cognition and behavior in other contexts profoundly. As highlighted in Equation (1) and discussed above, we stress the importance of defining and quantifying social engineering cyberattack metrics, which are largely missing and will become an indispensable component of the broader family of cybersecurity metrics as discussed in Pendleton et al. (2016), Cho et al. (2019), and Xu (2019).

Table 3 highlights the five future research directions and their relationships to the insights.

Table 3.

Relationships between future research directions and insights.

Future research direction Insights
Human-Machine interactions Reduce trust (Insights 1, 2, 10, 14); Increase suspicion (Insights 3, 5, 7, 9)
Immunity to social engineering cyber attacks Identify and quantify underlying causes of immunity (Insights 5, 7, 9); Security and UI design (Insight 1, 2, 3); Improve message detection (Insight 10, 13, 14)
Human-Centric System Design Incorporate psychological state of computer user during system design (Insight 1, 2, 3, 11)
Designing effective training Training based on susceptibility elements (Insight 4, 6, 8, 10, 11, 12, 14, 15)
Understanding and quantifying the impact of short-term factors Increase research focus on short term factors impact in security (Insight 1, 2, 3)

3.4. Further Discussion

First, we observe that the 15 insights mentioned above are all qualitative rather than quantitative. Moreover, the factors are typically investigated standalone. Furthermore, even the qualitative effects are discussed in specific scenarios, meaning that they may not be universally true. Summarizing most of the insights mentioned above, the state-of-the understanding is the following: (i) cognitive workload, stress, and attack effort increase one's vulnerability to social engineering cyberattacks; (ii) the effect of vigilance, personality, awareness, and culture remains to be investigated to be conclusive; (iii) domain knowledge, (certain kinds of) experience, and age (together with certain other factors) reduce one's vulnerability to social engineering cyberattacks; and (iv) gender may not have a significant effect on one's vulnerability to social engineering cyberattacks.

Figure 3 depicts our speculation of the impact on one's susceptibility to social engineering cyberattacks. Specifically, we suspect that expertise can decrease one's susceptibility to the largest extent among the factors because expertise equips one the capability to detect the deceptive cues that are used by social engineering cyberattacks. We suspect that vigilance and domain knowledge have a significant, but smaller, impact on reducing one's susceptibility because it is perhaps harder for an expert to fall into victim of social engineering cyberattacks. Since there is no evidence to show which one of these two factors would have a bigger impact than the other, we subjectively treat them as if they have the same impact. We suspect awareness would have a significant impact on reducing one's susceptibility, despite that the literature does not provide any evidence. According to Insight 8 (which is drawn from a body of literature reviewed above), gender has little or no impact. We suspect that stress would decrease one's capability in detecting deception cues, but attack effort would have an even more significant impact on increasing one's susceptibility. We suspect that workload may have the biggest impact on one's susceptibility because it would substantially reduce one's ability in detecting deception cues. For other factors like age and culture, we suspect that their impacts might have to be considered together with other factors, explaining why we do not include them in Figure 3. In summary, our understanding of the factors that have impacts on human's susceptibility to social engineering cyberattackers is superficial. This was indeed one of our motivations for proposing the mathematical framework outlined in Equation (1).

Figure 3.

Figure 3

Our speculation of the impact on one's susceptibility to social engineering cyberattacks: a factor on the left-hand side means that substantially increasing the factor (e.g., expertise) will decrease one's susceptibility, with further left indicating a bigger degree in decreasing susceptibility; a factor on the right-hand side means that substantially increasing the factor (e.g., workload) will increase one's susceptibility, with further right indicating a bigger degree in increasing susceptibility; gender has little or no effect on one's susceptibility.

Second, it is a fascinating research problem to fulfill the quantitative framework envisioned in the paper because its fulfillment will permit us to identify cost-effective, if not optimal, defense strategies against social engineering cyberattacks. This will also help identify the most important factors. However, we suspect that the optimal defense strategies will vary with, for example, different combinations of short-term factors and long-term factors. For example, we suspect that the importance of factors may be specific to attack scenarios. This is supported by two very recent studies: van der Heijden and Allodi (2019) observed that certain short-term and long-term factors (e.g., workload) may be exploited to wage phishing attacks because malicious emails can coincide with high email volume; and Jalali et al. (2020) showed that certain short-term and long-term factors (e.g., high workload and lack of expertise) are two important factors against medical workers. For example, Insight 6 says that awareness and general technical knowledge do not necessarily reduce one's susceptibility to social engineering cyberattacks; however, this may not hold when taking awareness and human cognition functions into consideration. In other words, we can speculate that the effect of considering one factor alone and the effect of considering multiple interacting factors together may be different. This phenomenon is also manifested by Insight 7, showing that self-efficacy, knowledge, and previous encounter of social engineering cyberattacks collectively reduce one's susceptibility to social engineering cyberattacks. In particular, costly phishing experiences would greatly reduce one's susceptibility to social engineering cyberattacks, while non-costly experiences do not. Putting another way, a certain previous encounter may or may not have a big effect when considered together with other factors.

Third, Insight 12 says that effective training should not ask people to consciously think about social engineering cyberattacks, but making people to formulate an unconscious habit in coping with these attacks. This points out an important research direction on designing future training systems.

Fourth, studies in the context of social engineering cyberattacks inevitably involve human subject, meaning that ethical aspects of these studies must be taken into adequate consideration when designing such experiments and an IRB approval must be sought before conducting any such experiment. For ethical considerations in phishing experiments, we refer to (Finn and Jakobsson, 2007) for a thorough treatment.

4. Conclusion

We have presented a framework for systematizing human cognition through the lens of social engineering cyberattacks, which exploit weaknesses in human's cognition functions. The framework is extended from the standard cognitive psychology to accommodate components that emerge from the cybersecurity context. In particular, the framework leads to a representation of a victim's behavior, or more precisely, the degree a victim is persuaded by an attacker to act as the attacker intended, as some mathematical function(s) of many aspects, including victim's cognition functions and attacker's effort. We articulate a number of research directions for future research. We hope that this mathematical representation will guide future research endeavors toward a systematic and quantitative theory of Cybersecurity Cognitive Psychology.

Author Contributions

All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We thank the reviewers for their constructive comments, which have guided us in improving the paper. In particular, the entire section 3.4 is inspired by the reviewers' comments.

Footnotes

Funding. This paper was supported in part by the Mitre Basic Education Assistance Program (BEAP) and ARO Grant #W911NF-17-1-0566 and NSF Grants #1814825 and #1736209. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. The opinions expressed in the paper are those of the authors and do not reflect the funding agencies' policies in any sense.

References

  1. Abass I. A. M. (2018). Social engineering threat and defense: a literature survey. J. Inform. Secur. 9:257 10.4236/jis.2018.94018 [DOI] [Google Scholar]
  2. Abbasi A., Zahedi F. M., Chen Y. (2016). Phishing susceptibility: the good, the bad, and the ugly, in 2016 IEEE Conference on Intelligence and Security Informatics (ISI) (Tucson: IEEE; ), 169–174. 10.1109/ISI.2016.7745462 [DOI] [Google Scholar]
  3. Acquisti A., Grossklags J. (2005). Privacy and rationality in individual decision making. IEEE Secur. Privacy 3, 26–33. 10.1109/MSP.2005.22 [DOI] [Google Scholar]
  4. Al'Absi M., Hugdahl K., Lovallo W. R. (2002). Adrenocortical stress responses and altered working memory performance. Psychophysiology 39, 95–99. 10.1111/1469-8986.3910095 [DOI] [PubMed] [Google Scholar]
  5. Al-Hamar M., Dawson R., Guan L. (2010). A culture of trust threatens security and privacy in Qatar, in 2010 10th IEEE International Conference on Computer and Information Technology (Bradford: IEEE; ), 991–995. 10.1109/CIT.2010.182 [DOI] [Google Scholar]
  6. Anderson R. J. (2008). Security Engineering: A Guide to Building Dependable Distributed Systems, 2nd Edn. New York, NY: Wiley Publishing. [Google Scholar]
  7. Arachchilage N. A. G., Love S. (2014). Security awareness of computer users: a phishing threat avoidance perspective. Comput. Hum. Behav. 38, 304–312. 10.1016/j.chb.2014.05.046 [DOI] [Google Scholar]
  8. Arnett J. J. (2008). The neglected 95%: why American psychology needs to become less American. Am. Psychol. 63:602. 10.1037/0003-066X.63.7.602 [DOI] [PubMed] [Google Scholar]
  9. Arnsten A. F. (2009). Stress signalling pathways that impair prefrontal cortex structure and function. Nat. Rev. Neurosci. 10:410. 10.1038/nrn2648 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Baars B. J. (1997). In the theatre of consciousness. global workspace theory, a rigorous scientific theory of consciousness. J. Conscious. Stud. 4, 292–309. 10.1093/acprof:oso/9780195102659.001.1 [DOI] [Google Scholar]
  11. Bohm M. (2011). Why Russians Don't Smile. Available online at: https://themoscowtimes.com/articles/why-russians-dont-smile-6672
  12. Bullee J.-W., Montoya L., Junger M., Hartel P. (2017). Spear phishing in organisations explained. Inform. Comput. Secur. 25, 593–613. 10.1108/ICS-03-2017-0009 [DOI] [Google Scholar]
  13. Byrne Z. S., Dvorak K. J., Peters J. M., Ray I., Howe A., Sanchez D. (2016). From the user's perspective: perceptions of risk relative to benefit associated with using the internet. Comput. Hum. Behav. 59, 456–468. 10.1016/j.chb.2016.02.024 [DOI] [Google Scholar]
  14. Cahill L. (2006). Why sex matters for neuroscience. Nat. Rev. Neurosci. 7:477. 10.1038/nrn1909 [DOI] [PubMed] [Google Scholar]
  15. Cain A. A., Edwards M. E., Still J. D. (2018). An exploratory study of cyber hygiene behaviors and knowledge. J. Inform. Secur. Appl. 42, 36–45. 10.1016/j.jisa.2018.08.002 [DOI] [Google Scholar]
  16. Canfield C. I., Fischhoff B., Davis A. (2016). Quantifying phishing susceptibility for detection and behavior decisions. Hum. Factors 58, 1158–1172. 10.1177/0018720816665025 [DOI] [PubMed] [Google Scholar]
  17. Cho J.-H., Cam H., Oltramari A. (2016). Effect of personality traits on trust and risk to phishing vulnerability: modeling and analysis, in 2016 IEEE International Multi-Disciplinary Conference on Cognitive Methods in Situation Awareness and Decision Support (CogSIMA) (San Diego, CA: IEEE; ), 7–13. [Google Scholar]
  18. Cho J.-H., Xu S., Hurley P. M., Mackay M., Benjamin T., Beaumont M. (2019). STRAM: Measuring the trustworthiness of computer-based systems. ACM Comput. Surv (New York, NY: ), 51 10.1145/3277666 [DOI] [Google Scholar]
  19. Chu P. C., Spires E. E., Sueyoshi T. (1999). Cross-cultural differences in choice behavior and use of decision aids: a comparison of Japan and the United States. Organ. Behav. Hum. Decis. Process. 77, 147–170. 10.1006/obhd.1998.2817 [DOI] [PubMed] [Google Scholar]
  20. Cialdini R. (2016). Pre-suasion: A Revolutionary Way to Influence and Persuade. New York, NY: Simon and Schuster. [Google Scholar]
  21. da Veiga A., Martins N. (2017). Defining and identifying dominant information security cultures and subcultures. Comput. Secur. 70, 72–94. 10.1016/j.cose.2017.05.002 [DOI] [Google Scholar]
  22. Damon W., Lerner R. M., Kuhn D., Siegler R. S. (2006). Handbook of Child Psychology, Cognition, Perception, and Language, Vol. 2 Hoboken, NJ: John Wiley & Sons. [Google Scholar]
  23. Darwish A., El Zarka A., Aloul F. (2012). Towards understanding phishing victims' profile, in 2012 International Conference on Computer Systems and Industrial Informatics (Sharjah: IEEE; ), 1–5. 10.1109/ICCSII.2012.6454454 [DOI] [Google Scholar]
  24. DeValois R. L., DeValois K. K. (1990). Spatial Vision, Vol. 14 Oxford University Press; 10.1093/acprof:oso/9780195066579.001.0001 [DOI] [Google Scholar]
  25. Dhamija R., Tygar J. D., Hearst M. (2006). Why phishing works, in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (New York, NY: ACM; ), 581–590. 10.1145/1124772.1124861 [DOI] [Google Scholar]
  26. Digman J. M. (1997). Higher-order factors of the big five. J. Pers. Soc. Psychol. 73:1246. 10.1037/0022-3514.73.6.1246 [DOI] [PubMed] [Google Scholar]
  27. Donnellan M. B., Robins R. W. (2009). The development of personality across the lifespan, in The Cambridge Handbook of Personality Psychology, eds Corr P. J., Matthews G. (Cambridge: Cambridge University Press; ), 191. 10.1017/CBO9780511596544.015 [DOI] [Google Scholar]
  28. Downs J. S., Holbrook M. B., Cranor L. F. (2006). Decision strategies and susceptibility to phishing, in Proceedings of the Second Symposium on Usable Privacy and Security (Pittsburgh, PA: ACM; ), 79–90. 10.1145/1143120.1143131 [DOI] [Google Scholar]
  29. Elzinga B. M., Roelofs K. (2005). Cortisol-induced impairments of working memory require acute sympathetic activation. Behav. Neurosci. 119:98. 10.1037/0735-7044.119.1.98 [DOI] [PubMed] [Google Scholar]
  30. Evans J. S. B. (2008). Dual-processing accounts of reasoning, judgment, and social cognition. Annu. Rev. Psychol. 59, 255–278. 10.1146/annurev.psych.59.103006.093629 [DOI] [PubMed] [Google Scholar]
  31. Farhat N. F. N. (2017). Scam Alert - Blackmail Email. Available online at: https://www.linkedin.com/pulse/scam-alert-blackmail-email-ned-farhat
  32. Ferreira A., Coventry L., Lenzini G. (2015). Principles of persuasion in social engineering and their use in phishing, in International Conference on Human Aspects of Information Security, Privacy, and Trust (Los Angeles, CA: Springer; ), 36–47. 10.1007/978-3-319-20376-8_4 [DOI] [Google Scholar]
  33. Ferreira A., Lenzini G. (2015). An analysis of social engineering principles in effective phishing, in 2015 Workshop on Socio-Technical Aspects in Security and Trust (Verona: IEEE; ), 9–16. 10.1109/STAST.2015.10 [DOI] [Google Scholar]
  34. Finn P., Jakobsson M. (2007). Designing ethical phishing experiments. IEEE Technol. Soc. Mag. 26, 46–58. 10.1109/MTAS.2007.335565 [DOI] [Google Scholar]
  35. Franklin D. W., Wolpert D. M. (2011). Computational mechanisms of sensorimotor control. Neuron 72, 425–442. 10.1016/j.neuron.2011.10.006 [DOI] [PubMed] [Google Scholar]
  36. Gavett B. E., Zhao R., John S. E., Bussell C. A., Roberts J. R., Yue C. (2017). Phishing suspiciousness in older and younger adults: the role of executive functioning. PLoS ONE 12:e0171620. 10.1371/journal.pone.0171620 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Gigerenzer G. (2008). Why heuristics work. Perspect. Psychol. Sci. 3, 20–29. 10.1111/j.1745-6916.2008.00058.x [DOI] [PubMed] [Google Scholar]
  38. Goel S., Williams K., Dincelli E. (2017). Got phished? Internet security and human vulnerability. J. Assoc. Inform. Syst. 18:2 10.17705/1jais.00447 [DOI] [Google Scholar]
  39. Gragg D. (2003). A multi-level defense against social engineering. SANS Reading Room 13, 1–21. 10.1093/acprof:oso/9780199253890.003.0002 [DOI] [Google Scholar]
  40. Grandstrand O. (2013). Cultural differences and their mechanisms, in The Oxford Handbook of Cognitive Psychology, ed Reisberg D. (Oxford: Oxford University Press; ), 970–985. [Google Scholar]
  41. Gupta S., Singhal A., Kapoor A. (2016). A literature survey on social engineering attacks: phishing attack, in 2016 International Conference on Computing, Communication and Automation (ICCCA) (Greater Noida: IEEE; ), 537–540. 10.1109/CCAA.2016.7813778 [DOI] [Google Scholar]
  42. Halevi T., Lewis J., Memon N. (2013). A pilot study of cyber security and privacy related behavior and personality traits, in Proceedings of the 22nd International Conference on World Wide Web (Singapore: ACM; ), 737–744. 10.1145/2487788.2488034 [DOI] [Google Scholar]
  43. Halevi T., Memon N., Lewis J., Kumaraguru P., Arora S., Dagar N., et al. (2016). Cultural and psychological factors in cyber-security, in Proceedings of the 18th International Conference on Information Integration and Web-based Applications and Services, iiWAS '16 (New York, NY: ACM; ), 318–324. 10.1145/3011141.3011165 [DOI] [Google Scholar]
  44. Halevi T., Memon N., Nov O. (2015). Spear-phishing in the wild: a real-world study of personality, phishing self-efficacy and vulnerability to spear-phishing attacks. SSRN Electron. J. 10.2139/ssrn.2544742 [DOI] [Google Scholar]
  45. Harrison B., Svetieva E., Vishwanath A. (2016). Individual processing of phishing emails: how attention and elaboration protect against phishing. Online Inform. Rev. 40, 265–281. 10.1108/OIR-04-2015-0106 [DOI] [Google Scholar]
  46. Herley C. (2012). “Why do Nigerian scammers say they are from Nigeria?” in WEIS (Berlin: ). [Google Scholar]
  47. Hirsh J., Kang S., Bodenhausen G. (2012). Personalized persuasion: tailoring persuasive appeals to recipients' personality traits. Psychol. Sci. 23, 578–581. 10.1177/0956797611436349 [DOI] [PubMed] [Google Scholar]
  48. Hof P. R., Mobbs C. V. (2001). Functional Neurobiology of Aging. Amsterdam: Elsevier. [Google Scholar]
  49. Hofstede G. H., Hofstede G. J., Minkov M. (2010). Cultures and Organizations: Software of the Mind, 3rd Edn McGraw-Hill. [Google Scholar]
  50. Hong K. W., Kelley C. M., Tembe R., Murphy-Hill E., Mayhorn C. B. (2013). Keeping up with the joneses: assessing phishing susceptibility in an email task, in Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Los Angeles, CA: SAGE Publications; ), 1012–1016. 10.1177/1541931213571226 [DOI] [Google Scholar]
  51. Horn J. L., Cattell R. B. (1967). Age differences in fluid and crystallized intelligence. Acta Psychol. 26, 107–129. 10.1016/0001-6918(67)90011-X [DOI] [PubMed] [Google Scholar]
  52. Howe A. E., Ray I., Roberts M., Urbanska M., Byrne Z. (2012). The psychology of security for the home computer user, in Proceedings of the 2012 IEEE Symposium on Security and Privacy, SP '12 (Washington, DC: IEEE Computer Society; ), 209–223. 10.1109/SP.2012.23 [DOI] [Google Scholar]
  53. Hutchins E. M., Cloppert M. J., Amin R. M. (2011). Intelligence-driven computer network defense informed by analysis of adversary campaigns and intrusion kill chains. Lead. Issues Inform. Warfare Secur. Res. 1:80. [Google Scholar]
  54. Indrajit R. E. (2017). Social engineering framework: Understanding the deception approach to human element of security. Int. J. Comput. Sci. Issues 14, 8–16. 10.20943/01201702.816 [DOI] [Google Scholar]
  55. Jagatic T. N., Johnson N. A., Jakobsson M., Menczer F. (2007). Social phishing. Commun. ACM 50, 94–100. 10.1145/1290958.1290968 [DOI] [Google Scholar]
  56. Jakobsson M. (2007). The human factor in phishing. Privacy Secur. Cons. Inform. 7, 1–19. [Google Scholar]
  57. Jalali M. S., Bruckes M., Westmattelmann D., Schewe G. (2020). Why employees (still) click on phishing links: investigation in hospitals. J. Med. Internet Res. 22:e16775. 10.2196/16775 [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Jansen J., Leukfeldt R. (2016). Phishing and malware attacks on online banking customers in the Netherlands: a qualitative analysis of factors leading to victimization. Int. J. Cyber Criminol. 10:79 10.5281/zenodo.58523 [DOI] [Google Scholar]
  59. Junger M., Montoya L., Overink F.-J. (2017). Priming and warnings are not effective to prevent social engineering attacks. Comput. Hum. Behav. 66(Suppl. C), 75–87. 10.1016/j.chb.2016.09.012 [DOI] [Google Scholar]
  60. Kahneman D. (2011). Thinking, Fast and Slow. New York, NY: Farrar, Straus and Giroux. [Google Scholar]
  61. Kaivanto K. (2014). The effect of decentralized behavioral decision making on system-level risk. Risk Anal. 34, 2121–2142. 10.1111/risa.12219 [DOI] [PubMed] [Google Scholar]
  62. Kandel E. R., Schwartz J. H., Jessell T. M. (2000). Principles of Neural Science, Vol. 4 New York, NY: McGraw-Hill. [Google Scholar]
  63. Keil F. C. (2003). Folkscience: Coarse interpretations of a complex reality. Trends Cogn. Sci. 7, 368–373. 10.1016/S1364-6613(03)00158-X [DOI] [PubMed] [Google Scholar]
  64. Kenrick D. T., Funder D. C. (1988). Profiting from controversy: lessons from the person-situation debate. Am. Psychol. 43:23. 10.1037/0003-066X.43.1.23 [DOI] [PubMed] [Google Scholar]
  65. Kimball D. R., Holyoak K. J. (2000). Transfer and expertise, in The Oxford Handbook of Memory, eds Tulving E., Craik F. I. M. (New York, NY: Oxford University Press; ), 109–122. [Google Scholar]
  66. Kirmani A., Zhu R. (2007). Vigilant against manipulation: the effect of regulatory focus on the use of persuasion knowledge. J. Market. Res. 44, 688–701. 10.1509/jmkr.44.4.688 [DOI] [Google Scholar]
  67. Klein G. A., Calderwood R. (1991). Decision models: some lessons from the field. IEEE Trans. Syst. Man Cybernet. 21, 1018–1026. 10.1109/21.120054 [DOI] [Google Scholar]
  68. Kruger J., Dunning D. (1999). Unskilled and unaware of it: how difficulties in recognizing one's own incompetence lead to inflated self-assessments. J. Pers. Soc. Psychol. 77:1121. 10.1037/0022-3514.77.6.1121 [DOI] [PubMed] [Google Scholar]
  69. Kumaraguru P., Acquisti A., Cranor L. F. (2006). Trust modelling for online transactions: a phishing scenario, in Proceedings of the 2006 International Conference on Privacy, Security and Trust: Bridge the Gap Between PST Technologies and Business Services (Markham, ON: ACM; ), 11 10.1145/1501434.1501448 [DOI] [Google Scholar]
  70. Lawson P. A., Crowson A. D., Mayhorn C. B. (2018). Baiting the hook: exploring the interaction of personality and persuasion tactics in email phishing attacks, in Congress of the International Ergonomics Association (Florence: Springer; ), 401–406. 10.1007/978-3-319-96077-7_42 [DOI] [Google Scholar]
  71. Lin T., Capecci D. E., Ellis D. M., Rocha H. A., Dommaraju S., Oliveira D. S., et al. (2019). Susceptibility to spear-phishing emails: Effects of internet user demographics and email content. ACM Trans. Comput. Hum. Interact. 26:32. 10.1145/3336141 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Linvill D. L., Boatwright B. C., Grant W. J., Warren P. L. (2019). “The Russians are hacking my brain!” investigating Russia's internet research agency twitter tactics during the 2016 United States presidential campaign. Comput. Hum. Behav. 99, 292–300. 10.1016/j.chb.2019.05.027 [DOI] [Google Scholar]
  73. Luo X. R., Zhang W., Burd S., Seazzu A. (2013). Investigating phishing victimization with the heuristic-systematic model: a theoretical framework and an exploration. Comput. Secur. 38, 28–38. 10.1016/j.cose.2012.12.003 [DOI] [Google Scholar]
  74. Lupien S. J., McEwen B. S., Gunnar M. R., Heim C. (2009). Effects of stress throughout the lifespan on the brain, behaviour and cognition. Nat. Rev. Neurosci. 10:434. 10.1038/nrn2639 [DOI] [PubMed] [Google Scholar]
  75. Mackworth N. H. (1948). The breakdown of vigilance during prolonged visual search. Q. J. Exp. Psychol. 1, 6–21. 10.1080/17470214808416738 [DOI] [Google Scholar]
  76. Mather M., Sutherland M. R. (2011). Arousal-biased competition in perception and memory. Perspect. Psychol. Sci. 6, 114–133. 10.1177/1745691611400234 [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. McBride M., Carter L., Warkentin M. (2012). Exploring the role of individual employee characteristics and personality on employee compliance with cybersecurity policies. RTI Int. Instit. Homeland Secur. Solut. 5:1. [Google Scholar]
  78. Mesulam M.-M. (1998). From sensation to cognition. Brain 121, 1013–1052. 10.1093/brain/121.6.1013 [DOI] [PubMed] [Google Scholar]
  79. Mitnick K., Simon W. L. (2003). The Art of Deception: Controlling the Human Element of Security. Indianapolis, IN: Wiley Publishing. [Google Scholar]
  80. Miyake A., Shah P. (1999). Models of Working Memory: Mechanisms of Active Maintenance and Executive Control. Cambridge, UK: Cambridge University Press; 10.1017/CBO9781139174909 [DOI] [Google Scholar]
  81. Navon D., Gopher D. (1979). On the economy of the human-processing system. Psychol. Rev. 86:214 10.1037/0033-295X.86.3.214 [DOI] [Google Scholar]
  82. Nisbett R. E., Wilson T. D. (1977). Telling more than we can know: verbal reports on mental processes. Psychol. Rev. 84:231. 10.1037/0033-295X.84.3.23117049881 [DOI] [Google Scholar]
  83. Nosek B. A., Hawkins C. B., Frazier R. S. (2011). Implicit social cognition: from measures to mechanisms. Trends Cogn. Sci. 15, 152–159. 10.1016/j.tics.2011.01.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Ovelgönne M., Dumitras T., Prakash B. A., Subrahmanian V. S., Wang B. (2017). Understanding the relationship between human behavior and susceptibility to cyber attacks: a data-driven approach. ACM Trans. Intell. Syst. Technol. 8:51:1–51: 25 10.1145/2890509 [DOI] [Google Scholar]
  85. Parasuraman R., Rizzo M. (2008). Neuroergonomics: The Brain at Work, Vol. 3 New York, NY: Oxford University Press. [Google Scholar]
  86. Park D. C., Reuter-Lorenz P. (2009). The adaptive brain: aging and neurocognitive scaffolding. Annu. Rev. Psychol. 60, 173–196. 10.1146/annurev.psych.59.103006.093656 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Pattinson M., Jerram C., Parsons K., McCormac A., Butavicius M. (2012). Why do some people manage phishing e-mails better than others? Inform. Manage. Comput. Secur. 20, 18–28. 10.1108/09685221211219173 [DOI] [Google Scholar]
  88. Pendleton M., Garcia-Lebron R., Cho J.-H., Xu S. (2016). A survey on systems security metrics. ACM Comput. Surv. 49, 1–35. 10.1145/3005714 [DOI] [Google Scholar]
  89. Pfleeger S. L., Caputo D. D. (2012). Leveraging behavioral science to mitigate cyber security risk. Comput. Secur. 31, 597–611. 10.1016/j.cose.2011.12.010 [DOI] [Google Scholar]
  90. Pinker S. (2009). How the Mind Works (1997/2009). New York, NY: WW Norton & Company. [Google Scholar]
  91. Purkait S., Kumar De S., Suar D. (2014). An empirical investigation of the factors that influence internet user's ability to correctly identify a phishing website. Inform. Manage. Comput. Secur. 22, 194–234. 10.1108/IMCS-05-2013-0032 [DOI] [Google Scholar]
  92. Rajivan P., Gonzalez C. (2018). Creative persuasion: a study on adversarial behaviors and strategies in phishing attacks. Front. Psychol. 9:135. 10.3389/fpsyg.2018.00135 [DOI] [PMC free article] [PubMed] [Google Scholar]
  93. Redmiles E. M., Chachra N., Waismeyer B. (2018). Examining the demand for spam: who clicks? in Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (Montreal, ON: ACM; ), 212 10.1145/3173574.3173786 [DOI] [Google Scholar]
  94. Rocha Flores W., Holm H., Svensson G., Ericsson G. (2014). Using phishing experiments and scenario-based surveys to understand security behaviours in practice. Inform. Manage. Comput. Secur. 22, 393–406. 10.1108/IMCS-11-2013-0083 [DOI] [Google Scholar]
  95. Salahdine F., Kaabouch N. (2019). Social engineering attacks: a survey. Future Internet 11:89 10.3390/fi11040089 [DOI] [Google Scholar]
  96. Salthouse T. (2012). Consequences of age-related cognitive declines. Annu. Rev. Psychol. 63, 201–226. 10.1146/annurev-psych-120710-100328 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Sample C., Cowley J., Hutchinson S., Bakdash J. (2018). Culture + cyber: exploring the relationship, in Advances in Human Factors in Cybersecurity, AHFE 2017 International Conference on Human Factors in Cybersecurity, eds D. Nicholson (Los Angeles, CA: Springer International Publishing; ), 185–196. 10.4018/978-1-5225-4053-3.ch004 [DOI] [Google Scholar]
  98. Sawyer B. D., Hancock P. A. (2018). Hacking the human: the prevalence paradox in cybersecurity. Human Factors 60, 597–609. 10.1177/0018720818780472 [DOI] [PubMed] [Google Scholar]
  99. Schaie K. W. (2005). What can we learn from longitudinal studies of adult development? Res. Hum. Dev. 2, 133–158. 10.1207/s15427617rhd0203_4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  100. Schechter S. E., Dhamija R., Ozment A., Fischer I. (2007). The emperor's new security indicators, in 2007 IEEE Symposium on Security and Privacy (SP'07) (Oakland, CA: IEEE; ), 51–65. 10.1109/SP.2007.35 [DOI] [Google Scholar]
  101. Schwabe L., Wolf O. T. (2013). Stress and multiple memory systems: from ‘thinking? to ‘doing’. Trends Cogn. Sci. 17, 60–68. 10.1016/j.tics.2012.12.001 [DOI] [PubMed] [Google Scholar]
  102. Shaffer L. (1975). Control processes in typing. Q. J. Exp. Psychol. 27, 419–432. 10.1080/14640747508400502 [DOI] [Google Scholar]
  103. Sharevski F., Treebridge P., Jachim P., Li A., Babin A., Westbrook J. (2019). Social engineering in a post-phishing era: ambient tactical deception attacks. arXiv preprint arXiv:1908.11752. 10.1145/3368860.3368863 [DOI] [Google Scholar]
  104. Sheng S., Holbrook M., Kumaraguru P., Cranor L. F., Downs J. (2010). Who falls for phish?: a demographic analysis of phishing susceptibility and effectiveness of interventions, in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI '10 (New York, NY: ACM; ), 373–382. 10.1145/1753326.1753383 [DOI] [Google Scholar]
  105. Shepherd G. M. (2004). The Synaptic Organization of the Brain. New York, NY: Oxford University Press; 10.1093/acprof:oso/9780195159561.001.1 [DOI] [Google Scholar]
  106. Simons D. J. (2000). Attentional capture and inattentional blindness. Trends Cogn. Sci. 4, 147–155. 10.1016/S1364-6613(00)01455-8 [DOI] [PubMed] [Google Scholar]
  107. Stajano F., Wilson P. (2009). Understanding Scam Victims: Seven Principles for Systems Security. Technical report, University of Cambridge, Computer Laboratory. [Google Scholar]
  108. Starcke K., Brand M. (2012). Decision making under stress: a selective review. Neurosci. Biobehav. Rev. 36, 1228–1248. 10.1016/j.neubiorev.2012.02.003 [DOI] [PubMed] [Google Scholar]
  109. Tembe R., Zielinska O., Liu Y., Hong K. W., Murphy-Hill E., Mayhorn C., et al. (2014). Phishing in international waters: exploring cross-national differences in phishing conceptualizations between Chinese, Indian and American sample, in Proceedings of the 2014 Symposium and Bootcamp on the Science of Security (Raleigh: ACM; ), 8 10.1145/2600176.2600178 [DOI] [Google Scholar]
  110. Tulving E., Craik F. I. (2000). The Oxford Handbook of Memory. New York, NY: Oxford University Press. [Google Scholar]
  111. Tversky A., Kahneman D. (1974). Judgment under uncertainty: heuristics and biases. Science 185, 1124–1131. 10.1126/science.185.4157.1124 [DOI] [PubMed] [Google Scholar]
  112. Valecha R., Gonzalez A., Mock J., Golob E. J., Rao H. R. (2020). Investigating phishing susceptibility–an analysis of neural measures, in Information Systems and Neuroscience (Vienna: Springer; ), 111–119. 10.1007/978-3-030-28144-1_12 [DOI] [Google Scholar]
  113. van der Heijden A., Allodi L. (2019). Cognitive triaging of phishing attacks. arXiv preprint arXiv:1905.02162. [Google Scholar]
  114. Van Schaik P., Jeske D., Onibokun J., Coventry L., Jansen J., Kusev P. (2017). Risk perceptions of cyber-security and precautionary behaviour. Comput. Hum. Behav. 75, 547–559. 10.1016/j.chb.2017.05.038 [DOI] [Google Scholar]
  115. Vishwanath A., Harrison B., Ng Y. J. (2018). Suspicion, cognition, and automaticity model of phishing susceptibility. Commun. Res. 45, 1146–1166. 10.1177/0093650215627483 [DOI] [Google Scholar]
  116. Vishwanath A., Herath T., Chen R., Wang J., Rao H. R. (2011). Why do people get phished? Testing individual differences in phishing vulnerability within an integrated, information processing model. Decis. Support Syst. 51, 576–586. 10.1016/j.dss.2011.03.002 [DOI] [Google Scholar]
  117. Voyer D., Voyer S., Bryden M. P. (1995). Magnitude of sex differences in spatial abilities: a meta-analysis and consideration of critical variables. Psychol. Bull. 117:250. 10.1037/0033-2909.117.2.250 [DOI] [PubMed] [Google Scholar]
  118. Wang J., Herath T., Chen R., Vishwanath A., Rao H. R. (2012). Phishing susceptibility: An investigation into the processing of a targeted spear phishing email. IEEE Trans. Profess. Commun. 55, 345–362. 10.1109/TPC.2012.2208392 [DOI] [Google Scholar]
  119. Wickens C. D. (2008). Multiple resources and mental workload. Human Factors 50, 449–455. 10.1518/001872008X288394 [DOI] [PubMed] [Google Scholar]
  120. Workman M. (2008). Wisecrackers: a theory-grounded investigation of phishing and pretext social engineering threats to information security. J. Am. Soc. Inform. Sci. Technol. 59, 662–674. 10.1002/asi.20779 [DOI] [Google Scholar]
  121. Wright R. T., Jensen M. L., Thatcher J. B., Dinger M., Marett K. (2014). Research note-influence techniques in phishing attacks: an examination of vulnerability and resistance. Inform. Syst. Res. 25, 385–400. 10.1287/isre.2014.0522 [DOI] [Google Scholar]
  122. Wright R. T., Marett K. (2010). The influence of experiential and dispositional factors in phishing: an empirical investigation of the deceived. J. Manage. Inform. Syst. 27, 273–303. 10.2753/MIS0742-1222270111 [DOI] [Google Scholar]
  123. Xu S. (2019). Cybersecurity dynamics: a foundation for the science of cybersecurity, in Proactive and Dynamic Network Defense, Vol. 74, eds Lu Z., Wang C. (Cham: Springer International Publishing; ), 1–31. 10.1007/978-3-030-10597-6_1 [DOI] [Google Scholar]
  124. Zhou Z., Zhao J., Xu K. (2016). Can online emotions predict the stock market in China? in International Conference on Web Information Systems Engineering (Shanghai: Springer; ), 328–342. 10.1007/978-3-319-48740-3_24 [DOI] [Google Scholar]

Articles from Frontiers in Psychology are provided here courtesy of Frontiers Media SA

RESOURCES