Skip to main content
Frontiers in Human Neuroscience logoLink to Frontiers in Human Neuroscience
. 2019 Mar 5;13:85. doi: 10.3389/fnhum.2019.00085

Use of a Portable Functional Near-Infrared Spectroscopy (fNIRS) System to Examine Team Experience During Crisis Event Management in Clinical Simulations

Jie Xu 1,2,*, Jason M Slagle 2,3, Arna Banerjee 2,3, Bethany Bracken 4, Matthew B Weinger 2,3,5
PMCID: PMC6412154  PMID: 30890926

Abstract

Objective: The aim of this study was to investigate the utilization of a portable functional near-infrared spectroscopy (fNIRS) system, the fNIRS PioneerTM, to examine team experience in high-fidelity simulation-based crisis event management (CEM) training for anesthesiologists in operating rooms.

Background: Effective evaluation of team performance and experience in CEM simulations is essential for healthcare training and research. Neurophysiological measures with wearable devices can provide useful indicators of team experience to compliment traditional self-report, observer ratings, and behavioral performance measures. fNIRS measured brain blood oxygenation levels and neural synchrony can be used as indicators of workload and team engagement, which is vital for optimal team performance.

Methods: Thirty-three anesthesiologists, who were attending CEM training in two-person teams, participated in this study. The participants varied in their expertise level and the simulation scenarios varied in difficulty level. The oxygenated and de-oxygenated hemoglobin (HbO and HbR) levels in the participants’ prefrontal cortex were derived from data recorded by a portable one-channel fNIRS system worn by all participants throughout CEM training. Team neural synchrony was measured by HbO/HbR wavelet transformation coherence (WTC). Observer-rated workload and self-reported workload and mood were also collected.

Results: At the individual level, the pattern of HbR level corresponded to changes of workload for the individuals in different roles during different phases of a scenario; but this was not the case for HbO level. Thus, HbR level may be a better indicator for individual workload in the studied setting. However, HbR level was insensitive to differences in scenario difficulty and did not correlate with observer-rated or self-reported workload. At the team level, high levels of HbO and HbR WTC were observed during active teamwork. Furthermore, HbO WTC was sensitive to levels of scenario difficulty.

Conclusion: This study showed that it was feasible to use a portable fNIRS system to study workload and team engagement in high-fidelity clinical simulations. However, more work is needed to establish the sensitivity, reliability, and validity of fNIRS measures as indicators of team experience.

Keywords: functional near-infrared spectroscopy (fNIRS), neural synchrony, team engagement, workload, clinical simulation

Introduction

Simulation Training in Healthcare

Simulation is “a technique that uses a situation or environment created to allow persons to experience a representation of a real event for the purpose of practice, learning, evaluation, testing, or to gain understanding of systems or human actions” (Huang et al., 2008). Simulation has been widely used in complex work system domains, such as aviation, ground transportation, process control, military command and control, and healthcare (Rosen, 2008; Vincenzi et al., 2009; Salas and Maurino, 2010; Fisher et al., 2011; Skjerve and Bye, 2011). In healthcare, simulation exists in many different forms, from verbal simulations with role-playing “what if” discussions, standardized patient visits which involve actors to simulate clinical conversations, part-task trainers which utilize anatomical models, screen-based simulations which involve interactive software-based computer patients, to high-fidelity simulations with replica of the clinical environments and settings (Gaba, 2007; Rosen, 2008). Simulations have been used to address aspects of the clinicians’ knowledge, skill, attitude, behavior, and other characteristics in training, human factors research, and performance assessment (Gaba, 2007; Weinger et al., 2017).

Simulation-based training enables the trainees to learn and practice patient care away from the bedside and without putting a patient at risk (Okuda et al., 2009). This is particularly important for crisis event management (CEM) training, which aims to prepare the trainees for dynamic decision-making and teamwork in high-risk, stressful situations in which the patient’s life is at stake (Gaba et al., 2001; Lighthall et al., 2003; Easdown et al., 2013). A randomized control trial showed that trainees who participated in high-fidelity simulation training performed better in real-life cardiopulmonary bypass than those who participated in traditional interactive seminar (Bruppacher et al., 2010).

During complex high-tempo, high-risk, and multi-person work, effective teamwork is essential to manage and monitor individual and team workload, allocate tasks, and maintain situation awareness. The ability of operating room (OR) teams (i.e., surgeons, anesthesia providers, nurses, techs, and other OR staff) to deliver high quality, safe care to patients depends on acting quickly and effectively, both individually and as a team. Effective training and evaluation must go beyond individual skills to include interactions among team members, and how those interactions transfer to operational environments. In addition, teamwork is a learned skill that can be improved with training (Easdown et al., 2013). The Anesthesia Crisis Resource Management curriculum emphasize the training of individuals to work in teams (Howard et al., 1992; Gaba et al., 2001).

Measuring Performance and Experience in High-Fidelity Simulations

A recent review of simulation-based training research in healthcare concluded that outcome measurement was one of the greatest challenges in the field (McGaghie et al., 2010). There are four main methods to measure team performance and experience in high-fidelity simulations in healthcare: observer ratings, self-reports, behavioral performance measures, and neurophysiological indicators (McGaghie et al., 2010; Doumouras et al., 2012; Forsyth et al., 2017; Robertson et al., 2017). There are advantages and disadvantages for each method. Observer-rated measurement systems, such as the Anaesthetists’ Non-Technical Skills (ANTS) (Fletcher et al., 2003) and the Observational Teamwork Assessment for Surgery (OTAS) (Hull et al., 2011), usually require experienced clinicians or researchers to observe the participants’ behaviors, compare those behaviors with the provided behavioral markers, and produce standardized ratings. However, some systems are easier to use that others (Watkins et al., 2017) and these ratings are subject to potential biases from cognitive, social and environmental sources (Williams et al., 2003). Self-reported measures, such as the Surgery-specific Task Load Index, SURG-TLX (Wilson et al., 2011), and the Perceived Stress Scale (Cohen et al., 1983; Larkin et al., 2010), can provide an understanding of participants’ subjective experience. However, they have to be administered after the fact because except for very brief tools (e.g., the Borg workload scale; Weinger et al., 2004), completing them during a simulation can influence native task performance and even the domain being measured (e.g., workload) (Weinger and Slagle, 2002). Behavioral performance measures and neurophysiological indicators are usually technology-based assessment methods. Behavioral performance can be obtained through haptic sensors or optical systems embedded in the simulator (Rutherford et al., 2015). For example, D’Angelo et al. (2015) used optical motion tracking and analysis to capture the hand movement path length and suture time to show how participants improved their performance over time in a simulated suturing task. Neurophysiological indicators can measure participants’ cognitive states, such as attention and stress. For example, heart rate, respiration rate, and electrodermal activity can be used to measure stress levels (Rutherford et al., 2015). However, these methods are rarely used to measure team-level experience.

Feasible approaches to address the measurement issue in high-fidelity simulations are to integrate multiple measurement methods and to continue to develop new methods, especially technology-based assessment methods (Forsyth et al., 2017). Traditional physiological measures, such as cardiovascular and electrodermal activity measures, are influenced by cognitive, affective, physical movement, and other systemic physiological factors (Boucsein and Backs, 2000). As a result, it is difficult to establish their validity and sensitivity in realistic environments (Xu et al., 2017). Direct measurement of brain activity may be more useful in measuring cognitive status. Recent advances in neuroergonomics, which consider the neural mechanisms of human performance (Parasuraman and Mehta, 2015), and portable neurophysiological sensing technologies provided promising new methods to study human cognition and performance in realistic environments. One of those methods is to use functional near-infrared spectroscopy (fNIRS) to measure the participants’ level of brain activation as an indicator of workload and team engagement.

Functional Near-Infrared Spectroscopy Measures

fNIRS technology infers changes of the oxygenated hemoglobin (HbO) and deoxygenated hemoglobin (HbR) levels in the cortical surface (Ferrari and Quaresima, 2012; Boas et al., 2014). fNIRS technology takes advantage of the fact that the human tissues are relatively transparent to near-infrared (NIR) light and the main absorption agent of NIR light in human brain is hemoglobin. Furthermore, HbO and HbR absorb NIRS light differently depending on the light’s wavelength. Thus, fNIRS technology is able to measure changes of HbO and HbR by shining NIR light with two or more wavelengths, detecting the reflected light, and quantifying the relative light attenuation. The measured changes of HbO and HbR levels can be related to neuronal activity in the corresponding region (Naseer and Hong, 2015). When a brain region is activated due to performance of relevant tasks, the cerebral blood flow (CBF) would increase to meet the metabolism requirement of the brain. While neural activity is associated with the conversion of HbO to HbR, the CBF increase provides an oversupply of HbO and “pushes out” HbR, thus leads to an increase of HbO and decrease of HbR in the brain region (Scholkmann et al., 2014).

Compared to other neural activity measures, such as electroencephalography (EEG) and functional magnetic resonance imaging (fMRI), fNIRS has unique advantages (Derosière et al., 2013). While hemodynamic signal measured by fNIRS is slower in detecting neural activation than the electrical signal measured by EEG, fNIRS provides better space localization capability and better resistance to muscular and movement artifacts than EEG (Naseer and Hong, 2015; Curtin and Ayaz, 2018). fNIRS has lower spatial resolution and penetration depth compared to fMRI, but it has higher temporal resolution; furthermore, the data can be captured using a portable form factor to enable studies in naturalistic environments (Balardin et al., 2017). For example, McKendrick et al. (2016) used a fNIRS system to evaluate portable and wearable device users’ prefrontal cortex (PFC) hemodynamics in outdoor navigation tasks. Yoshino et al. (2013) mounted a fNIRS device in a vehicle to study driver brain activation in actual highway driving.

fNIRS measures have been used as indicators of workload in both laboratory and applied settings. Laboratory studies have found that HbO increases and HbR decreases in the PFC region as workload increases in both n-back working memory tasks (Herff et al., 2014; Keshmiri et al., 2017) and mathematical tasks (Mandrick et al., 2013; Maior et al., 2014). In applied settings, fNIRS was used to measure workload in more complex tasks such as web form usability testing (Lukanov et al., 2016) and air traffic control (Harrison et al., 2014). Afergan et al. (2014) developed an adaptive system for unmanned aerial vehicles that can change task difficulty based on fNIRS-measured metrics of workload. However, few studies have used fNIRS measures as indicators of workload during acute event management in realistic environments.

The development of hyperscanning techniques, which enable the measurement of between-person brain activity dynamics (Scholkmann et al., 2013), facilitated research on groups or teams with fNIRS and other neural activity measures. Studies have investigated neural synchrony ranging from simple tasks, such as synchronizing button pressing or finger movements, to more complex ones, such as music production and trust games (Babiloni and Astolfi, 2014). Neural synchrony can be used as an indicator of the level of collective engagement in shared activities. For example, Osaka et al. (2015) found that pairs of singers showed a higher level of fNIRS neural synchrony in a cooperative singing condition than in a sing-alone condition. In face-to-face communication, Jiang et al. (2012) found that individuals showed higher level of fNIRS neural synchrony in conversations than in monologs; furthermore, Zhang et al. (2018) observed that in psychological counseling, client and counselor dyads showed higher levels of fNIRS neural synchrony during counseling than chatting. Other studies have used neural synchrony to predict team performance (Funane et al., 2011; Stevens et al., 2016). However, to our knowledge, no study to date has used fNIRS-derived neural synchrony to study crisis management teams in high-fidelity simulations.

The Current Study

In the current study, we aimed to investigate the utilization of a portable fNIRS system to examine team experience in high-fidelity simulation-based CEM training for anesthesiologists in operating rooms. Specifically, we used PFC HbO and HbR levels as indicators of workload on the individual level, and neural synchrony as an indicator of team engagement on the team level.

A typical simulation session consisted of a full-scale recreation of a challenging medical crisis situation using a computer-based patient mannequin, real clinical equipment and supplies, actors trained to portray other clinicians (e.g., surgeons, nurses, and technicians), and closely followed scripts, all set in a realistic clinical environment (McIvor et al., 2017; Weinger et al., 2017). Each simulation session consisted of two stages: the 10–20 min simulation scenario itself, which involved two and occasionally three trainees, and a typically 10–25 min instructor-facilitated debriefing, which included all of the trainees in the training session. During a simulation scenario, one trainee started the scenario on their own (the “initial provider”) while a second trainee (the “responder”) was sequestered in a quiet conference room to await being called to assist. Thus, data can be collected under three different conditions: single provider phase (i.e., initial provider only trainee in the simulation room, responder sequestered), team phase (both initial provider and responder in the simulation), and a debriefing phase. Figure 1 depicts the process of a typical simulation session.

FIGURE 1.

FIGURE 1

The process of a simulation session in the current study.

Due to the structure of the simulation sessions and our review of the literature, we speculated that: (1) the workload of the initial provider (in simulation) should be higher than that of the responder (in sequestration) during the single provider phase; (2) the workload of the responder should be higher during the team phase than the single provider phase; and (3) the team should have more team engagement during the team phase (i.e., when working together) than during the single provider phase (not interacting). Therefore, we proposed the following hypotheses:

  • simple

    H-1a: The PFC HbO (HbR) level of the initial provider will be higher (lower) than that of the responder during the single provider phase of the scenario.

  • simple

    H-1b: The PFC HbO (HbR) level of the responder will be higher (lower) in the team phase than in the single provider phase of the scenario.

  • simple

    H-2: The team’s level of neural synchrony will be higher in the team phase than in the single provider phase of the scenario.

In addition to the hypotheses, we concurrently proposed a series of exploratory research questions (RQs) to guide our investigation of the use of fNIRS in the simulations. Specifically, we were interested in questions related to the sensitivity, convergent validity, and discriminant validity of PFC HbO and HbR as indicators of workload on the individual level:

  • simple

    RQ-1a (sensitivity): Is the PFC HbO (HbR) level sensitive to different levels of scenario difficulty?

  • simple

    RQ-1b (sensitivity): What is the pattern of PFC hemodynamics across different time segments in the scenario?

  • simple

    RQ-1c (convergent validity): Is there a significant correlation between the PFC HbO (HbR) and observer-rated or self-reported workload?

  • simple

    RQ-1d (discriminant validity): Is there a significant correlation between the PFC HbO (HbR) and self-reported mood?

We were also interested the sensitivity of PFC neural synchrony as an indicator of team engagement on the team level:

  • simple

    RQ-2a (sensitivity): Is the PFC neural synchrony level sensitive to different levels of scenario difficulty?

  • simple

    RQ-2b (sensitivity): What is the pattern of PFC neural synchrony across different time segments in the scenario?

Materials and Methods

Design

A typical CEM training session consisted of three simulation sessions involving two to four trainees. At any given training session, two to four trainees were enrolled in the study as participants. Most of the simulation scenarios involved two trainees and those were the target of our data collection. The two trainees were randomly assigned as initial provider and responder. All the scenarios included three phases: single provider, team, and debriefing. Each scenario followed one of the fifteen scenario scripts that required the trainees to manage a crisis event. For example, in the “Blown Intravenous (IV)” scenario, the trainees have to ascertain that their usual treatment (via IV drugs) of a deteriorating patient is not working, determine that it is due to an infiltrated IV, and institute alternative (non-IV) treatments. In the “ENT Airway Fire” script, the trainees have to manage a laser-instigated fire in the patient’s mouth and determine what to do next upon extinguishing it.

Trainees with different levels of experience went through different scenarios commensurate with their clinical experience. Scenario difficulty was adjusted to optimize learning (i.e., scenarios for experienced physicians were more difficult than were those for novice physicians). The relative difficulty levels of the scenarios were determined from ratings by three simulation instructors who were all experienced anesthesiologists and simulation instructors. All three instructors had all taught all of the studied scenarios. They rated each scenario’s difficulty on a five-point scale (independent of who might do the scenario). The intra-class correlation (ICC) among the ratings of the three instructors was 0.77, indicating an excellent level of inter-rater reliability (Hallgren, 2012). The rating for each scenario was then averaged across the three instructors. Finally, by applying the Jenks natural breaks classification method (Jenks, 1967) to the ratings, the scenario difficulty was categorized as low, medium, or high for the corresponding experience level of the trainees.

Thus, there were three independent variables (IVs): scenario phase (single provider, team, and debriefing), role (initial provider and responder), and scenario difficulty (low, medium, and high).

Sample

This study was conducted in an academic medical center located in Nashville, TN, United States. The participants were resident and attending physicians in anesthesiology. An attending physician is a board-certified physician who can practice medicine independently. A resident physician is a medical school graduate who is training to become an attending physician. The participating residents were in the first, second, or third year of their residency training program (henceforth, they were referred as Y1, Y2, and Y3 residents). All participants were attending scheduled CEM training sessions at the time of study recruitment. The residents were required to attend the training while the attending physicians chose to do the training as part of their maintenance of certification in anesthesiology (MOCA). Table 1 provides a summary of the characteristics of the 33 individual participants. The study protocol was approved by the medical center’s Institutional Review Board and by the U.S. Army Medical Research and Materiel Command’s Office of Research Protections.

Table 1.

Sample characteristics.

Experience level Year 1 residents Year 2 residents Year 3 residents Attending physicians Total
Sample size 13 5 7 8 33
Age (years) 29.54 ± 2.47a 29.20 ± 1.64 30.14 ± 0.69 42.88 ± 6.08 32.85 ± 6.65
Gender (female) 3 (23.1%) 3 (60.0%) 3 (42.9%) 1 (12.5%) 10 (30.3%)
Prior residency training (months) 7.85 ± 3.44 25.00 ± 8.22 38.57 ± 5.86 19.88 ± 14.54
Prior clinical experience (years)b 9.88 ± 4.02
Prior times doing simulation training 2.54 ± 1.27 4.25 ± 0.50 11.00 ± 1.10 3.13 ± 6.88 4.55 ± 4.75
Prior night’s sleep duration (hours) 6.73 ± 1.02 7.35 ± 1.02 7.54 ± 1.81 7.37 ± 1.35 7.15 ± 1.30
Had difficulty falling asleep (“yes”) 4 (30.8%) 1 (20.0%) 3 (42.9%) 1 (12.5%) 9 (27.3%)
Had caffeinated drinks (“yes”) 11 (84.6%) 3 (60.0%) 5 (71.4%) 8 (100%) 27 (81.8%)

aData in this table are presented as Mean ± SD or Count (Percentage) when applicable. bYears of experience since becoming an attending physician.

A total number of 25 simulation sessions (11, 4, 5, and 5 for Y1 residents, Y2 residents, Y3 residents, and attending physicians, respectively) were included in the analysis. The number of different scenarios were 6, 2, 3, and 4 for Y1, Y2, and Y3 residents, and attending physicians, respectively.

Procedure

Each participant was informed via email about the study opportunity prior to their scheduled training session. Upon arrival to the session, the study procedures and risks were explained, any questions were answered, and interested participants then signed a written informed consent document. Participants then filled out a pre-study survey that included questions about general demographics, clinical experience, and selected factors that could affect their performance and physiologic response to the stress of the training experience (Table 1). The baseline fNIRS measures were taken over a 5 min period where the participant sat still and filled out the pre-study survey.

During each scenario, using a tablet computer with a specialized software program, a trained research assistant (RA) noted any events of interest, including the start and end times of each phase of the scenario session. The RA also rated the initial provider’s workload (see section Workload and Mood) at 4–6 min random intervals as prompted by custom computer software. At the end of each scenario, all participants completed a survey to report their mood and workload during the simulation.

Measures

fNIRS

We used the head-worn fNIRS PioneerTM sensor designed to be part of the MEDIC II system (Charles River Analytics, Cambridge, MA, United States)1 to collect HbO and HbR levels in the PFC. The device consisted of a head-worn sensor probe and a mobile-phone-sized hardware unit. The head-worn sensor probe, embedded in a flexible, black headband, included two light sources, which emitted infrared and visible light, and one light detector. The participants wore the sensor-embedded headband and then wore a surgical cap on top (all the trainees were required to wear surgical caps and scrubs during the CEM training). Pilot tests had demonstrated that this setup could hold the sensor probe stably in place and shield it from ambient light. The probe was secured without adhesion to the participant’s Fp2 region of the PFC, according to the international EEG 10-20 system (Klem et al., 1999). Thus, the fNIRS channel was located on the right side of the anterior PFC. Prior to placement, alcohol was used to clean the skin surface. The probe was connected via a thin cable to a lightweight hardware unit secured to the headband at the back of the participant’s head. The sensor recorded data at a sample rate of 100 Hz.

Workload and Mood

The Borg workload scale was the observer-rated workload measure. The Borg workload scale is a visual analog scale, ranging from 6 (no exertion) to 20 (maximum exertion), that has been shown to yield continuous parametric data (Borg, 1982) and has been used to assess health care providers’ overall workload in clinical settings (Weinger et al., 1994, 1997, 2004). The Borg scale correlates well with physiologic and other measures of workload and stress (Weinger et al., 2004).

The NASA Task Load Index (NASA-TLX) is a self-reported workload measure that includes six dimensions: mental demand, physical demand, temporal demand, performance, effort, and frustration (Hart and Staveland, 1988; Hoonakker et al., 2011). For each dimension, participants rated their experience during the scenario from 0 (“lowest”) to 100 (“highest”). An overall score was calculated by summing the ratings on the six dimensions.

Self-reported mood was measured using the Positive and Negative Affective Schedule (PANAS), a 20-item five-point Likert scale survey instrument (Watson et al., 1988; Crawford and Henry, 2004). A positive affect score and a negative affect score ranging from 10 to 50 were derived from participants’ ratings.

Data Analysis

Case Segmentation

The cases were segmented to extract fNIRS data for subsequent analysis. The first kind of segments was the 1 min period immediately prior to each observer workload rating. The observer typically took less than 1 min to provide the workload rating once prompted by the computer software.

The second kind of segments involved different time points across a scenario. The different scenario phases (i.e., single provider, team, and debriefing) varied appreciably in duration across cases. For example, debriefings varied from 7 to 31 min in duration. Thus, to facilitate data analysis across cases, within each phase, we considered data at three time points (segments): the first minute, the middle minute, and the last minute. These were calculated as the 60 s period at the beginning, the middle, and the end of each scenario phase. The 1 min length was chosen to avoid overlapping time windows as well as provide consistency with the first kind of segmentation.

fNIRS Data Processing

The raw data from the MEDIC II fNIRS module included changes in optical density (OD) for two channels of light with different wavelengths. Changes in concentrations of HbO and HbR were calculated from the OD data using the modified Beer-Lambert Law (Cope and Delpy, 1988). We applied the wavelet-based motion artifact removal procedure with a tuning parameter value α = 0.15 to all HbO and HbR time series (Behnam and Guy, 2012; Cooper et al., 2012). We accounted for global drift (low-frequency noise) and high-frequency noise, by applying a band-pass third-order Butterworth filter with 0.01–0.1 Hz cutoff frequencies. After artifact removal and filtering, the data were visually inspected to ensure data quality; no manual corrections were made. The HbO and HbR data were then normalized within each participant. To express HbO and HbR data at any given moment as a relative change from the baseline, the mean HbO and HbR value during the baseline period was subtracted from all recorded data before further data processing and analysis. Data processing was performed in R (R Core Team, 2018) using the “signal” (Ligges et al., 2015) and “wavelets” (Aldrich, 2013) packages.

For the analysis at the individual level, mean levels of HbO and HbR were calculated for the time periods of interests. First, mean HbO and HbR were calculated for each of the three phases in each simulation session. Then, mean HbO and HbR were calculated for the first, middle, and last minute of each phase in a simulation session. In addition, mean HbO and HbR were calculated for the 1 min period immediately prior to each observer workload rating.

For the analysis on the team level, the wavelet transformation coherence (WTC) (Torrence and Compo, 1998; Grinsted et al., 2004) of the HbO and HbR between the two participants in a scenario session were calculated as indicators of neural synchrony (Cui et al., 2012; Osaka et al., 2015; Baker et al., 2016; Nozawa et al., 2016). WTC measures the cross-correlation of two time series based on the continuous wavelet transform at the given frequency and time. The HbO and HbR time series were re-sampled to 1 Hz for this analysis. The Morlet wavelet function with the parameter ω0 = 6 was used in the wavelet transformation. Only the coherence values outside of the cone of influence (COI) was considered to control the edge-effects of the WTC estimation. The frequency band to be considered in WTC varied in previous fNIRS hyperscanning studies depending on the task-related frequency band of each study, however, task-related bands are difficult to define for activities occur in natural and unstructured settings (Nozawa et al., 2016). This study selected a relatively wide frequency range of 0.01–0.1 Hz according to previous studies in natural settings (Jiang et al., 2012; Zhang et al., 2018). Mean coherence values at 0.01–0.1 Hz were calculated for the baseline and other time periods of interest (i.e., the 1 min periods prior to the observed workload ratings, the different scenario phases, and the three segments within each scenario phase). An illustrative example of the results of WTC analysis from one scenario session is shown in Figure 2. The final WTC value for each time period was calculated as the mean coherence value of that time period minus the mean coherence value of the corresponding baseline period. The WTC analysis was performed in R using the “biwavelet” package (Gouhier et al., 2018).

FIGURE 2.

FIGURE 2

An illustrative example of the results of WTC analysis for HbO (A) and HbR (B) from a scenario session.

Statistical Analysis

The main statistical analysis was conducted using Bayesian linear mixed effects (LME) models with Markov Chain Monte Carlo (MCMC) estimation (Hadfield, 2010). The LME models accounted for the fixed effects of the IVs as well as the random effects of the participants, teams, or scenarios. Table 2 shows the model specifications for the LME models used in the analysis. The random effects were modeled as random intercepts. These statistical analyses were conducted using R with the “MCMCglmm” package (Hadfield, 2010). The MCMCglmm algorithm used non-informative priors and default parameter estimation settings (iterations = 13,000, thinning interval = 10, burn-in = 3,000). All the models were checked for convergence and autocorrelation of the estimates. An effect was considered significant if the 95% highest density interval (HDI) did not contain zero and the Bayesian “p-value” pMCMC was smaller than 0.05.

Table 2.

LME model specifications.

Model ID Dependent variable Independent variables (fixed effects) Independent variables (random effects)
1 HbO or HbR mean • Scenario phase • Participant
• Role • Team
• Scenario difficulty • Scenario
• All possible interactions of the previous variables
• Experience level
2 HbO or HbR mean • Scenario phase (with first/middle/last minute) • Participant
• Role • Team
• Scenario difficulty • Scenario
• All possible interactions of the previous variables
• Experience level
3 Observer-rated workload • Scenario phase • Participant
• Scenario difficulty • Team
• All possible interactions of the previous variables • Scenario
• Experience level
4 Self-reported workload, positive mood, or negative mood • Role • Participant
• Scenario difficulty • Team
• All possible interactions of the previous variables • Scenario
• Experience level
5 HbO or HbR WTC • Scenario phase • Team
• Scenario difficulty • Scenario
• All possible interactions of the previous variables
• Experience level
6 HbO or HbR WTC • Scenario phase (with first/middle/last minute) • Team
• Scenario difficulty • Scenario
• All possible interactions of the previous variables
• Experience level

The correlations between the dependent variables (DVs) were estimated using a multivariate multilevel modeling approach (Baldwin et al., 2014; Luo et al., 2015). The multivariate multilevel models were fitted using MCMCglmm. A correlation was considered significant if the 95% HDI did not contain zero.

Results

Individual Level Results

H-1a and H-1b Were Supported by HbR but Not HbO

The effects of scenario phase, role, and scenario difficulty on the mean HbO and HbR were tested using Model 1 (Table 2). The results are visualized in Figure 3. The difference of HbO mean between initial provider and responder in the single provider phase was not significant (b = 0.02, 95% HDI = [-0.04, 0.09], pMCMC = 0.43). The difference of the HbO mean of the responder between the single provider phase and the team phase was also not significant (b = 0.04, 95% HDI = [-0.01, 0.10], pMCMC = 0.18). HbR mean of the initial provider was significantly lower than the responder in the single provider phase (b = 0.14, 95% HDI = [0.02, 0.23], pMCMC = 0.01). The responder’s HbR mean was significantly lower in the team phase than in the single provider phase (b = 0.12, 95% HDI = [0.02, 0.20], pMCMC = 0.01).

FIGURE 3.

FIGURE 3

The predicted means and 95% HDIs of the initial provider and the responder’s HbO (A) and HbR (B) levels at different scenario phases. The values of all the other variables were held at their means.

In summary, H-1a and H-1b were both supported by the results from the HbR but not the HbO data.

The Effect of Scenario Difficulty (RQ-1a)

Scenario difficulty influenced the effect of scenario phases on HbO mean. When scenario difficulty was held at its mean, HbO mean was higher in the team/debriefing phases than in the single provider phase (b = 0.05, 95% HDI = [0.01, 0.09], pMCMC = 0.01) for both the initial provider and responder. This effect was smaller in the medium/high difficulty than in the low difficulty (b = -0.13, 95% HDI = [-0.21, -0.05], pMCMC < 0.01); so that the difference between the team/debriefing phases and the single provider phase was not significant in both medium difficulty (b = 0.003, 95% HDI = [-0.05, 0.06], pMCMC = 0.94) and high difficulty (b = 0.007, 95% HDI = [-0.05, 0.07], pMCMC = 0.85). These significant effects were not observed in HbR mean.

In summary, an HbO increase in team/debriefing phase compared to the initial provider phase was observed in the low difficulty scenarios but not in the medium/high difficulty scenarios. No scenario difficulty moderation effect was observed in the HbR data.

The Hemodynamics Across Time Segments (RQ-1b)

Model 2 (Table 2) was used to analyze the effect of scenario phase (using first, middle, and last minute) on HbO and HbR. The results are visualized in Figure 4. For HbO, a small but significant increase was observed at the last minute of the team phase compared to the last minute of the single provider phase (b = 0.11, 95% HDI = [0.03, 0.19], pMCMC < 0.01). There was a significant interaction effect that the initial providers’ HbO increased in the first minute of the debriefing phase compared to the last minute of the team phase, while the responders’ HbO decreased (b = 0.23, 95% HDI = [0.07, 0.37], pMCMC < 0.01). Responders showed higher levels of HbR than the initial providers in the first (b = -0.18, 95% HDI = [-0.33, -0.03], pMCMC = 0.02) and last (b = -0.16, 95% HDI = [-0.31, -0.01], pMCMC = 0.04) minutes of the single provider phase. The HbR level of the responder decreased in the first minute of the team phase compared to the last minute of the single provider phase, while the HbR level of the initial provider increased in the same period (b = 0.29, 95% HDI = [0.09, 0.49], pMCMC < 0.01).

FIGURE 4.

FIGURE 4

The predicted means and 95% HDIs of the initial provider and the responder’s HbO (A) and HbR (B) levels at the first/middle/last minutes (FM, MM, and LM) of the different scenario phases. The values of all the other variables were held at their means.

In summary, the hemodynamics analysis across time segments revealed additional information regarding the HbO/HbR change in different time points across the scenarios and the different patterns of HbO/HbR change related to different roles.

The Correlations Among HbO, HbR, Workload, and Mood (RQ-1c and RQ-1d)

The effects of the IVs on the observer-rated workload and self-reported workload and mood were first tested. The results from Model 3 (Table 2) indicated that there was no significant effect of scenario phase and scenario difficulty on observer-rated workload of the initial provider. Model 4 (Table 2) was used to explore the effects of role and scenario difficulty on self-reported workload and mood. No significant effect of the IVs on self-reported workload or negative mood was observed. Initial providers reported higher positive mood than responders in low difficulty scenarios (b = 10.95, 95% HDI = [3.75, 17.29], pMCMC < 0.01).

A multivariate model based on Model 3 was fitted to the data to explore the correlations between HbO/HbR and observer-rated workload. Mean levels of HbO and HbR were calculated for 1 min prior to the moment the observer-rated workload rating was given. The correlations were not significant (-0.05 < r < 0.05). Correlations among HbO, HbR, and self-reported workload and mood were, similarly, tested based on Model 4. Using mean levels in the team phase, HbO and HbR correlated poorly with the self-reported metrics (-0.05 < r < 0.05). The only significant correlation was between self-reported workload and negative mood (r = 0.39, 95% HDI = [0.07, 0.66]).

In summary, HbO/HbR were not significantly correlated with the workload and mood measures.

Team Level Results

H-2 Was Supported by Both HbO and HbR WTCs

Model 5 (Table 2) was fitted to the data to test the effects of scenario phase and difficulty on HbO and HbR WTCs. WTC was higher in the team phase than in the single provider phase for both HbO (b = 0.08, 95% HDI = [0.06, 0.11], pMCMC < 0.01) and HbR (b = 0.08, 95% HDI = [0.05, 0.10], pMCMC < 0.01). Also note that HbO WTC was significantly higher in the team phase than in the debriefing phase (b = 0.04, 95% HDI = [0.01, 0.07], pMCMC < 0.01), while the same effect was not significant for HbR WTC (b = 0.02, 95% HDI = [-0.01, 0.04], pMCMC = 0.17).

In summary, H-2 was supported by findings from both HbO and HbR data.

The Effect of Scenario Difficulty (RQ-2a)

The interaction effect showed that the HbO WTC difference between the single provider phase and the team phase was larger in the high difficulty scenarios than in the low difficulty scenarios (b = 0.07, 95% HDI = [0.01, 0.14], pMCMC = 0.04). No significant effect associated with scenario difficulty was observed for HbR WTC. The results are visualized in Figure 5.

FIGURE 5.

FIGURE 5

The predicted means and 95% HDIs of the WTCs of HbO (A) and HbR (B) between the initial provider and the responder in a team at different scenario phases. The values of all the other variables were held at their means.

In summary, more difficult scenarios were correlated with an increased HbO WTC in team phase compared to the single provider phase. However, similar effects were not observed in HbR WTC.

The Neural Synchrony Across Time Segments (RQ-2b)

The temporal dynamics of the HbO/HbR WTCs in the first, middle, and last minutes of the different scenario phases were explored using Model 6 (Table 2). As seen in Figure 6, for HbO, the WTC level increased in the last minute compared to the middle minute of the single provider phase (b = 0.05, 95% HDI = [0.01, 0.08], pMCMC = 0.02). The WTC also increased in the last minute compared to the middle minute of the team phase (b = 0.11, 95% HDI = [0.08, 0.15], pMCMC < 0.01). WTC then decreased in the last minute compared to the middle minute of the debriefing phase (b = -0.16, 95% HDI = [-0.19, -0.12], pMCMC < 0.01). The HbR WTC showed a similar pattern; there was a significant increase in the last minute compared to the middle minute of the team phase (b = 0.10, 95% HDI = [0.06, 0.14], pMCMC < 0.01) and a decrease in the last minute compared to the middle minute of the debriefing phase (b = -0.15, 95% HDI = [-0.18, -0.11], pMCMC < 0.01).

FIGURE 6.

FIGURE 6

The predicted means and 95% HDIs of the WTCs of HbO (A) and HbR (B) between the initial provider and the responder in a team at first/middle/last minutes (FM, MM, and LM) of the different scenario phases. The values of all the other variables were held at their means.

In summary, dramatic HbO/HbR WTC increases were observed around the end of the team phase and the high WTC levels maintained until around the end of the debriefing phase.

Discussion

The current study used a portable fNIRS system, the fNIRS PioneerTM, to measure PFC hemodynamics in clinicians doing CEM high-fidelity simulations. On the individual level, we found that the pattern of HbR mean for the different roles in different scenario phases was consistent with our hypotheses while HbO mean was not. However, HbR mean was not sensitive to different levels of scenario difficulty. Furthermore, the HbO and HbR means were not significantly correlated with any of the observer-rated or self-reported workload and mood measures. The teams showed higher HbO/HbR neural synchrony during periods of teamwork, supporting our hypothesis. In addition, HbO neural synchrony was sensitive to different levels of scenario difficulty; higher level of neural synchrony was observed in scenarios with higher level of difficulty. Analysis of the time segments indicated that the neural synchrony of the teams increased dramatically in the last minute of the team phase compared to previous segments, and this higher level was maintained until the end of the debriefing phase.

fNIRS Measured Brain Activity as Indicator of Workload for Individual Clinicians

fNIRS Measures in Different Scenario Phases

The findings suggest that HbR may be a better indicator for workload than HbO in the current study setting. Regarding the use of HbO vs. HbR, some prior studies found that HbR was more sensitive than HbO (e.g., Peck et al., 2013), however, some studies have reported that HbO and HbR are, similarly, sensitive to changes in workload (e.g., Herff et al., 2014; McKendrick et al., 2016) while others have suggested that HbR is less sensitive than HbO (e.g., Derosière et al., 2015). In the brain-computer interface studies, HbR was less sensitive for real time cortical activation detection (Naseer et al., 2014) and HbO provided better test-retest reliability (Plichta et al., 2006). Overall, the results presented in the literature are not conclusive.

Physiologically, increased cognitive workload typically associates with increased activation in the PFC region (Parasuraman and Caggiano, 2005), as a result, one should observe an HbO increase and an HbR decrease in high workload conditions compared to low workload conditions. However, fNIRS measures are influenced not only by the neurovascular coupling which reflects cerebral neuronal activity, but also systemic physiological changes in the cerebral and extracerebral compartments (Tachtsidis and Scholkmann, 2016). Studies have found that anesthesiologists’ workload correlates with systemic physiological changes, such as heart rate and heart rate variability (Weinger and Slagle, 2002; Weinger et al., 2004). Thus, in such a study setting, as HbR may be less influenced by global hemodynamic changes (Kirilina et al., 2012), when the signal-to-noise ratio is optimized, it can be a reliable indicator for task-related cortical activation (Piper et al., 2014).

fNIRS Measures in Scenarios With Different Difficulty Level

While HbR level was able to differentiate periods of resting (i.e., responder in the single provider phase) and working (i.e., responder in the team phase and initial provider in the single provider phase) and revealed an interesting pattern across time segments within a scenario, it was not sensitive to scenario difficulty. It was assumed that the more difficult a scenario was, the more cognitive resource the participants must use to work. In more difficult scenarios, the participants’ PFC may activate to a greater extent or for longer durations. As a result, the mean HbO level should be higher and the mean HbR level should be lower over the scenario period. For example, studies have found the associations between fNIRS measured hemodynamics and different levels of task difficulty in driving (Foy and Chapman, 2018) and flight (Gateau et al., 2015). However, anesthesiology work involves more physical activities, which introduces additional systemic physiological changes, than driving and flight. More research is needed to improve the sensitivity and specificity of the fNIRS measures before they will be useful for evaluating workload in this domain.

Correlations With Other Measures of Workload

In the present study, neither HbO nor HbR correlated with observer-rated or self-reported workload. Previous studies have found that different workload measures are sensitive, but they have poor correlation with each other (Myrtek et al., 1994; Matthews et al., 2015). Workload is a multidimensional construct; therefore, different measures may reflect different dimensions of workload (Matthews et al., 2015; Young et al., 2015). At the same time, different measurement methods may provide unique advantages in certain situations. Based on changes in HbR levels, this study found that the initial providers’ workload decreased in the transition between the single provider phase and the team phase, while the responders’ workload increased during the same time period. These types of dynamic changes would be very difficult to capture with self-reported workload measures.

Neural Synchrony in Anesthesiology Teams

fNIRS Neural Synchrony as a Measure of Team Engagement

The results from the team level neural synchrony analysis showed that fNIRS measures are feasible and sensitive indicators of team engagement. Both HbO and HbR WTCs were able to differentiate periods of individual work (i.e., single provider phase) from teamwork (i.e., team phase). These findings are consistent with previous study results regarding cooperative work in other settings, such as laboratory n-back task (Dommer et al., 2012) and singing (Osaka et al., 2015), where significantly higher levels of fNIRS PFC neural synchrony were observed in cooperative work compared to individual work.

Furthermore, HbO WTC was sensitive to changes in scenario difficulty. More difficult scenarios required closer cooperation between the team members compared to less difficult ones, thus the team members might have shown a higher level of engagement in teamwork. The increased level of team engagement corresponded to an increased level of neural synchrony. For example, Toppi et al. (2016) have found that pairs of pilots showed an increased level of EEG neural synchrony in the frontal and parietal brain areas in more cooperation-demanding task scenarios in simulated flight. Zhang et al. (2018) found that the fNIRS neural synchrony in the right temporo-parietal junction area between psychological counselors and clients increased as the dyads increased their engagement in the conversation. This study demonstrated that fNIRS neural synchrony can be useful in detecting team engagement levels in anesthesiology teams. In practice, fNIRS neural synchrony may be useful as a quantitative guide to the design of simulation scenario difficulty to fit trainee expertise level. Note that although HbO was less sensitive than HbR in measuring workload (possibly due to systemic physiological noise), it was more sensitive in measuring team engagement. Future research should investigate how systemic physiological noise may influence WTC and why HbR was a less sensitive team engagement measure.

Temporal Dynamics of Neural Synchrony in Anesthesiology Scenarios

This study revealed interesting patterns in the level of neural synchrony across different time segments within a simulation session. First, there was a dramatic increase in neural synchrony at the end of the team phase, when the teams were actively coordinating to implement treatment in the face of a crisis situation. Simulation scenarios are written to allow the instructor to conclude when the team has achieved key treatment milestones. Thus, one might postulate that the team neural synchrony is an indicator of a coherent team mental model that either allowed or was a consequence of successful interventions. Future studies should examine team neural synchrony in successful vs. unsuccessful event management. Second, there was a sustained level of neural synchrony throughout most of the debriefing. One of the main goals of instructor-facilitated debriefing is to achieve a common understanding (i.e., common ground) of what happened during the scenario. The debriefing is a period of high participant engagement in the discussion regarding what happened and why, and how performance could be improved. Third, the level of neural synchrony decreased in the last segment of the debriefing phase when the group was wrapping up the discussion and either preparing for the next simulation session or concluding the day’s training.

Limitations

This study has limitations. First, the study sample size was small due to limited supplies of clinicians and other logistical and technological constraints. As a result, we were not able to directly compare clinician groups of different expertise levels and had to control their effects statistically. Second, the three 1 min time points in a scenario phase may not have provided enough information on the temporal dynamics of HbO/HbR means and WTCs. Third, due to logistic restrictions, baseline fNIRS measure was collected when the participants were filling out surveys. Compared to a traditional resting baseline (Fishel et al., 2007), our baseline involved a certain degree of cognitive activity. Individual variations in cortical activation during the baseline might have decreased the power to detect significant changes over time. Fourth, the portable fNIRS system used in the current study only provided one channel of HbO/HbR data from one light source – detector pair. Without multiple data channels, we had to use univariate methods to process the fNIRS data which is not as robust as multivariate methods in removing artifacts from sources such as motion and physiological noise (Scholkmann et al., 2014; Naseer and Hong, 2015).

Conclusion

In this paper we studied a portable fNIRS system, the fNIRS PioneerTM, to measure team experience in high-fidelity CEM clinical simulation. We found that HbR level may be a useful indicator of workload on the individual level, but its sensitivity needs further study. Neural synchrony based on HbO/HbR WTCs appears to be a promising fNIRS measure of team engagement. The use of a portable fNIRS system in CEM is feasible but challenging. In practice, fNIRS may best be used in combination with other physiological and psychological (i.e., observer ratings and self-reports) measures, to provide a comprehensive perspective on team experience and performance. Specific direction for future studies include: (1) test the use of fNIRS in CEM simulation with a larger sample size to compare the effect of clinician expertise; (2) compare neural synchrony measures with other team level measures, such as observer-rated and self-reported teamwork and team performance; (3) improve the design of the portable fNIRS system to achieve a balance of size, weight, performance (e.g., more available data channels), and usability in realistic high-tempo practice environments.

Data Availability

The datasets generated for this study are available on request to the corresponding author.

Ethics Statement

This study was carried out in accordance with the recommendations of Vanderbilt Human Research Protection Program policies and procedures, Behavioral Sciences Committee – Institutional Review Board, with written informed consent from all subjects. All subjects gave written informed consent in accordance with the Declaration of Helsinki. The protocol was approved by the Vanderbilt University Behavioral Sciences Committee – Institutional Review Board.

Author Contributions

JX, JS, AB, BB, and MW designed the research. JX and JS performed data collection with the assistance from AB and MW. JX analyzed the data and drafted the manuscript. BB provided technical support for data collection and analysis. The other authors revised the manuscript.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The views, opinions, and/or findings contained herein are those of the authors and should not be construed as an official Department of the Army position, policy or decision unless so designated by other documentation. In the conduct of research where humans are the subjects, the investigator(s) adhered to the policies regarding the protection of human subjects as prescribed by Code of Federal Regulations (CFR) Title 45, Volume 1, Part 46; Title 32, Chapter 1, Part 219; and Title 21, Chapter 1, Part 50 (Protection of Human Subjects). The technical assistance of Katherine Tippey, Ph.D. and Tao Liu, Ph.D. were greatly appreciated.

Funding. This work was supported in part by a subcontract to Vanderbilt University Medical Center from Charles River Analytics (Cambridge, MA) under Contract W81XWH-14-C-0018 with the United States Army Medical Research and Materiel Command (USAMRMC), Fort Detrick, MD. During data analysis and manuscript preparation, some of MW time was supported by Vanderbilt University Medical Center and by the VA Tennessee Valley Healthcare System while some of JX time was supported by Zhejiang University.

References

  1. Afergan D., Peck E. M., Solovey E. T., Jenkins A., Hincks S. W., Brown E. T., et al. (2014). “Dynamic difficulty using brain metrics of workload,” in Proceedings of the 32nd Annual ACM Conference on Human Factors in Computing Systems Toronto, ON: 3797–3806. 10.1145/2556288.2557230 [DOI] [Google Scholar]
  2. Aldrich E. (2013). Wavelets: A Package of Functions for Computing Wavelet Filters, Wavelet Transforms and Multiresolution Analyses. Available at: https://CRAN.R-project.org/package=wavelets [Google Scholar]
  3. Babiloni F., Astolfi L. (2014). Social neuroscience and hyperscanning techniques: past, present and future. Neurosci. Biobehav. Rev. 44 76–93. 10.1016/j.neubiorev.2012.07.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Baker J. M., Liu N., Cui X., Vrticka P., Saggar M., Hosseini S. M. H., et al. (2016). Sex differences in neural and behavioral signatures of cooperation revealed by fNIRS hyperscanning. Sci. Rep. 6:26492. 10.1038/srep26492 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Balardin J. B., Zimeo Morais G. A., Furucho R. A., Trambaiolli L., Vanzella P., Biazoli C., et al. (2017). imaging brain function with functional near-infrared spectroscopy in unconstrained environments. Front. Hum. Neurosci. 11:258. 10.3389/fnhum.2017.00258 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Baldwin S. A., Imel Z. E., Braithwaite S. R., Atkins D. C. (2014). Analyzing multiple outcomes in clinical research using multivariate multilevel models. J. Consult. Clin. Psychol. 82 920–930. 10.1037/a0035628 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Behnam M., Guy A. D. (2012). Wavelet-based motion artifact removal for functional near-infrared spectroscopy. Physiol. Meas. 33 259–270. 10.1088/0967-3334/33/2/259 [DOI] [PubMed] [Google Scholar]
  8. Boas D. A., Elwell C. E., Ferrari M., Taga G. (2014). Twenty years of functional near-infrared spectroscopy: introduction for the special issue. Neuroimage 85 1–5. 10.1016/j.neuroimage.2013.11.033 [DOI] [PubMed] [Google Scholar]
  9. Borg G. A. (1982). Psychophysical bases of perceived exertion. Med. Sci. Sports Exerc. 14 377–381. 10.1249/00005768-198205000-00012 [DOI] [PubMed] [Google Scholar]
  10. Boucsein W., Backs R. W. (2000). “Engineering psychophysiology as a discipline: historical and theoretical aspects,” in Engineering Psychophysiology. Issues and Applications eds Backs R. W., Boucsein W. (Mahwah, NJ: Lawrence Erlbaum Associates, Inc; ) 3–30. [Google Scholar]
  11. Bruppacher H. R., Alam S. K., LeBlanc V. R., Latter D., Naik V. N., Savoldelli G. L. (2010). Simulation-based training improves physicians’ performance in patient care in high-stakes clinical setting of cardiac surgery. Anesthesiology 112 985–992. 10.1097/ALN.0b013e3181d3e31c [DOI] [PubMed] [Google Scholar]
  12. Cohen S., Kamarck T., Mermelstein R. (1983). A global measure of perceived stress. J. Health Soc. Behav. 24 385–396. 10.2307/2136404 [DOI] [PubMed] [Google Scholar]
  13. Cooper R., Selb J., Gagnon L., Phillip D., Schytz H., Iversen H., et al. (2012). A systematic comparison of motion artifact correction techniques for functional near-infrared spectroscopy. Front. Neurosci. 6:147. 10.3389/fnins.2012.00147 [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Cope M., Delpy D. T. (1988). System for long-term measurement of cerebral blood and tissue oxygenation on newborn infants by near infra-red transillumination. Med. Biol. Eng. Comput. 26 289–294. 10.1007/BF02447083 [DOI] [PubMed] [Google Scholar]
  15. R Core Team (2018). R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing. [Google Scholar]
  16. Crawford J. R., Henry J. D. (2004). The Positive and Negative Affect Schedule (PANAS): construct validity, measurement properties and normative data in a large non-clinical sample. Br. J. Clin. Psychol. 43 245–265. 10.1348/0144665031752934 [DOI] [PubMed] [Google Scholar]
  17. Cui X., Bryant D. M., Reiss A. L. (2012). NIRS-based hyperscanning reveals increased interpersonal coherence in superior frontal cortex during cooperation. Neuroimage 59 2430–2437. 10.1016/j.neuroimage.2011.09.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Curtin A., Ayaz H. (2018). The Age of Neuroergonomics: towards ubiquitous and continuous measurement of brain function with fNIRS. Jpn. Psychol. Res. 60 374–386. 10.1111/jpr.12227 [DOI] [Google Scholar]
  19. D’Angelo A.-L. D., Rutherford D. N., Ray R. D., Mason A., Pugh C. M. (2015). Operative skill: quantifying surgeon’s response to tissue properties. J. Surg. Res. 198 294–298. 10.1016/j.jss.2015.04.078 [DOI] [PubMed] [Google Scholar]
  20. Derosière G., Billot M., Ward T. E., Perrey S. (2015). Adaptations of motor neural structures’ activity to lapses in attention. Cereb. Cortex 25 66–74. 10.1093/cercor/bht206 [DOI] [PubMed] [Google Scholar]
  21. Derosière G., Mandrick K., Dray G., Ward T. E., Perrey S. (2013). NIRS-measured prefrontal cortex activity in neurogonomics: strengths and weaknesses. Front. Hum. Neurosci. 7:583. 10.3389/fnhum.2013.00583 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Dommer L., Jäger N., Scholkmann F., Wolf M., Holper L. (2012). Between-brain coherence during joint n-back task performance: a two-person functional near-infrared spectroscopy study. Behav. Brain Res. 234 212–222. 10.1016/j.bbr.2012.06.024 [DOI] [PubMed] [Google Scholar]
  23. Doumouras A. G., Keshet I., Nathens A. B., Ahmed N., Hicks C. M. (2012). A crisis of faith? A review of simulation in teaching team-based, crisis management skills to surgical trainees. J. Surg. Educ. 69 274–281. 10.1016/j.jsurg.2011.11.004 [DOI] [PubMed] [Google Scholar]
  24. Easdown L. J., Banerjee A., Weinger M. B. (2013). “Simulation to assess human responses to critical events,” in The Oxford Handbook of Cognitive Engineering eds Lee J. D., Kirlik A. (New York, NY: Oxford University Press; ) 336–351. [Google Scholar]
  25. Ferrari M., Quaresima V. (2012). A brief review on the history of human functional near-infrared spectroscopy (fNIRS) development and fields of application. Neuroimage 63 921–935. 10.1016/j.neuroimage.2012.03.049 [DOI] [PubMed] [Google Scholar]
  26. Fishel S. R., Muth E., Hoover A. W. (2007). Establishing appropriate physiological baseline procedures for real-time physiological measurement. J. Cognit. Eng. Decis. Mak. 1 286–308. 10.1518/155534307X255636 [DOI] [Google Scholar]
  27. Fisher D. L., Rizzo M., Caird J., Lee J. D. (2011). Handbook of Driving Simulation for Engineering, Medicine, and Psychology. Boca Raton, FL: CRC Press. [Google Scholar]
  28. Fletcher G., Flin R., McGeorge P., Glavin R., Maran N., Patey R. (2003). Anaesthetists’ Non-Technical Skills (ANTS): evaluation of a behavioural marker system. Br. J. Anaesth. 90 580–588. 10.1093/bja/aeg112 [DOI] [PubMed] [Google Scholar]
  29. Forsyth K. L., D’Angelo A.-L., Cohen E. M., Pugh C. M. (2017). “Improving clinical performance by analyzing surgical skills and operative errors,” in Surgical Patient Care: Improving Safety, Quality and Value eds Sanchez J. A., Barach P., Johnson J. K., Jacobs J. P. (Cham: Springer International Publishing; ) 555–568. [Google Scholar]
  30. Foy H. J., Chapman P. (2018). Mental workload is reflected in driver behaviour, physiology, eye movements and prefrontal cortex activation. Appl. Ergonom. 73 90–99. 10.1016/j.apergo.2018.06.006 [DOI] [PubMed] [Google Scholar]
  31. Funane T., Kiguchi M., Atsumori H., Sato H., Kubota K., Koizumi H. (2011). Synchronous activity of two people’s prefrontal cortices during a cooperative task measured by simultaneous near-infrared spectroscopy. J. Biomed. Opt. 16:077011. 10.1117/1.3602853 [DOI] [PubMed] [Google Scholar]
  32. Gaba D. M. (2007). The future vision of simulation in healthcare. Simul. Healthc. 2 126–135. 10.1097/01.sih.0000258411.38212.32 [DOI] [PubMed] [Google Scholar]
  33. Gaba D. M., Howard S. K., Fish K. J., Smith B. E., Sowb Y. A. (2001). Simulation-based training in Anesthesia Crisis Resource Management (ACRM): a decade of experience. Simul. Gaming 32 175–193. 10.1177/104687810103200206 [DOI] [Google Scholar]
  34. Gateau T., Durantin G., Lancelot F., Scannella S., Dehais F. (2015). Real-time state estimation in a flight simulator using fNIRS. PLoS One 10:e0121279. 10.1371/journal.pone.0121279 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Gouhier T. C., Grinsted A., Simko V. (2018). R Package Biwavelet: Conduct Univariate and Bivariate Wavelet Analyses. Available at: https://github.com/tgouhier/biwavelet [Google Scholar]
  36. Grinsted A., Moore J. C., Jevrejeva S. (2004). Application of the cross wavelet transform and wavelet coherence to geophysical time series. Nonlin. Process. Geophys. 11 561–566. 10.5194/npg-11-561-2004 [DOI] [Google Scholar]
  37. Hadfield J. D. (2010). MCMC methods for multi-response generalized linear mixed models: the MCMCglmm R Package. J. Stat. Softw. 33 1–22. 10.18637/jss.v033.i0220808728 [DOI] [Google Scholar]
  38. Hallgren K. A. (2012). Computing inter-rater reliability for observational data: an overview and tutorial. Tutor. Quant. Methods Psychol. 8 23–34. 10.20982/tqmp.08.1.p023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Harrison J., Izzetoglu K., Ayaz H., Willems B., Hah S., Ahlstrom U., et al. (2014). Cognitive workload and learning assessment during the implementation of a next-generation air traffic control technology using functional near-infrared spectroscopy. IEEE Trans. Hum. Mach. Syst. 44 429–440. 10.1109/thms.2014.2319822 [DOI] [Google Scholar]
  40. Hart S. G., Staveland L. E. (1988). “Development of NASA-TLX (Task Load Index): Results of empirical and theoretical research,” in Human Mental Workload eds Hancock P. A., Meshkati N. (Amsterdam: North Holland Press; ). [Google Scholar]
  41. Herff C., Heger D., Fortmann O., Hennrich J., Putze F., Schultz T. (2014). Mental workload during n-back task-quantified in the prefrontal cortex using fNIRS. Front. Hum. Neurosci. 7:935. 10.3389/fnhum.2013.00935 [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Hoonakker P., Carayon P., Gurses A. P., Brown R., Khunlertkit A., McGuire K., et al. (2011). Measuring workload of ICU nurses with a questionnaire survey: the NASA Task Load Index (TLX). IIE Trans. Healthc. Syst. Eng. 1 131–143. 10.1080/19488300.2011.609524 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Howard S. K., Gaba D. M., Fish K. J., Yang G., Sarnquist F. H. (1992). Anesthesia crisis resource management training: teaching anesthesiologists to handle critical incidents. Aviat. Space Environ. Med. 63 763–770. [PubMed] [Google Scholar]
  44. Huang Y. M., Pliego J. F., Henrichs B., Bowyer M. W., Siddall V. J., McGaghie W. C., et al. (2008). 2007 simulation education summit. Simul. Healthc. 3 186–191. 10.1097/SIH.0b013e3181828156 [DOI] [PubMed] [Google Scholar]
  45. Hull L., Arora S., Kassab E., Kneebone R., Sevdalis N. (2011). Observational teamwork assessment for surgery: content validation and tool refinement. J. Am. Coll. Surg. 212 234–243.e1-5. 10.1016/j.jamcollsurg.2010.11.001 234-243.e235PMID:NOPMID [DOI] [PubMed] [Google Scholar]
  46. Jenks G. F. (1967). The data model concept in statistical mapping. Int. Yearb. Cartogr. 7 186–190. [Google Scholar]
  47. Jiang J., Dai B., Peng D., Zhu C., Liu L., Lu C. (2012). Neural synchronization during face-to-face communication. J. Neurosci. 32 16064–16069. 10.1523/jneurosci.2926-12.2012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Keshmiri S., Sumioka H., Yamazaki R., Ishiguro H. (2017). A non-parametric approach to the overall estimate of cognitive load using NIRS time series. Front. Hum. Neurosci. 11:15. 10.3389/fnhum.2017.00015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Kirilina E., Jelzow A., Heine A., Niessing M., Wabnitz H., Brühl R., et al. (2012). The physiological origin of task-evoked systemic artefacts in functional near infrared spectroscopy. Neuroimage 61 70–81. 10.1016/j.neuroimage.2012.02.074 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Klem G. H., Lüders H. O., Jasper H., Elger C. (1999). The ten-twenty electrode system of the International Federation. Electroencephalogr. Clin. Neurophysiol. Suppl. 52 3–6. [PubMed] [Google Scholar]
  51. Larkin A. C., Cahan M. A., Whalen G., Hatem D., Starr S., Haley H.-L., et al. (2010). Human Emotion and Response in Surgery (HEARS): a simulation-based curriculum for communication skills, systems-based practice, and professionalism in surgical residency training. J. Am. Coll. Surg. 211 285–292. 10.1016/j.jamcollsurg.2010.04.004 [DOI] [PubMed] [Google Scholar]
  52. Ligges U., Short T., Kienzle P., Schnackenberg S., Billinghurst D., Borchers H.-W., et al. (2015). Signal: Signal Processing. Retrieved from Available at: https://CRAN.R-project.org/package=signal [Google Scholar]
  53. Lighthall G. K., Barr J., Howard S. K., Gellar E., Sowb Y., Bertacini E., et al. (2003). Use of a fully simulated intensive care unit environment for critical event management training for internal medicine residents. Crit. Care Med. 31 2437–2443. 10.1097/01.ccm.0000089645.94121.42 [DOI] [PubMed] [Google Scholar]
  54. Lukanov K., Maior H. A., Wilson M. L. (2016). “Using fNIRS in usability testing: understanding the effect of web form layout on mental workload,” in Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems San Jose, CA: 4011–4016. 10.1145/2858036.2858236 [DOI] [Google Scholar]
  55. Luo J., D’Angelo G., Gao F., Ding J., Xiong C. (2015). Bivariate correlation coefficients in family-type clustered studies. Biom. J. 57 1084–1109. 10.1002/bimj.201400131 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Maior H. A., Pike M., Wilson M. L., Sharples S. (2014). Continuous detection of workload overload: an fNIRS approach. Paper Presented at the Contemporary Ergonomics and Human Factors Southampton: 10.1201/b16742-79 [DOI] [Google Scholar]
  57. Mandrick K., Derosiere G., Dray G., Coulon D., Micallef J. P., Perrey S. (2013). Utilizing slope method as an alternative data analysis for functional near-infrared spectroscopy-derived cerebral hemodynamic responses. Int. J. Indust. Ergonom. 43 335–341. 10.1016/j.ergon.2013.05.003 [DOI] [Google Scholar]
  58. Matthews G., Reinerman-Jones L. E., Barber D. J., Abich J. (2015). The psychometrics of mental workload:multiple measures are sensitive but divergent. Hum. Fact. 57 125–143. 10.1177/0018720814539505 [DOI] [PubMed] [Google Scholar]
  59. McGaghie W. C., Issenberg S. B., Petrusa E. R., Scalese R. J. (2010). A critical review of simulation-based medical education research: 2003-2009. Med. Educ. 44 50–63. 10.1111/j.1365-2923.2009.03547.x [DOI] [PubMed] [Google Scholar]
  60. McIvor W. R., Banerjee A., Boulet J. R., Bekhuis T., Tseytlin E., Torsher L., et al. (2017). A taxonomy of delivery and documentation deviations during delivery of high-fidelity simulations. Simul. Healthc. 12 1–8. 10.1097/SIH.0000000000000184 [DOI] [PubMed] [Google Scholar]
  61. McKendrick R., Parasuraman R., Murtza R., Formwalt A., Baccus W., Paczynski M., et al. (2016). Into the wild: neuroergonomic differentiation of hand-held and augmented reality wearable displays during outdoor navigation with functional near infrared spectroscopy. Front. Hum. Neurosci. 10:216. 10.3389/fnhum.2016.00216 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Myrtek M., Deutschmann-Janicke E., Strohmaier H., Zimmermann W., Lawerenz S., BrÜGner G., et al. (1994). Physical, mental, emotional, and subjective workload components in train drivers. Ergonomics 37 1195–1203. 10.1080/00140139408964897 [DOI] [PubMed] [Google Scholar]
  63. Naseer N., Hong K. S. (2015). fNIRS-based brain-computer interfaces: a review. Front. Hum. Neurosci. 9:3. 10.3389/fnhum.2015.00003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Naseer N., Hong M. J., Hong K.-S. (2014). Online binary decision decoding using functional near-infrared spectroscopy for the development of brain–computer interface. Exp. Brain Res. 232 555–564. 10.1007/s00221-013-3764-1 [DOI] [PubMed] [Google Scholar]
  65. Nozawa T., Sasaki Y., Sakaki K., Yokoyama R., Kawashima R. (2016). Interpersonal frontopolar neural synchronization in group communication: an exploration toward fNIRS hyperscanning of natural interactions. Neuroimage 133 484–497. 10.1016/j.neuroimage.2016.03.059 [DOI] [PubMed] [Google Scholar]
  66. Okuda Y., Bryson E. O., DeMaria S., Jacobson L., Quinones J., Shen B., et al. (2009). The utility of simulation in medical education: What is the evidence? Mt. Sinai J. Med. 76 330–343. 10.1002/msj.20127 [DOI] [PubMed] [Google Scholar]
  67. Osaka N., Minamoto T., Yaoi K., Azuma M., Shimada Y. M., Osaka M. (2015). How two brains make one synchronized mind in the inferior frontal cortex: fNIRS-based hyperscanning during cooperative singing. Front. Psychol. 6:1811. 10.3389/fpsyg.2015.01811 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Parasuraman R., Caggiano D. (2005). “Neural and genetic assays of human mental workload,” in Quantifying Human Information Processing eds McBride D., Schmorrow D. (Oxford: Rowman and Littlefeld; ) 123–149. [Google Scholar]
  69. Parasuraman R., Mehta R. (2015). “Neuroergonomic methods for the evaluation of physical and cognitive work,” in Evaluation of Human Work 4th Edn eds Wilson J. R., Sharples S. (Boca Raton, FL: CRC Press; ) 609–640. [Google Scholar]
  70. Peck E. M., Yuksel B. F., Ottley A., Jacob R. J. K., Chang R. (2013). “Using fNIRS brain sensing to evaluate information visualization interfaces,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems Paris: 473–482. 10.1145/2470654.2470723 [DOI] [Google Scholar]
  71. Piper S. K., Krueger A., Koch S. P., Mehnert J., Habermehl C., Steinbrink J., et al. (2014). A wearable multi-channel fNIRS system for brain imaging in freely moving subjects. Neuroimage 85 64–71. 10.1016/j.neuroimage.2013.06.062 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Plichta M., Herrmann M., Baehne C., Ehlis A.-C., Richter M., Pauli P., et al. (2006). Event-related functional near-infrared spectroscopy (fNIRS): are the measurements reliable? Neuroimage 31 116–124. 10.1016/j.neuroimage.2005.12.008 [DOI] [PubMed] [Google Scholar]
  73. Robertson J. M., Dias R. D., Yule S., Smink D. S. (2017). Operating room team training with simulation: a systematic review. J. Laparoendosc. Adv. Surg. Tech. 27 475–480. 10.1089/lap.2017.0043 [DOI] [PubMed] [Google Scholar]
  74. Rosen K. R. (2008). The history of medical simulation. J. Crit. Care 23 157–166. 10.1016/j.jcrc.2007.12.004 [DOI] [PubMed] [Google Scholar]
  75. Rutherford D. N., D’Angelo A.-L. D., Law K. E., Pugh C. M. (2015). Advanced engineering technology for measuring performance. Surg. Clin. 95 813–826. 10.1016/j.suc.2015.04.005 [DOI] [PubMed] [Google Scholar]
  76. Salas E., Maurino D. (2010). Human Factors in Aviation 2nd Edn. Amsterdam: Elsevier. [Google Scholar]
  77. Scholkmann F., Holper L., Wolf U., Wolf M. (2013). A new methodical approach in neuroscience: assessing inter-personal brain coupling using functional near-infrared imaging (fNIRI) hyperscanning. Front. Hum. Neurosci. 7:813. 10.3389/fnhum.2013.00813 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Scholkmann F., Kleiser S., Metz A. J., Zimmermann R., Pavia J. M., Wolf U., et al. (2014). A review on continuous wave functional near-infrared spectroscopy and imaging instrumentation and methodology. Neuroimage 85 6–27. 10.1016/j.neuroimage.2013.05.004 [DOI] [PubMed] [Google Scholar]
  79. Skjerve A. B., Bye A. (2011). Simulator-Based Human Factors Studies Across 25 Years. London: Springer. [Google Scholar]
  80. Stevens R., Galloway T., Willemson-Dunlap A. (2016). “Intermediate neurodynamic representations: a pathway towards quantitative measurements of teamwork?,” in Proceedings of the Human Factors and Ergonomics Society Annual Meeting Los Angeles, CA: 1996–2000. 10.1177/1541931213601454 [DOI] [Google Scholar]
  81. Tachtsidis I., Scholkmann F. (2016). False positives and false negatives in functional near-infrared spectroscopy: issues, challenges, and the way forward. Neurophotonics 3 030401–030401. 10.1117/1.NPh.3.3.030401 [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Toppi J., Borghini G., Petti M., He E. J., De Giusti V., He B., et al. (2016). Investigating cooperative behavior in ecological settings: an EEG hyperscanning study. PLoS One 11:e0154236. 10.1371/journal.pone.0154236 [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Torrence C., Compo G. P. (1998). A practical guide to wavelet analysis. Bull. Am. Meteorol. Soc. 79 61–78. [DOI] [Google Scholar]
  84. Vincenzi D. A., Wise J. A., Mouloua M., Hancock P. A. (2009). Human Factors in Simulation and Training. Boca Raton, FL: CRC Press. [Google Scholar]
  85. Watkins S. C., Roberts D. A., Boulet J. R., McEvoy M. D., Weinger M. B. (2017). Evaluation of a simpler tool to assess non-technical skills during simulated critical events. Simul. Healthc. 12 69–75. 10.1097/SIH.0000000000000199 [DOI] [PubMed] [Google Scholar]
  86. Watson D., Clark L. A., Tellegen A. (1988). Development and validation of brief measures of positive and negative affect: the PANAS scales. J. Pers. Soc. Psychol. 54:1063 10.1037/0022-3514.54.6.1063 [DOI] [PubMed] [Google Scholar]
  87. Weinger M. B., Banerjee A., Burden A., Boulet J. R., Cooper J., Steadman R., et al. (2017). Simulation-based assessment of the management of critical events by board-certified anesthesiologists. Anesthesiology 127 475–489. 10.1097/ALN.0000000000001739 [DOI] [PubMed] [Google Scholar]
  88. Weinger M. B., Herndon O. W., Gaba D. M. (1997). The effect of electronic record keeping and transesophageal echocardiography on task distribution, workload, and vigilance during cardiac anesthesia. J. Am. Soc. Anesthesiol. 87 144–155. 10.1097/00000542-199707000-00019 [DOI] [PubMed] [Google Scholar]
  89. Weinger M. B., Herndon O. W., Zornow M. H., Paulus M. P., Gaba D. M., Dallen L. T. (1994). An objective methodology for task analysis and workload assessment in anesthesia providers. Anesthesiology 80 77–92. 10.1097/00000542-199401000-00015 [DOI] [PubMed] [Google Scholar]
  90. Weinger M. B., Reddy S. B., Slagle J. M. (2004). Multiple measures of anesthesia workload during teaching and nonteaching cases. Anesth. Analg. 98 1419–1425. 10.1213/01.ANE.0000106838.66901.D2 [DOI] [PubMed] [Google Scholar]
  91. Weinger M. B., Slagle J. M. (2002). Human factors research in anesthesia patient safety: techniques to elucidate factors affecting clinical task performance and decision-making. J. Am. Med. Inform. Assoc. 9 S58–S63. 10.1197/jamia.M1229 [DOI] [Google Scholar]
  92. Williams R. G., Klamen D. A., McGaghie W. C. (2003). SPECIAL ARTICLE: cognitive, social and environmental sources of bias in clinical performance ratings. Teach. Learn. Med. 15 270–292. 10.1207/S15328015TLM1504_11 [DOI] [PubMed] [Google Scholar]
  93. Wilson M. R., Poolton J. M., Malhotra N., Ngo K., Bright E., Masters R. S. W. (2011). Development and validation of a surgical workload measure: the surgery task load index (SURG-TLX). World J. Surg. 35 1961–1969. 10.1007/s00268-011-1141-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Xu J., Anders S., Pruttianan A., France D. J., Lau N., Adams J. A., et al. (2017). “Human performance measures for the evaluation of nuclear power plant control room interfaces: a systematic review,” in Proceedings of The 10th International Topical Meeting on Nuclear Plant Instrumentation, Control and Human-Machine Interface Technology (NPIC & HMIT 2017) (Switzerland: Springer; ). [Google Scholar]
  95. Yoshino K., Oka N., Yamamoto K., Takahashi H., Kato T. (2013). Functional brain imaging using near-infrared spectroscopy during actual driving on an expressway. Front. Hum. Neurosci. 7:882. 10.3389/fnhum.2013.00882 [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Young M. S., Brookhuis K. A., Wickens C. D., Hancock P. A. (2015). State of science: mental workload in ergonomics. Ergonomics 58 1–17. 10.1080/00140139.2014.956151 [DOI] [PubMed] [Google Scholar]
  97. Zhang Y., Meng T., Hou Y., Pan Y., Hu Y. (2018). Interpersonal brain synchronization associated with working alliance during psychological counseling. Psychiatry Res. Neuroimag. 282 103–109. 10.1016/j.pscychresns.2018.09.007 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The datasets generated for this study are available on request to the corresponding author.


Articles from Frontiers in Human Neuroscience are provided here courtesy of Frontiers Media SA

RESOURCES