Skip to main content
Seminars in Hearing logoLink to Seminars in Hearing
. 2023 Apr 6;44(2):188–210. doi: 10.1055/s-0043-1766105

Brightening the Study of Listening Effort with Functional Near-Infrared Spectroscopy: A Scoping Review

Hannah E Shatzer 1,, Frank A Russo 1
PMCID: PMC10147513  PMID: 37122884

Abstract

Listening effort is a long-standing area of interest in auditory cognitive neuroscience. Prior research has used multiple techniques to shed light on the neurophysiological mechanisms underlying listening during challenging conditions. Functional near-infrared spectroscopy (fNIRS) is growing in popularity as a tool for cognitive neuroscience research, and its recent advances offer many potential advantages over other neuroimaging modalities for research related to listening effort. This review introduces the basic science of fNIRS and its uses for auditory cognitive neuroscience. We also discuss its application in recently published studies on listening effort and consider future opportunities for studying effortful listening with fNIRS. After reading this article, the learner will know how fNIRS works and summarize its uses for listening effort research. The learner will also be able to apply this knowledge toward generation of future research in this area.

Keywords: fNIRS, listening effort, auditory cognitive neuroscience


Speech perception involves the extraction of meaning from a continuous stream of sound. This process appears deceptively simple until it starts to falter, and it varies in difficulty with respect to factors that are both external and internal to the listener. 1 Real-world communication often involves listening to speech in challenging conditions. These challenges may arise from environmental sources (e.g., background noise, competing talkers), degraded or difficult speech (e.g., conversation via phone, a foreign-accented talker, speech in the listener's nonnative language), or internal factors such as hearing loss or disordered auditory processing. Decades of research have been directed at understanding how listeners perceive speech under a variety of challenging conditions, with common techniques involving behavioral and/or physiological measures of listening effort. Behavioral approaches include self-report and experimental measures, such as dual-task procedures and memory paradigms. Physiological approaches can be divided into measures of the autonomic nervous system (ANS), such as skin conductance or pupil dilation, and neurophysiological measures such as those which can be obtained with functional magnetic resonance imaging (fMRI), electroencephalography (EEG), or magnetoencephalography (MEG; see Table 1 for abbreviations used throughout this article). 2 Neurophysiological measures are likely to be the most reliable method of assessing listening effort, as they directly index brain activity associated with increased cognitive effort. A systematic review by Ohlenforst et al 3 concluded that only studies with brain-based measures of listening effort were able to demonstrate greater listening effort in individuals with hearing loss compared to those with normal hearing, suggesting that behavioral and autonomic measures are unable to reliably identify differences in listening effort based on hearing status. However, the most commonly used neuroimaging techniques (e.g., fMRI, MEG/EEG) each pose challenges for obtaining measures of listening effort from hard-of-hearing or deaf individuals who use hearing instruments, such as hearing aids or cochlear implants. In the case of fMRI and MEG, the metallic components in hearing instruments amount to an exclusion criterion in the majority of cases. In addition, motion artifacts are highly problematic in fMRI research, which leads to paradigms that necessitate severe restriction of movement. In the case of EEG, hearing instruments generate an electromagnetic field which leads to artifacts that can be difficult to fully remove.

Table 1. Important abbreviations.

ANS Autonomic nervous system
DLPFC Dorsolateral prefrontal cortex
EEG Electroencephalography
fMRI Functional magnetic resonance imaging
fNIRS Functional near-infrared spectroscopy
HbO Oxygenated hemoglobin
HbR Deoxygenated hemoglobin
HbT Total hemoglobin
HD-DOT High-density diffuse optical tomography
IFG Inferior frontal gyrus
IPS Intraparietal sulcus
IPL Inferior parietal lobule
L1 First (native) language
L2 Second language
MEG Magnetoencephalography
MFG Middle frontal gyrus
MTG Middle temporal gyrus
PET Positron emission tomography
PFC Prefrontal cortex
SIN Speech-in-noise
SNR Signal-to-noise ratio
STG Superior temporal gyrus
VAE Virtual acoustic environment

Recently, functional near-infrared spectroscopy (fNIRS) has increased in popularity for use in cognitive neuroscience research, and it has great potential for overcoming some of the challenges faced with other brain-based measures of listening effort. fNIRS is a noninvasive technique which directs near-infrared (NIR) light into the brain to assess changes in oxygenated and deoxygenated blood flow to cortical brain tissue over time. It has a higher temporal resolution than fMRI, and higher spatial resolution than EEG. It is also compatible with hearing instruments, robust to motion artifacts, and is able to image several areas of cortex that have been implicated in listening effort. All of these features suggest that fNIRS has great potential for studying the neural underpinnings of listening effort in diverse populations and in a variety of ecological contexts. The goals of the current article are to (1) explain why fNIRS is poised to play an essential role in listening effort research, (2) survey the work that has been completed thus far using fNIRS to study listening effort, and (3) identify areas for future research that would benefit from fNIRS within the body of work on listening effort.

Previous Work in Listening Effort

A consensus paper by Pichora-Fuller and colleagues 4 established a framework that defined listening effort as “the deliberate allocation of mental resources to overcome obstacles in goal pursuit when carrying out a [listening] task.” By this definition, effortful listening involves three primary elements: task demands, hearing difficulties, and the listener's level of motivation to understand what they are hearing. Many different types of tasks and measures of listening effort have been used and compared in previous research, which can be broadly categorized into behavioral and physiological measures. Behavioral tasks include both subjective and experimental measures. In subjective measures, participants self-report the level of effort they feel they are expending to complete a challenging listening task. Commonly-used self-report measures include standardized and validated scales like the NASA-TLX, 5 6 or simpler indicators such as rating effort on a multi-point Likert scale. 7 8 Experimental paradigms largely quantify listening effort by accuracy and/or reaction time: As listening challenges increase, accuracy typically decreases, and reaction time increases. Common experimental methods to assess listening effort include dual-task paradigms, in which participants complete a speech recognition task simultaneously with a secondary task involving other cognitive processes (e.g., visual perception or memory). 9 10 11 An underlying assumption in dual-task procedures is that domain-general resources that are used to support listening effort should be less available to support secondary tasks that are unrelated to speech understanding. Hence, we should be able to assess listening effort by way of performance on the secondary task. Memory paradigms are also often used on their own to assess listening effort, with the reasoning that as listening challenges increase, fewer domain-general resources are available for memory processing, therefore leading to lower accuracy and slower reaction times in the memory task. Listening span tasks, 7 12 the cognitive spare capacity test, 13 14 and running memory tasks 15 are some common memory-based paradigms for behavioral assessments of listening effort.

Behavioral tasks can often be used simultaneously or in addition to physiological measures of listening effort, which involve recording physiological responses from the ANS or the brain. The sympathetic nervous system branch of the ANS responds to listening challenges with a host of physiological changes. Pupillometry has become a standard methodology in the assessment of listening effort, as changes in pupil dilation tend to predict changes in cognitive effort across several tasks and sensory domains. 16 17 18 Pupil dilation increases with reductions in the signal-to-noise ratio (SNR) of speech in noise, 19 greater degradation of speech via noise vocoding, 20 and with increasing syntactic complexity of speech, 21 suggesting that it is sensitive to a variety of listening challenges. Another ANS-based measure of listening effort is galvanic skin response, as increased skin conductance has been found with increased difficulty of listening conditions. 6 22 Finally, heart-related measures have been found to correlate with increased listening challenges, including heart rate variability, 6 23 preejection period, 24 and blood volume pulse amplitude. 25 Elevated cortisol levels have also been linked to effortful listening—however, the endocrine response appears to be physiologically distinct from ANS measures, reflecting changes in stress that persist over time (i.e., chronic) due to hearing challenges. 26 27 Consistent with this view, momentary elevations in cortisol have not been found in relationship to situational factors that affect listening challenges.

Brain-based measures of listening effort have typically involved use of fMRI or MEG/EEG. Each neuroimaging technique has highlighted neural signatures related to effortful listening. In the fMRI literature, processing of easily intelligible speech generally involves domain-specific regions including bilateral temporal regions including superior temporal gyrus (STG) and middle temporal gyrus (MTG), as well as inferior frontal gyrus (IFG). 28 29 30 As listening challenge increases, domain-general networks are recruited to process the target signal (for a more detailed review, see Peelle, 2018). 31 The cingulo-opercular (or “salience”) network, including the anterior insula, cingulate cortex, and thalamus, may become active as part of a general compensatory response leading to the intrinsic maintenance of tonic alertness. 32 33 34 In some circumstances, a listening challenge may preferentially engage the frontoparietal attention (or “central executive”) network which directs attention in favor of relevant information while suppressing irrelevant information. 35 36 This network runs dorsally from frontal to parietal cortex, including the dorsolateral prefrontal cortex (DLPFC), intraparietal sulcus (IPS), and to the inferior parietal lobule (IPL). Listening effort work with EEG and MEG has also identified oscillatory signatures and auditory-evoked potentials that correlate with listening challenge. Notably, increases in the power of alpha oscillations have been linked to the acoustic degradation of a signal, and therefore may reflect increased engagement of cognitive resources required to suppress noise such as selective attention or working memory. 37 38 39 This alpha-modulated noise suppression has been localized to both frontal and parietal sources, 40 41 which is consistent with activations observed in fMRI studies of effortful listening. Slower cortical oscillations in the delta and theta range have also been observed to phase-lock with the speech envelope. 42 This phase-locking increases when speech is degraded or presented in noise, suggesting modulation of the cortical response to challenging speech by top-down cognitive functions. 43 44 45 In terms of event-related potentials, a larger and slower auditory-evoked N1 has been linked to more challenging listening conditions, 46 and the N400 amplitude has been linked to linguistic challenge when listening to nonnative accented speech. 47

Some of these neuroimaging techniques have also been used successfully to study listening effort in individuals with hearing loss and hearing instruments (see Kuchinsky & Vaden, 2020 for more detail). 48 However, as noted earlier, there are major challenges with fMRI and MEG/EEG when trying to study listening effort in populations with hearing instruments. Hearing aids and cochlear implants are incompatible for use in an fMRI scanner due to the metallic components in these devices, which also leave electromagnetic artifacts in recording from most MEG systems. While hearing aids and cochlear implants can be worn during EEG recording, they also inevitably result in electrical artifacts. Some data cleaning methods (e.g., independent components analysis) have been developed to reduce these artifacts, though they cannot remove artifacts completely without introducing some distortion of the signal. It is also possible to use insert earphones that are calibrated against probe microphone measurements taken in situ with hearing aids. 49 50 This approach avoids electrical artifacts, but it does not generalize to dynamic contexts outside the lab where sound cannot be fully controlled, and even in the case of a lab study, the realism of such simulations can be called into question because of possible effects related to differences in occlusion between real-life listening with hearing aids and laboratory simulations. These difficulties of standard neuroimaging systems complicate the ability to study listening effort in hard-of-hearing populations who use hearing instruments, and to determine how use of a hearing instrument may affect listening effort.

Introduction to Functional Near-Infrared Spectroscopy

fNIRS has a relatively recent history for use in cognitive neuroscience research. It has many benefits for auditory research and great potential for future work understanding listening effort, especially with regard to testing in ecological contexts and with those with hearing loss who use hearing instruments. fNIRS is a form of optical imaging, as it measures the hemodynamic response function of the brain with the use of NIR light (see Pinti et al, 2020, for complete review). 51 Fig. 1 illustrates the mechanics of the fNIRS signal. NIR light generated from low-power lasers and/or LEDs is shone into the head from a light source placed on the scalp. This light scatters through the scalp, skull, cerebrospinal fluid, and other layers of tissue. As the body is primarily composed of water, it minimally absorbs the NIR light, and the light is therefore able to continue traveling through the tissue. 52 Comparatively, hemoglobin is the most absorptive chromophore within the NIR optical window, so it absorbs more of the NIR light relative to surrounding tissue and cerebrospinal fluid. Two wavelengths of NIR light are typically used to measure absorption for oxygenated and deoxygenated hemoglobin. NIR light at 700 nM is “redder” and tends to be better absorbed by deoxygenated hemoglobin (HbR), while NIR light at 900 nM is less visible and is better absorbed by oxygenated hemoglobin (HbO). 53 A light detector (also called a “target”) is placed a specific distance away from the light source on the scalp, and it measures the optical density of backscattered light at both wavelengths that has not been absorbed.

Figure 1.

Figure 1

Illustration of the mechanics of the functional near-infrared spectroscopy signal. Near-infrared photons are emitted from the light source and travel through the layers of the head to the detectors. The channel indicating the measurement point on the cortex is indicated with C . The depth of the channel is proportional to the distance between the source and detector on the scalp ( C indicates a standard-depth channel; SSC indicates a shallower short-separation channel).

A source–detector pair creates a channel, which is located at the midpoint between the source and the detector. The area of cortex being imaged can be found perpendicular to the scalp beneath the midpoint at a depth corresponding to half of the distance between the source and detector pair. When the source and detector are farther apart, the detector is able to measure backscattered light at a greater depth than when they are closer together. However, if the distance between the source and detector pair is too close, light will reach the detector directly without traveling through the scalp. If the distance is too far, the detector is unable to detect any meaningful amount of light. Research suggests that to reach the cortex, the optimal distance between a source and a detector is approximately 3 to 5 cm in adults and 2 to 3 cm in infants. 54 Some fNIRS systems also include “short separation channels,” which are source–detector pairs placed at a closer distance (approximately 1–1.5 cm) so that the channel measures at a shallower depth than other source–detector pairs. These short channels enable tracing of physiological processes that are not indicative of neural activity, such as respiration or changes in noncortical blood flow that result from ANS activity. The activity captured by short channels can then be regressed out of the fNIRS recording during analysis as a means of removing physiological noise from standard channels, thus providing a cleaner picture of the hemodynamic activity related to local cortical circuits. During analysis, optical density measurements captured by the detectors over time are converted into hemoglobin concentrations that correspond to the two different wavelengths of NIR light emitted by the sources. HbO and HbR concentrations may be analyzed separately, or HbR and HbO may be added together to analyze the total changes in hemoglobin concentrations (HbT). The majority of studies focus on analyses related to hypothesis testing on HbO, while reporting descriptive statistics on HbO as well as HbR. Some studies have used the difference between HbO and HbR as a measure for analysis due to its relative sensitivity to both the increase in HbO and the decrease in HbR. 55 56

These analyses of oxygenation and deoxygenation often stand on their own, but the same measures may also be subjected to connectivity analysis in order to better understand network activity. These connectivity analyses can be conducted with measures obtained from different regions of interest while the participant is in a resting state, often for the purposes of understanding individual differences or neuroplasticity. 57 Alternatively, the analyses may be performed on measures collected while the participant is engaged in a task, often for the purposes of understanding differences in circuitry underlying different tasks. 58 The assessment of connectivity can be as simple as correlating two time series from different regions of interest, but may also take on more complex operations in the form of wavelet coherence analysis. 59

While the use of optical imaging for cognitive neuroscience is fairly recent, fNIRS has been in development for neuroimaging purposes since the early 1990s. At first, single-channel devices were developed and could detect hemodynamic changes in the brain in response to basic functional tasks. 60 61 62 63 Eventually, several single-channel devices were combined and used simultaneously to measure from multiple regions at once 64 before the development of multichannel systems (see Ferrari & Quaresima, 2012, for more complete history). 65 66 67

Most commercially available systems for optical imaging of the brain are continuous wave (CW) instruments, which continuously emit NIR light at 2 to 3 wavelengths for measurement of HbO and HbR (see Scholkmann et al, 2014, for detailed review). 68 However, time-domain and frequency-domain devices can also be used to collect optical data. Frequency-domain devices produce intensity-modulated NIR light, while time-domain devices emit rapid pulses of NIR light. Current fNIRS systems are either fiber-based or wearable devices and generally involve attaching the optodes to a cap, band, or other headgear. Fiber-based systems use optical fibers containing low-power lasers, so the participant remains stationary during recording and mobility is limited. However, optical fibers have higher power than LED light and can image the cortex with greater separation between the source and detector pairs (up to 6 cm), thereby allowing deeper penetration of the cortex. Wearable systems do not use fibers, and instead employ LEDs to emit the NIR light into the scalp. Because wearable systems are fiber-less, they are more portable, and can be used to record from a participant in motion 69 and even outside the lab. 70 Both types of systems have varying degrees of optode density and can generally accommodate up to approximately 64 channels. Recently, high-density diffuse optical tomography (HD-DOT) systems that are both fiber-based and LED-based have been developed to allow measurement with NIR light from a larger number of optodes (more than 90 sources and detectors). These systems are designed so that sources and detectors are placed at several different separation distances, which allows for recording from spatially overlapped channels and brings the spatial resolution of fNIRS approximately equal to that of 3-T fMRI. 71 72

Several features of fNIRS make it a particularly good choice for studies in auditory cognitive neuroscience (see Table 2 ). The system itself is generally quiet, so it is ideal for auditory tasks where spatial imaging is desired. The fMRI scanner produces noise during imaging, though some techniques have been developed to facilitate presentation of auditory stimuli in the scanner. 73 fNIRS systems vary in the amount of noise they emit: Most fNIRS systems are either silent or generate low levels of ambient noise that are far less disruptive than fMRI scanner noise. One example in which imaging noise may have affected results is in the study of auditory working memory. Rovetti et al (2021) 74 found that there was no difference in DLPFC activity for auditory versus visual n-back tasks while recording with a silent fNIRS system, which was consistent with findings from a previous positron emission tomography (PET) study 75 but contrary to prior fMRI results suggesting increased DLPFC engagement for auditory relative to visual n-back tasks. 76 77 One potential explanation for this pattern of findings across studies is that the fMRI scanner noise contributed to additional listening effort, thus increasing engagement of DLPFC for the auditory task relative to the visual task. PET and fNIRS are both very quiet by comparison, and therefore would not suffer from the same confound. As previously mentioned, fNIRS is also compatible with hearing instruments such as hearing aids and cochlear implants. There is no interference from the metallic components of the devices (which prohibits these devices from being used in an MRI scanner), and as fNIRS measures hemodynamic activity, the signal is not affected by the electrical components of hearing devices.

Table 2. Comparison of popular neuroimaging methods.

Feature fMRI EEG fNIRS
Acoustic noise level during imaging High Silent Silent-low
Compatible with hearing instruments No Yes a Yes
Spatial resolution Good Poor Fair
Temporal resolution Poor Good Fair
Portability No Few systems Some systems
Robust to motion artifacts No No Yes
a

While hearing aids and cochlear implants can be worn during EEG recording, they produce an electromagnetic artifact.

The separate measurement of HbO and HbR in fNIRS can be viewed as another advantage over fMRI. HbR measured with fNIRS correlates most strongly with the blood oxygen level–dependent signal (BOLD) measured with fMRI. 78 For example, in a concurrent fMRI-fNIRS study, Huppert et al (2006) 79 found correlation coefficients between 1/BOLD and HbT, HbO, and HbR to be 0.53, 0.71, and 0.98, respectively. To explain the discrepancy in correlation strengths, we can consider biophysical models of hemodynamics. 80 When a population of neurons becomes active, HbO is expected to rush into the area. As a consequence of this increase in HbO, HbT will temporarily increase and HbR will be washed out. Thus, over the course of an entire block of trials, HbO and HbR will be inversely correlated but over the short term (e.g., the first 10 seconds following the onset of a stimulus), HbO and HbR may differ to some extent. The ability to measure HbO directly, and to assess HbO and HbR independently, affords fNIRS a more nuanced temporal view of brain hemodynamics than fMRI. This nuanced temporal view is further supported by the relatively high sampling rate of fNIRS (10–100 Hz) compared to fMRI (typically 0.5 Hz), which facilitates use of event-related, block, and mixed designs. 81

fNIRS has several additional advantages over other imaging systems that may make it especially suitable for listening effort research. It has higher spatial resolution than EEG (10–20 mm for fNIRS compared to 5–9 cm for EEG), so it is a comparably better choice for localizing cortical activity. 82 The locations of source and detectors on the scalp can also be digitized and converted to coordinates in MNI space and/or coregistered to fMRI brain atlases for better visualizations of localized activity. Certain fNIRS systems can also interface with other imaging modalities to provide a more thorough, multimodal assessment of neural activity. Systems have been developed for recording from concurrent EEG-fNIRS, 83 84 fMRI-fNIRS, 85 86 87 and MEG-fNIRS. 88 The cost of the average fNIRS system is comparable to EEG, and therefore relatively low when compared to fMRI, MEG, or PET systems. 51 fNIRS systems also do not have specific environmental requirements, such as magnetically or electrically shielded rooms. They can be used for longer imaging sessions up to approximately 1 hour and can be easily implemented for repeated measures designs. 82 An additional benefit for fNIRS systems is that they are relatively robust to motion artifacts, so they can be used in paradigms that involve substantial participant movement. 70 Wearable fNIRS systems are portable and can be used in more naturalistic settings outside the lab. Overall, the many benefits of fNIRS systems indicate great potential for brain-based research.

fNIRS systems are not without their limitations. As noted, the penetration depth of fNIRS channels is only about 1.5 to 3.5 cm from the scalp in adults, depending on the distance between source–detector pairs. This means that fNIRS cannot measure activity from deeper cortical and subcortical structures. This poses a challenge for imaging the auditory cortex, as it has a transverse orientation within the temporal lobe, and the NIR light cannot reach the entire auditory cortex. 89 Nevertheless, attempts to image parts of the auditory cortex have been successful, including the primary auditory cortex, secondary auditory cortex, and Wernicke's area. Fig. 2 depicts the locations of these regions of interest, in addition to Broca's area and core nodes of the frontoparietal attention network.

Figure 2.

Figure 2

Model representation of cortical regions of interest for listening effort, including left temporal, frontal, and parietal sources.

Additionally, NIR light transmission is attenuated by darker hair and skin pigmentation, which has higher absorption in the NIR optical window. 90 91 Neuroimaging applications of fNIRS therefore risk limited racial and ethnic diversity because a reliable signal may not be attainable from individuals with darker hair and skin tones. However, the extent of this problem appears to be negligible in recent fNIRS systems and may be mitigated with the inclusion of short channels. 92 93

Finally, while some laser- and LED-based fNIRS systems are silent, others generate noticeable levels of noise that could be distracting or serve as confound in auditory experiments. The noise level tends to correlate with the size of the system; so, wearable/portable systems are more likely to be silent while fiber-based systems are likely to generate more low-level noise from the cooling fans and other components inside the amplifier box. If using a fiber-based system for research involving auditory stimuli (e.g., listening effort studies), it is recommended to place the fNIRS amplifier outside the sound-attenuated booth to avoid this confound.

fNIRS and Auditory Cognitive Neuroscience

The earliest study to successfully image auditory cortex with fNIRS was completed by Hoshi and Tamura, 60 who combined five single-channel devices placed over different areas of the cortex and found systematic increases in activation to sound in the temporal lobe. Ohnishi et al 94 followed up on this work to more specifically measure from auditory cortex. They first used MEG to localize the auditory cortex response to a pure tone, then placed an fNIRS channel over this region and confirmed an increase in both HbO and HbR in response to the tone stimulus. Further work extended use of fNIRS to capture the hemodynamic response to sound in auditory cortex in children and infants. 95 96 97 Differential activations to various sound categories in auditory cortex have also been confirmed using fNIRS, 98 99 indicating that while fNIRS has lower spatial resolution than fMRI, it can be quite sensitive to changes in the type of sound stimulus being used. To our knowledge, no study to date has confirmed whether fNIRS has the necessary sensitivity for tonotopic mapping of the auditory cortex.

Once early research confirmed that fNIRS can capture the hemodynamic response to acoustic stimuli, it became a more popular tool in several areas of research within auditory cognitive neuroscience. One of its most prominent uses is to study auditory processing in deaf and hard-of-hearing individuals, including those who use hearing instruments (see Oghalai, 2017; Bortfeld, 2019, for detailed review). 100 101 As previously mentioned, fNIRS is uniquely compatible with hearing aids and cochlear implants, and is therefore an ideal choice for noninvasive neuroimaging in individuals who use hearing devices. Furthermore, fNIRS is simple to implement over multiple testing sessions; so, it is ideal for longitudinal designs and may be an important tool to assess the efficacy of hearing devices over time (e.g., before and after cochlear implantation). 102 Sevy et al 103 were the first group to identify a reliable response to speech in auditory cortex for cochlear implant users. Bisconti et al 104 found overall similar activation patterns in auditory and nonauditory regions between cochlear implant users and normal-hearing adults during passive listening to speech, suggesting that hearing loss in adulthood does not significantly alter the neural organization of speech perception. Cortical activation to speech in cochlear implant users has also been associated with speech perception accuracy, which links neural variability to behavioral variability in cochlear implant outcomes. 105 106 More recently, HD-DOT has been explored as a means to increase spatial resolution of imaging in cochlear implant users. 107 In contrast, relatively few studies have investigated auditory perception in hearing aid users with fNIRS. Two studies to be discussed later tested adult and children who use hearing aids in listening effort tasks, but other fNIRS work in hearing aid users is currently limited. 108 109

Some studies have investigated cross-modal plasticity in cochlear implant users by measuring the cortical response to auditory and visual stimuli, as prior fNIRS evidence suggests that neural reorganization occurs during deafness to maximize processing of visual and other sensory stimuli in absence of auditory stimulation. 110 111 Chen et al completed concurrent EEG-fNIRS recording during presentation of unimodal and multimodal stimuli, finding greater visual adaptation and lower activation of visual cortex in cochlear implant users relative to normal-hearing adults. 112 113 Anderson et al 114 measured oxygenation of superior temporal cortex in cochlear implant recipients in response to visual speech stimuli both preimplantation and 6 months postimplantation, finding that an increased response to visual speech in auditory cortex postimplantation was associated with higher speech perception scores with a cochlear implant, therefore suggesting that increased cross-modal plasticity in auditory cortex has adaptive benefits for speech perception postimplantation. Follow-up work by the same group showed that fNIRS can be used preimplantation to predict cochlear implant outcomes in deaf adults, as stronger auditory cortex activation to visual speech preimplantation was associated with poorer auditory speech perception 6 months postimplantation. 115 They also investigated cross-modal activations to speech in pediatric cochlear implant recipients, finding that cochlear-implanted children had greater auditory cortex responses to visual speech but similar responses to auditory speech compared to normal-hearing children. 116

In normal-hearing listeners, fNIRS has been used successfully with a variety of speech stimuli to replicate and extend findings from other neuroimaging methods. Differential cortical activations were identified for normal speech versus speech that was distorted to simulate the signal transmitted via a cochlear implant, confirming that fNIRS can detect differences in the cortical response to speech based on intelligibility. 117 In infants, differential activations were found for listening to speech in the infants' native versus nonnative language at 1-day-old 118 and for native-accented versus foreign-accented speech within the first few days of birth. 119 Audiovisual speech stimuli have also been shown to increase activity in temporal regions relative to auditory-only stimuli in adults 120 as well as infants. 121 Shader et al 122 identified differential activation in specific regions of interest that responded to auditory speech near the auditory core, and visual speech near middle occipital gyrus.

fNIRS has also been used to understand the neural processing of complex sound beyond the types of stimuli that are used in typical speech perception studies. For example, differential cortical activations have been observed for music versus noise, 123 various genres of music, 124 familiarity of music, 125 and for whether or not music is playing during memory encoding. 126 127 128 Differential responses have also been observed across categories of vocal emotion in adults 129 130 131 as well as infants. 132 Given the portable nature of many systems, fNIRS also has great potential for studying dynamic communication in real-world environments via a process called hyperscanning. 133 134 135

fNIRS and Listening Effort

Listening effort is a burgeoning area of hearing research with important implications for real-world scenarios that involve auditory–cognitive interactions. The neural substrates of these interactions can be effectively studied using fNIRS. Researchers are able to take advantage of the benefits of optical imaging for auditory research, as well as the flexibility and ease of use of fNIRS systems across diverse populations, including people living with hearing loss who use hearing instruments. Because of its versatility, the use of fNIRS to study listening effort spans several disciplines, including psychology, neuroscience, audiology, linguistics, and cognitive science. An online search was conducted in February 2022 to locate fNIRS studies of listening effort using Google Scholar, as the diversity of disciplines potentially involved in this work is unlikely to be fully captured using a single-research database. The search included the following search terms: (listening effort) and (fNIRS or functional near infrared spectroscopy). After limiting the search results to peer-reviewed original research articles, 61 studies were identified. Next, the article titles and abstracts were reviewed to determine whether each article used fNIRS to study listening effort. On the basis of this assessment, 42 of the 61 articles were eliminated. The authors then read the method and results sections of the remaining 19 articles to further determine eligibility, and eliminated all articles that did not have listening effort as the primary focus of the research. A replication of the same search process conducted using PubMed and PsycInfo yielded no new studies that were not already indexed by Google Scholar.

Table 3 lists the 11 currently published original research articles as of February 2022 that have used fNIRS to study listening effort. These studies have all been published within the past 5 years, which is an indication of the relatively recent adoption of fNIRS in listening effort research. A variety of listening challenges and stimulus manipulations have been used, allowing for comparison across studies of different sources of listening challenge and the accompanying effects on neural processing.

Table 3. Currently published fNIRS studies of listening effort.

Citation Subject population fNIRS system Cortical ROIs fNIRS signal analyzed Listening challenge Stimulus type
Bell, Peng, Pausch et al (2020) Children with normal hearing and hearing aids Hitachi ETG-4000 system with 44 channels Bilateral STG and IFG HbT Acoustic Speech-in-noise (virtual acoustic environment)
Defenderfer, Forbes, Wijeakumar, Hedrick, Plyler, & Buss (2021) Young adults with normal hearing Techen CW7 system with 13 channels, 1 short channel Left frontal and temporal cortices HbO; HbR Acoustic Speech-in-noise; degraded speech
Defenderfer, Kerr-German, Hedrick, & Buss (2017) Young adults with normal hearing System not stated; 12 channels Bilateral auditory cortex, STG HbO; HbR Acoustic Speech-in-noise, degraded speech
Lawrence, Wiggins, Anderson, Davies-Thompson, & Hartley (2018) Young adults with normal hearing Hitachi ETG-4000 system with 44 channels Bilateral superior temporal cortex, IFG, IPL HbO; HbR Acoustic Degraded speech
Lawrence, Wiggins, Hodgson, & Hartley (2021) Children with normal hearing Hitachi ETG-4000 system with 26 channels Bilateral superior temporal cortex, IFG, rolandic operculum, posterior temporal cortex HbO Acoustic Degraded speech
Mushtaq, Wiggins, Kitterick, Anderson, & Hartley (2021) Children with normal hearing Hitachi ETG-4000 system with 44 channels Bilateral superior temporal cortex, IFG, rolandic operculum, posterior temporal cortex HbO; HbR Acoustic Degraded speech
Rovetti, Goy, Pichora-Fuller, & Russo (2019) Older adults with normal hearing and hearing aids fNIR Imager 1100 with 16 channels Bilateral prefrontal cortex HbO–HbR Acoustic Visual and auditory verbal n-back tasks, completed with/without hearing aids
Rovetti, Goy, Zara, & Russo (2021) Young adults with normal hearing fNIR Imager 1100 with 16 channels Bilateral prefrontal cortex HbO–HbR Acoustic; linguistic Speech-in-noise, sentences with high/low context
Rowland, Hartley, & Wiggins (2018) Young adults with normal hearing Hitachi ETG-4000 system with 52 channels bilateral prefrontal cortex, superior temporal cortex HbO; HbR Acoustic speech-in-noise (acoustic scenes)
White & Langdon (2021) Young adults with normal hearing NIRx NIRScout system with 50 channels Bilateral IFG, SMG, primary somatosensory cortex, MFG HbR Acoustic; linguistic Sentences varying in semantic plausibility, syntactic complexity, and speech rate; degraded speech
Wijayasiri, Hartley, & Wiggins (2017) Young adults with normal hearing Hitachi ETG-4000 system with 52 channels Bilateral prefrontal cortex, IFG HbO; HbR Acoustic Degraded speech

Acoustic Challenges to Listening

The majority of studies manipulated listening effort with some manipulation of acoustic challenge—most commonly, presenting one or more varieties of speech in noise or speech degradation. Wijayasiri et al 136 were one of the first groups to investigate potential neural correlates of effortful listening with fNIRS using degraded speech. Young adults with normal hearing were presented simultaneously with a sentence and a nonspeech auditory distractor. The sentence was either presented unaltered (clear) or degraded by noise-vocoding, which is created by splitting the speech signal into frequency bands, then extracting the amplitude envelope and using that envelope to modulate noise in that frequency band. The bands are then recombined to create the full noise-vocoded signal. 137 The authors used four channels in their degraded speech, which was intended to keep the speech intelligible while reducing spectral clarity. Participants were instructed to attend to the speech or the distractor. They found that left IFG activation increased for degraded relative to clear speech, but only when attention was directed toward the speech—when attention was directed toward the distractor, there was no difference in IFG activation for degraded versus clear speech. Additionally, when listening was directed to degraded speech, the peak in the hemodynamic response was delayed in the left IFG relative to the superior temporal cortex. These results aligned with prior fMRI work suggesting that understanding-degraded speech relies on attention. 138

Lawrence et al 139 expanded on the use of degraded speech in effortful listening—instead of a single level of degradation, they developed five levels of intelligibility for eight-channel noise-vocoded speech by raising the extracted acoustic envelope of each channel to a fractional power, thereby manipulating the depth of envelope modulation for each channel. Behavioral pilot testing allowed the fractional power of each condition to be set so that there was a target accuracy of 0, 25, 50, 75, and 100% correct keywords. Participants listened to the vocoded Bamford–Kowal–Bench (BKB) sentences at the five levels of intelligibility and indicated whether a probe word that appeared on the screen had been spoken in the sentence they heard on each trial. They found that activity in bilateral superior temporal cortex increased monotonically with speech intelligibility, suggesting that as speech becomes more intelligible (i.e., less effortful to understand), engagement of auditory regions increases. In left IFG, however, activation was highest for the 25 and 50% intelligibility conditions, with the lowest activation for the most challenging 0% intelligible condition. This seems to suggest that when speech is impossible to understand, the listener may not attend to the signal, therefore decreasing engagement of left IFG relative to circumstances when speech is challenging but still intelligible. Some channels over bilateral inferior parietal regions (IPS and IPL) and middle temporal lobe also showed deactivation that was most pronounced for intermediate levels of intelligibility, indicating a “ U -shaped” pattern of deactivation relative to silence where activity in those regions was lower for moderately intelligible speech compared to highly intelligible or unintelligible speech. The authors suggested that deactivation in these more posterior regions reflects attentional demands of effortful listening to moderately intelligible speech. Importantly, greater activation was found in left IFG and bilateral posterior superior temporal cortex for correct versus incorrect trials. This increased activation on correct trials might have been due to increased attention or comprehension.

Two recent studies from the same group extended their investigation of cortical activations to degraded speech to children with normal hearing. Lawrence et al 140 presented the same degraded speech stimuli from their 2018 study, in addition to a clear speech condition, to children between the ages of 6 and 13. While the children were slower to respond and less accurate overall relative to adults on the behavioral task, they displayed similarly monotonic increases in accuracy as speech intelligibility increased. A similar pattern of increases in activation as speech intelligibility increased from their 2018 study with adults was also observed in left superior temporal cortex in children, but was not found in right superior temporal cortex. This increased left lateralization of speech processing in children has been found in previous pediatric studies using both fMRI and fNIRS and may be related to greater brain–scalp distance in right frontal and temporal regions relative to the left hemisphere, which is less apparent in adults. Activation of left IFG also increased monotonically with increased intelligibility, which was a departure from the adult findings that left IFG activation was greatest for intermediate levels of intelligibility and lower for highly intelligible and unintelligible speech. The authors suggest that some listening effort was required even when speech was most intelligible, as children did not reach 100% accuracy in the task. In addition, while a trend toward similar patterns of activation for correct versus incorrect trials was found in children who approximated the adult findings from 2018, the result was not statistically significant after correcting for multiple comparisons.

Mushtaq et al 141 used the same BKB sentence stimuli with children aged 6 to 12, but with an additional method of signal degradation. Instead of all conditions using 8-channel noise-vocoded speech, they utilized three levels of vocoding with one, four, and eight channels, with one channel being least intelligible and eight channels being most intelligible. The eight-channel speech was also manipulated with full, reduced, and zero modulation depth of the amplitude envelope in a manner similar to Lawrence et al, 139 140 with resulting intelligibility decreases as modulation depth was reduced. They again found that activation in left temporal regions increased as speech intelligibility increased, as in their 2018 study in children, while deactivation relative to silence in right inferior parietal regions increased with greater intelligibility. Unlike the findings of the previous two studies, however, there was no significant relationship between intelligibility and activation of left IFG, which the authors suggest may be due to insufficient spatial resolution in this region of interest.

An additional source of acoustic challenge for listening may arise from the background—the speech signal itself remains intact, but background noise increases the effort required to understand the speech. Speech-in-noise (SIN) is a common manipulation of listening effort, and several fNIRS studies have used various kinds of noise to assess the cortical mechanisms associated with this type of effortful listening. Defenderfer et al 142 completed a direct comparison between SIN and degraded speech by presenting sentences either in quiet, with a −2.8 dB SNR, or with eight-channel noise vocoding to young adults with normal hearing while imaging the left temporal lobe. Participants were asked to repeat back the sentence after each presentation, and behavioral results suggested that they were most accurate for speech in quiet and least accurate for speech in noise. Oxygenation was greatest at several left temporal channels for SIN relative to speech in quiet and noise-vocoded speech, and oxygenation at a subset of channels was also higher for correct versus incorrect trials when collapsed over the three conditions. However, the authors did not find a correlation between individual differences in accuracy and oxygenation at any of the channels. In a follow-up study, Defenderfer et al 143 executed a similar design but with changes in the manipulation of intelligibility. They also included speech in quiet and eight-channel noise-vocoded speech as conditions, but used SIN at two different levels of intelligibility (+10 dB and −4 dB), and added a condition that combined SIN and noise vocoding. They also imaged bilateral inferior frontal and temporal cortices, rather than focusing solely on the left temporal cortex. The combination of SIN and vocoding was the most challenging condition, leading to lowest behavioral accuracy on the task. fNIRS channels corresponding to MTG and middle frontal gyrus (MFG), a large patch of the frontal lobe that encompasses the DLPFC, showed increased activation associated with low intelligibility conditions: MTG activity was related to processing the SIN-vocoded speech, while MFG activity increased for SIN. The authors suggested that this differential activation based on noise type indicates that there is less reliance on frontal attention mechanisms for degraded speech that is still intelligible, while frontal engagement for SIN implies an increase in attention to parse the signal out from the noise.

Noise can also reflect more naturalistic listening conditions, thereby increasing the external validity of findings regarding listening effort. Two fNIRS articles have explored using more ecologically valid types of noise to assess listening effort. Rowland et al 144 used stimuli from the RealSpeech content library, which combines narrative speech with binaural recordings of background sounds from various environments at different SNRs. For the current study, they chose passages with SNRs that varied over time between −16 dB and +25 dB, corresponding to transitions in the acoustic environment (e.g., a swimming pool, cafe, traveling car). The authors used intersubject correlation to analyze differences in fNIRS activation across different environments, which identified bilateral STG, IFG, DLPFC, and frontopolar cortex as regions that responded similarly across participants during attentive listening to speech in complex naturalistic scenes. However, the left IFG and DLPFC did not show the expected statistically significant relationship between levels of activation and the degree of self-reported effort during listening. The authors propose that this finding may be a result of either too much variability in the self-reported effort measure across participants, or the inherent variability in using complex acoustic scenes with higher external validity. Nevertheless, this study was an important step toward studying naturalistic sources of acoustic challenge using fNIRS.

Bell et al 109 took the ecological validity of stimuli a step further by presenting a complex acoustic scene via a virtual acoustic environment (VAE), which used a four-channel array of speakers at each corner of a sound booth to more closely approximate the acoustic features of sound encountered outside the lab. Bell et al also included an additional source of listening challenge, as half of their participants had hearing loss and used hearing aids. Hearing loss serves as an internal source of acoustic challenge and may compound upon external challenges faced from signal degradation or environmental noise—so it is especially important to understand the additional role of hearing loss/hearing aid use when studying the effort required for real-world listening challenges. This study served as a pilot test of the VAE, using a simulated classroom environment with three children who had normal hearing and three children who used hearing aids. Participants were seated in the booth with a target speaker positioned in front of them, which was emitting a five-keyword target sentence while two-talker distractor speech was presented elsewhere in the room with varying degrees of reverberation to simulate the acoustic properties of difference classrooms. The sound levels of the target speech and environment were calibrated to each child's level of hearing. Participants were asked to attend to and verbally repeat the five-keyword target sentence. Behaviorally, the three children with normal hearing benefitted from lower reverberation and greater spatial separation between the target speaker and distractor speakers. The children with hearing aids had more variable performance, though their performance was worse across all conditions compared to the normal-hearing children. The three children with normal hearing also demonstrated greater activation of left IFG for specific challenging listening conditions (e.g., when the target talker and distractor speech shared the same pitch); however, neural results across hearing-aided children were highly variable and there was no consistent effect in left IFG. Definitive conclusions regarding differential processing of naturalistic SIN for normal hearing versus hearing aids are not possible with such a small sample size, but the VAE shows promise for systematically varying acoustic cues to reflect more ecologically valid listening environments.

Rovetti et al 108 also investigated cortical correlates of listening effort in hearing aid users. Rather than children with hearing loss, this study tested older adults with hearing loss who use hearing aids. Participants with hearing loss completed a visual verbal n-back task and an auditory verbal n-back task at four levels of working memory demand with and without their hearing aids, comparing them to a control group of adults with normal hearing. The authors specifically targeted frontal regions with their fNIRS configuration, in opposition to other work discussed here that primarily focused on temporal regions with some channels in frontal and/or parietal regions. Participants showed slower reaction times for the highest level of working memory demand (recall of the item presented immediately, 1-back, 2-back, or 3-back), while reaction times were faster for the visual relative to auditory n-back task. Hearing aid use did not affect reaction time, and it did not modulate the relationship between self-reported effort and working memory demand. With respect to brain correlates, oxygenation across the entire prefrontal cortex (PFC) increased as working memory demand increased reaching asymptote at 2-back. Interestingly, there was no difference in PFC oxygenation for the auditory versus visual n-back task, and no effect of hearing aid use on PFC oxygenation. The authors concluded that PFC activation is associated with more challenging listening, as indicated by the increased oxygenation as working memory demands increased in the auditory n-back task. However, the lack of effect of hearing aid use on oxygenation suggests that hearing aids do not necessarily decrease listening effort. One plausible explanation for the lack of an effect of hearing aid on listening effort is that while hearing aids do amplify the target signal, they also amplify other irrelevant background sounds, so they do not reliably increase the SNR of the listening environment. Additionally, they can distort sound or produce feedback, which add further sources of noise.

Linguistic Challenges to Listening

While acoustic challenges tend to be most salient when considering effortful listening, the linguistic aspects of a speech signal can also vary in difficulty and therefore require different amounts of effort for successful understanding. These linguistic challenges range from semantic difficulty to syntactic and lexical complexity, and are inherent to the stimulus itself regardless of acoustic or environmental factors. A study by Rovetti et al 145 manipulated listening effort via both acoustic and linguistic challenge. They presented normal-hearing young adults with sentences that provided either high or low sentence context. For example, a high-context sentence (“She had to vacuum the rug .”) contains contextual information that aids in predicting the final word of the sentence, while a low-context sentence (“Tom could have thought about the sport .”) is lacking in contextual cues. These high- and low-context sentences were presented in multi-talker babble at two SNRs that varied in acoustic difficulty (+4 dB “easy” vs. −2 dB “hard”). Participants were asked to repeat the last word of the sentence. Similar to their earlier work, 108 only bilateral PFC was imaged. Behavioral accuracy was lowest for low context/hard SNR speech, and highest for high-context/easy SNR speech. Oxygenation increased across the entire PFC for the hard SNR compared to easy SNR speech, suggesting greater frontal engagement and attention for greater acoustic challenges. There was also a significant effect of context only in the left lateral PFC, with greater oxygenation for low- versus high-context sentences. The left lateral PFC encompasses left IFG as well as left DLPFC. These results suggest that while PFC engages more generally for understanding acoustically challenging speech, left lateral regions of PFC are responsive to both acoustic and linguistic challenges.

White and Langdon 146 further investigated the interplay between multiple sources of acoustic and linguistic challenges to listening with a plausibility judgment task. Normal-hearing young adults were presented with sentences that were either semantically plausible (e.g., people feeding birds ) or implausible (e.g., birds feeding people ). The sentences were also manipulated in terms of syntactic complexity, with simpler subject-relative clauses (e.g., women that feed birds ) versus more complex object-relative clauses (e.g., birds that women feed ). The acoustic properties of the sentences were manipulated by varying the speech rate (normal vs. fast) and degrading the speech via either eight-channel noise vocoding to simulate cochlear implant speech, or a hearing aid simulation which compressed the speech with a limited spectral range. This resulted in a total of four interacting manipulations of listening challenge: two linguistic and two acoustic in nature. Participants listened to each sentence and made judgments about their plausibility or implausibility, and also self-reported their effort on each trial. Both sources of acoustic challenge (speech degradation and speech rate) predicted task accuracy and reaction time, and self-reported effort correlated with behavioral performance. Additional interactions between speech degradation, speech rate, and syntactic complexity also predicted accuracy. As the degree of listening challenge increased, activity in right MFG increased until the most difficult condition (noise-vocoding × fast speech rate × high syntactic complexity × low plausibility), at which point activation dropped off. The authors suggested that this right MFG activity reflected sustained engagement of auditory attentional resources with increased listening challenge, and the drop in activation for the most difficult condition indicates decreased attention related to a loss in motivation to understand the signal. More generally, these findings are consistent with the proposal that the right MFG is a site of convergence between two neuroanatomically distinct attention networks. 147 148 The former is bilateral and thought to be involved with the top-down voluntary allocation of attentional resources, and the latter is largely right-lateralized and thought to be involved with reorientation of attention to unexpected events. 149

Increased listening challenge also correlated with greater engagement of regions involved in the core speech and language network: easier conditions primarily engaged IFG and STG, and as listening challenge increased, activation in posterior temporal regions and DLPFC increased. The authors noted that this shift in activation suggests a hierarchical organization of linguistic processing, and that these regions parallel the dorsal and ventral streams posed in Hickok and Poeppel's dual stream model of speech perception. 150 They proposed that easier listening conditions engage more of the ventral stream, and as listening challenges increase, the dorsal stream is recruited to activate associated motor plans that may aid in speech understanding. Note that this dorsal stream of the speech network is primarily left-lateralized and distinct from the bilateral frontoparietal (dorsal) attention network mentioned earlier, though both are engaged more heavily during challenging listening conditions.

Summary and Future Directions of Listening Effort and fNIRS Work

Several of the listening effort fNIRS studies described here have confirmed involvement of regions associated with effortful listening through prior studies conducted with fMRI, EEG, and MEG. Multiple fNIRS studies replicated increased activation of domain-specific resources inclusive of the STG and left IFG in response to greater listening challenges, particularly in regard to increased linguistic challenge. The engagement of right IFG and domain-general frontal areas like the DLPFC also increases with greater listening challenge. Importantly, the two studies that include hearing aid users also provide early evidence of changes in activation in left IFC and DLPFC that correspond to listening challenge, thereby supporting the use of fNIRS to study listening effort in hard-of-hearing populations with hearing instruments. 108 109

There is still much to learn about the interactions between frontal areas associated with motor planning and attention, and temporal regions associated with core auditory processing. Few studies have imaged areas beyond temporal cortices and IFG/DLPFC. Some work with fNIRS has implicated the role of parietal regions as part of a frontoparietal attention network, and additional frontal regions such as anterior cingulate cortex and premotor cortex that have not specifically been investigated here. 31 As the density of commercially available fNIRS systems increases, more opportunities to image a larger portion of the cortex arise. For example, a preprint by Sherafati et al 107 has used HD-DOT to compare cortical activations in cochlear implant users and normal-hearing adults, allowing high-resolution imaging across a wider scalp area at varying depths of cortical penetration.

There are also some inconsistencies across currently published fNIRS studies of listening effort, most notably in conflicting results regarding the relationship between behavioral indicators of effort (e.g., reaction time, accuracy, self-reported effort) and neural indicators of listening effort. For example, higher levels of cortical activity were found for correct versus incorrect trials in normal-hearing adults by Lawrence et al 139 and Defenderfer et al, 142 but a significant relationship between accuracy and neural activity was not found in children by Lawrence et al 140 and Mushtaq et al. 141 White and Langdon 146 found a correlation between self-reported effort and accuracy, but neither measure correlated with neural activity. Some of these inconsistencies are likely due to variability in types of listening challenge, task demands, and subject pool. 151 However, they may also be due to variability in analysis and statistical techniques. As fNIRS is a relatively newer technology in auditory cognitive neuroscience, there is a wide variety of approaches to analysis, and little consensus on best practices for collecting and analyzing data. Some of the studies described here analyzed concentrations of HbO, HbR, and/or HbT. Others quantified response amplitude after statistical techniques were used to regress out sources of physiological noise (e.g., beta weights after a general linear model was applied), and still others utilized transformations of the concentration data with a hemodynamic modality separation algorithm so that activation was quantified relative to silence. Each technique varies in complexity and has different implications for the conclusions drawn about cortical activation. 152 As the field of work in fNIRS progresses and best practices are established, more consistency in the techniques used to collect and analyze fNIRS data should help clarify the relationship between neural and behavioral indicators of listening effort.

Several opportunities exist for future research in listening effort using fNIRS. The variability of acoustic and linguistic challenges used in the papers reviewed here highlight the many possibilities to investigate interactions of different types of listening challenge, and how those interactions may affect activity in regions like left IFG and DLPFC. Importantly, as fNIRS is uniquely compatible with hearing instruments, new insights are possible in understanding how use of a cochlear implant or hearing aid impacts listening effort as a function of listening challenge. There is great potential for longitudinal work to be completed pre- and post-cochlear implantation, or with/without use of hearing aids, to better understand how hearing instruments affect listening effort. Additionally, an important component of listening effort that was emphasized by Pichora-Fuller et al 4 but not directly studied in any current work in fNIRS is the role of motivation in the listener. Paradigms that intentionally manipulate motivation at various levels of acoustic and/or linguistic challenge may shed further light on the role of nontemporal regions in listening effort, such as the cingulo-opercular network. More multimodal studies that interface fNIRS with fMRI, EEG, and/or MEG can further connect prior findings from those neuroimaging modalities with hemodynamic signatures of listening effort detected by fNIRS. Both multimodal and standalone fNIRS also have potential for future work analyzing resting-state or task-based functional connectivity between frontal, temporal, and parietal regions of interest that may be associated with behavioral data relating to listening effort. Finally, the portability of many fNIRS systems presents opportunities for more naturalistic listening situations outside of the laboratory. Bell et al 109 and Rowland et al 144 provided some preliminary evidence for what that may look like with simulated naturalistic acoustic environments and scenes, but fNIRS also has the capability to be used for testing in real-world challenging listening environments and situations, such as real-time conversation with a partner in a cafe, museum, or park.

Conclusion

Research using fNIRS is rapidly increasing, with the number of citations doubling every year. 51 fNIRS has many advantages for auditory cognitive neuroscience, including its compatibility with hearing instruments, resilience to motion, and silent or near-silent imaging. fNIRS has been used successfully in published research to corroborate results of previous studies of listening effort undertaken with more established neuroimaging techniques. More importantly, fNIRS has begun to be used in novel paradigms and settings that would have previously been untenable. There is much exciting opportunity for growth in the area of listening effort, and several unique features of fNIRS position it as an ideal technique for the field.

Acknowledgments

The authors would like to thank Karla Kovacek and Joseph Urban for their contributions to the tables and figures in this manuscript. This work was supported in part by the Discovery Grant from the Natural Science and Engineering Research Council (NSERC) of Canada (2017-06969) and the NSERC-Sonova Senior Research Chair in Auditory Cognitive Neuroscience awarded to the second author (IRC 537355-18).

Footnotes

Conflict of Interest None declared.

References

  • 1.Mattys S L, Davis M H, Bradlow A R, Scott S K.Speech recognition in adverse conditions: a review Lang Cogn Process 201227(7–8):953–978. [Google Scholar]
  • 2.McGarrigle R, Munro K J, Dawes P. Listening effort and fatigue: what exactly are we measuring? A British Society of Audiology Cognition in Hearing Special Interest Group ‘white paper’. Int J Audiol. 2014;53(07):433–440. doi: 10.3109/14992027.2014.890296. [DOI] [PubMed] [Google Scholar]
  • 3.Ohlenforst B, Zekveld A A, Jansma E P. Effects of hearing impairment and hearing aid amplification on listening effort: a systematic review. Ear Hear. 2017;38(03):267–281. doi: 10.1097/AUD.0000000000000396. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Pichora-Fuller M K, Kramer S E, Eckert M A. Hearing impairment and cognitive energy: the framework for understanding effortful listening (FUEL) Ear Hear. 2016;37 01:5S–27S. doi: 10.1097/AUD.0000000000000312. [DOI] [PubMed] [Google Scholar]
  • 5.Hart S G, Staveland L E. Development of NASA-TLX (task load index): results of empirical and theoretical research. Adv Psychol. 1988;52:139–183. [Google Scholar]
  • 6.Seeman S, Sims R. Comparison of psychophysiological and dual-task measures of listening effort. J Speech Lang Hear Res. 2015;58(06):1781–1792. doi: 10.1044/2015_JSLHR-H-14-0180. [DOI] [PubMed] [Google Scholar]
  • 7.Johnson J, Xu J, Cox R, Pendergraft P. A comparison of two methods for measuring listening effort as part of an audiologic test battery. Am J Audiol. 2015;24(03):419–431. doi: 10.1044/2015_AJA-14-0058. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Koelewijn T, de Kluiver H, Shinn-Cunningham B G, Zekveld A A, Kramer S E. The pupil response reveals increased listening effort when it is difficult to focus attention. Hear Res. 2015;323:81–90. doi: 10.1016/j.heares.2015.02.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Picou E M, Ricketts T A. Increasing motivation changes subjective reports of listening effort and choice of coping strategy. Int J Audiol. 2014;53(06):418–426. doi: 10.3109/14992027.2014.880814. [DOI] [PubMed] [Google Scholar]
  • 10.Sarampalis A, Kalluri S, Edwards B, Hafter E. Objective measures of listening effort: effects of background noise and noise reduction. J Speech Lang Hear Res. 2009;52(05):1230–1240. doi: 10.1044/1092-4388(2009/08-0111). [DOI] [PubMed] [Google Scholar]
  • 11.Desjardins J L, Doherty K A. The effect of hearing aid noise reduction on listening effort in hearing-impaired adults. Ear Hear. 2014;35(06):600–610. doi: 10.1097/AUD.0000000000000028. [DOI] [PubMed] [Google Scholar]
  • 12.Pichora-Fuller M K, Schneider B A, Daneman M. How young and old adults listen to and remember speech in noise. J Acoust Soc Am. 1995;97(01):593–608. doi: 10.1121/1.412282. [DOI] [PubMed] [Google Scholar]
  • 13.Keidser G, Best V, Freeston K, Boyce A. Cognitive spare capacity: evaluation data and its association with comprehension of dynamic conversations. Front Psychol. 2015;6:597. doi: 10.3389/fpsyg.2015.00597. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Mishra S, Stenfelt S, Lunner T, Rönnberg J, Rudner M. Cognitive spare capacity in older adults with hearing loss. Front Aging Neurosci. 2014;6:96. doi: 10.3389/fnagi.2014.00096. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.McCoy S L, Tun P A, Cox L C, Colangelo M, Stewart R A, Wingfield A. Hearing loss and perceptual effort: downstream effects on older adults' memory for speech. Q J Exp Psychol A. 2005;58(01):22–33. doi: 10.1080/02724980443000151. [DOI] [PubMed] [Google Scholar]
  • 16.Beatty J. Task-evoked pupillary responses, processing load, and the structure of processing resources. Psychol Bull. 1982;91(02):276–292. [PubMed] [Google Scholar]
  • 17.Zekveld A A, Koelewijn T, Kramer S E. The pupil dilation response to auditory stimuli: current state of knowledge. Trends Hear. 2018;22:2.331216518777174E15. doi: 10.1177/2331216518777174. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Winn M B, Wendt D, Koelewijn T, Kuchinsky S E. Best practices and advice for using pupillometry to measure listening effort: an introduction for those who want to get started. Trends Hear. 2018;22:2.331216518800869E15. doi: 10.1177/2331216518800869. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Kramer S E, Kapteyn T S, Festen J M, Kuik D J. Assessing aspects of auditory handicap by means of pupil dilatation. Audiology. 1997;36(03):155–164. doi: 10.3109/00206099709071969. [DOI] [PubMed] [Google Scholar]
  • 20.Winn M B, Edwards J R, Litovsky R Y. The impact of auditory spectral resolution on listening effort revealed by pupil dilation. Ear Hear. 2015;36(04):e153–e165. doi: 10.1097/AUD.0000000000000145. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Piquado T, Isaacowitz D, Wingfield A. Pupillometry as a measure of cognitive effort in younger and older adults. Psychophysiology. 2010;47(03):560–569. doi: 10.1111/j.1469-8986.2009.00947.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Mackersie C L, Cones H. Subjective and psychophysiological indexes of listening effort in a competing-talker task. J Am Acad Audiol. 2011;22(02):113–122. doi: 10.3766/jaaa.22.2.6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Mackersie C L, Calderon-Moultrie N. Autonomic nervous system reactivity during speech repetition tasks: heart rate variability and skin conductance. Ear Hear. 2016;37 01:118S–125S. doi: 10.1097/AUD.0000000000000305. [DOI] [PubMed] [Google Scholar]
  • 24.Richter M. The moderating effect of success importance on the relationship between listening demand and listening effort. Ear Hear. 2016;37 01:111S–117S. doi: 10.1097/AUD.0000000000000295. [DOI] [PubMed] [Google Scholar]
  • 25.Francis A L, MacPherson M K, Chandrasekaran B, Alvar A M. Autonomic nervous system responses during perception of masked speech may reflect constructs other than subjective listening effort. Front Psychol. 2016;7:263. doi: 10.3389/fpsyg.2016.00263. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Bess F H, Gustafson S J, Corbett B A, Lambert E W, Camarata S M, Hornsby B WY. Salivary cortisol profiles of children with hearing loss. Ear Hear. 2016;37(03):334–344. doi: 10.1097/AUD.0000000000000256. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Kramer S E, Teunissen C E, Zekveld A A. Cortisol, chromogranin A, and pupillary responses evoked by speech recognition tasks in normally hearing and hard-of-hearing listeners: a pilot study. Ear Hear. 2016;37 01:126S–135S. doi: 10.1097/AUD.0000000000000311. [DOI] [PubMed] [Google Scholar]
  • 28.Davis M H, Johnsrude I S. Hierarchical processing in spoken language comprehension. J Neurosci. 2003;23(08):3423–3431. doi: 10.1523/JNEUROSCI.23-08-03423.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Peelle J E, Johnsrude I S, Davis M H. Hierarchical processing for speech in human auditory cortex and beyond. Front Hum Neurosci. 2010;4:51. doi: 10.3389/fnhum.2010.00051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Rodd J M, Longe O A, Randall B, Tyler L K. The functional organisation of the fronto-temporal language system: evidence from syntactic and semantic ambiguity. Neuropsychologia. 2010;48(05):1324–1335. doi: 10.1016/j.neuropsychologia.2009.12.035. [DOI] [PubMed] [Google Scholar]
  • 31.Peelle J E. Listening effort: how the cognitive consequences of acoustic challenge are reflected in brain and behavior. Ear Hear. 2018;39(02):204–214. doi: 10.1097/AUD.0000000000000494. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Sadaghiani S, D'Esposito M. Functional characterization of the cingulo-opercular network in the maintenance of tonic alertness. Cereb Cortex. 2015;25(09):2763–2773. doi: 10.1093/cercor/bhu072. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Vaden K I, Jr, Kuchinsky S E, Cute S L, Ahlstrom J B, Dubno J R, Eckert M A. The cingulo-opercular network provides word-recognition benefit. J Neurosci. 2013;33(48):18979–18986. doi: 10.1523/JNEUROSCI.1417-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Vaden K I, Jr, Kuchinsky S E, Ahlstrom J B, Dubno J R, Eckert M A. Cortical activity predicts which older adults recognize speech in noise and when. J Neurosci. 2015;35(09):3929–3937. doi: 10.1523/JNEUROSCI.2908-14.2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Botvinick M M, Cohen J D, Carter C S. Conflict monitoring and anterior cingulate cortex: an update. Trends Cogn Sci. 2004;8(12):539–546. doi: 10.1016/j.tics.2004.10.003. [DOI] [PubMed] [Google Scholar]
  • 36.Eckert M A, Teubner-Rhodes S, Vaden K I., Jr Is listening in noise worth it? The neurobiology of speech recognition in challenging listening conditions. Ear Hear. 2016;37, Suppl 1 01:101S–110S. doi: 10.1097/AUD.0000000000000300. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Obleser J, Weisz N. Suppressed alpha oscillations predict intelligibility of speech and its acoustic details. Cereb Cortex. 2012;22(11):2466–2477. doi: 10.1093/cercor/bhr325. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Obleser J, Wöstmann M, Hellbernd N, Wilsch A, Maess B. Adverse listening conditions and memory load drive a common α oscillatory network. J Neurosci. 2012;32(36):12376–12383. doi: 10.1523/JNEUROSCI.4908-11.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.McMahon C M, Boisvert I, de Lissa P. Monitoring alpha oscillations and pupil dilation across a performance-intensity function. Front Psychol. 2016;7:745. doi: 10.3389/fpsyg.2016.00745. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Dimitrijevic A, Smith M L, Kadis D S, Moore D R. Neural indices of listening effort in noisy environments. Sci Rep. 2019;9(01):11278. doi: 10.1038/s41598-019-47643-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Paul B T, Chen J, Le T, Lin V, Dimitrijevic A. Cortical alpha oscillations in cochlear implant users reflect subjective listening effort during speech-in-noise perception. PLoS One. 2021;16(07):e0254162. doi: 10.1371/journal.pone.0254162. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Doelling K B, Arnal L H, Ghitza O, Poeppel D.Acoustic landmarks drive delta-theta oscillations to enable speech comprehension by facilitating perceptual parsing Neuroimage 201485(0 2):761–768. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Etard O, Reichenbach T. Neural speech tracking in the theta and in the delta frequency band differentially encode clarity and comprehension of speech in noise. J Neurosci. 2019;39(29):5750–5759. doi: 10.1523/JNEUROSCI.1828-18.2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Wisniewski M G. Indices of effortful listening can be mined from existing electroencephalographic data. Ear Hear. 2017;38(01):e69–e73. doi: 10.1097/AUD.0000000000000354. [DOI] [PubMed] [Google Scholar]
  • 45.Peelle J E, Gross J, Davis M H. Phase-locked responses to speech in human auditory cortex are enhanced during comprehension. Cereb Cortex. 2013;23(06):1378–1387. doi: 10.1093/cercor/bhs118. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Obleser J, Kotz S A. Multiple brain signatures of integration in the comprehension of degraded speech. Neuroimage. 2011;55(02):713–723. doi: 10.1016/j.neuroimage.2010.12.020. [DOI] [PubMed] [Google Scholar]
  • 47.Song J, Iverson P. Listening effort during speech perception enhances auditory and lexical processing for non-native listeners and accents. Cognition. 2018;179:163–170. doi: 10.1016/j.cognition.2018.06.001. [DOI] [PubMed] [Google Scholar]
  • 48.Kuchinsky S E, Vaden K I.Aging, hearing loss, and listening effort: imaging studies of the aging listenerIn: Helfer KS, Bartlett EL, Popper AN, Fay RR, eds. Aging and Hearing, Springer Handbook of Auditory Research. Springer;202072231–256. [Google Scholar]
  • 49.Wu Y-H, Aksan N, Rizzo M, Stangl E, Zhang X, Bentler R. Measuring listening effort: driving simulator versus simple dual-task paradigm. Ear Hear. 2014;35(06):623–632. doi: 10.1097/AUD.0000000000000079. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Kuruvilla-Mathew A, Purdy S C, Welch D. Cortical encoding of speech acoustics: effects of noise and amplification. Int J Audiol. 2015;54(11):852–864. doi: 10.3109/14992027.2015.1055838. [DOI] [PubMed] [Google Scholar]
  • 51.Pinti P, Tachtsidis I, Hamilton A. The present and future use of functional near-infrared spectroscopy (fNIRS) for cognitive neuroscience. Ann N Y Acad Sci. 2020;1464(01):5–29. doi: 10.1111/nyas.13948. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Sfareni R, Boffi A, Quaresima V, Ferrari M. Near infrared absorption spectra of human deoxy- and oxyhaemoglobin in the temperature range 20-40 degrees C. Biochim Biophys Acta. 1997;1340(02):165–169. doi: 10.1016/s0167-4838(97)00042-3. [DOI] [PubMed] [Google Scholar]
  • 53.Wray S, Cope M, Delpy D T, Wyatt J S, Reynolds E O. Characterization of the near infrared absorption spectra of cytochrome aa3 and haemoglobin for the non-invasive monitoring of cerebral oxygenation. Biochim Biophys Acta. 1988;933(01):184–192. doi: 10.1016/0005-2728(88)90069-2. [DOI] [PubMed] [Google Scholar]
  • 54.Quaresima V, Bisconti S, Ferrari M. A brief review on the use of functional near-infrared spectroscopy (fNIRS) for language imaging studies in human newborns and adults. Brain Lang. 2012;121(02):79–89. doi: 10.1016/j.bandl.2011.03.009. [DOI] [PubMed] [Google Scholar]
  • 55.Liang L-Y, Shewokis P A, Getchell N. Brain activation in the prefrontal cortex during motor and cognitive tasks in adults. J Behav Brain Sci. 2016;06(12):463–474. [Google Scholar]
  • 56.Pinti P, Scholkmann F, Hamilton A, Burgess P, Tachtsidis I. Current status and issues regarding pre-processing of fNIRS neuroimaging data: an investigation of diverse signal filtering methods within a general linear model framework. Front Hum Neurosci. 2019;12:505. doi: 10.3389/fnhum.2018.00505. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Zhang Y, Zhu C. Assessing brain networks by resting-state dynamic functional connectivity: an fNIRS-EEG study. Front Neurosci. 2020;13:1430. doi: 10.3389/fnins.2019.01430. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Hu Z, Liu G, Dong Q, Niu H. Applications of resting-state fNIRS in the developing brain: a review from the connectome perspective. Front Neurosci. 2020;14:476. doi: 10.3389/fnins.2020.00476. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Zhang X, Noah J A, Dravida S, Hirsch J. Optimization of wavelet coherence analysis as a measure of neural synchrony during hyperscanning using functional near-infrared spectroscopy. Neurophotonics. 2020;7(01):15010. doi: 10.1117/1.NPh.7.1.015010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Hoshi Y, Tamura M. Dynamic multichannel near-infrared optical imaging of human brain activity. J Appl Physiol. 1993;75(04):1842–1846. doi: 10.1152/jappl.1993.75.4.1842. [DOI] [PubMed] [Google Scholar]
  • 61.Villringer A, Planck J, Hock C, Schleinkofer L, Dirnagl U.Near infrared spectroscopy (NIRS): a new tool to study hemodynamic changes during activation of brain function in human adults Neurosci Lett 1993154(1-2):101–104. [DOI] [PubMed] [Google Scholar]
  • 62.Kato T, Kamei A, Takashima S, Ozaki T. Human visual cortical function during photic stimulation monitoring by means of near-infrared spectroscopy. J Cereb Blood Flow Metab. 1993;13(03):516–520. doi: 10.1038/jcbfm.1993.66. [DOI] [PubMed] [Google Scholar]
  • 63.Meek J H, Elwell C E, Khan M J.Regional changes in cerebral haemodynamics as a result of a visual stimulus measured by near infrared spectroscopy Proc Biol Sci 1995261(1362):351–356. [DOI] [PubMed] [Google Scholar]
  • 64.Villringer A, Chance B. Non-invasive optical spectroscopy and imaging of human brain function. Trends Neurosci. 1997;20(10):435–442. doi: 10.1016/s0166-2236(97)01132-6. [DOI] [PubMed] [Google Scholar]
  • 65.Watanabe E, Yamashita Y, Maki A, Ito Y, Koizumi H. Non-invasive functional mapping with multi-channel near infra-red spectroscopic topography in humans. Neurosci Lett. 1996;205(01):41–44. doi: 10.1016/0304-3940(96)12376-4. [DOI] [PubMed] [Google Scholar]
  • 66.Miyai I, Tanabe H C, Sase I. Cortical mapping of gait in humans: a near-infrared spectroscopic topography study. Neuroimage. 2001;14(05):1186–1192. doi: 10.1006/nimg.2001.0905. [DOI] [PubMed] [Google Scholar]
  • 67.Ferrari M, Quaresima V. A brief review on the history of human functional near-infrared spectroscopy (fNIRS) development and fields of application. Neuroimage. 2012;63(02):921–935. doi: 10.1016/j.neuroimage.2012.03.049. [DOI] [PubMed] [Google Scholar]
  • 68.Scholkmann F, Kleiser S, Metz A J.A review on continuous wave functional near-infrared spectroscopy and imaging instrumentation and methodology Neuroimage 201485(Pt 1):6–27. [DOI] [PubMed] [Google Scholar]
  • 69.Zhao H, Cooper R J. Review of recent progress toward a fiberless, whole-scalp diffuse optical tomography system. Neurophotonics. 2018;5(01):11012. doi: 10.1117/1.NPh.5.1.011012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Pinti P, Aichelburg C, Gilbert S. A review on the use of wearable functional near-infrared spectroscopy in naturalistic environments. Jpn Psychol Res. 2018;60(04):347–373. doi: 10.1111/jpr.12206. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Wheelock M D, Culver J P, Eggebrecht A T. High-density diffuse optical tomography for imaging human brain function. Rev Sci Instrum. 2019;90(05):51101. doi: 10.1063/1.5086809. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Frijia E M, Billing A, Lloyd-Fox S. Functional imaging of the developing brain with wearable high-density diffuse optical tomography: a new benchmark for infant neuroimaging outside the scanner environment. Neuroimage. 2021;225:117490. doi: 10.1016/j.neuroimage.2020.117490. [DOI] [PubMed] [Google Scholar]
  • 73.Peelle J E. Methodological challenges and solutions in auditory functional magnetic resonance imaging. Front Neurosci. 2014;8:253. doi: 10.3389/fnins.2014.00253. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Rovetti J, Goy H, Nurgitz R, Russo F A. Comparing verbal working memory load in auditory and visual modalities using functional near-infrared spectroscopy. Behav Brain Res. 2021;402:113102. doi: 10.1016/j.bbr.2020.113102. [DOI] [PubMed] [Google Scholar]
  • 75.Schumacher E H, Lauber E, Awh E, Jonides J, Smith E E, Koeppe R A. PET evidence for an amodal verbal working memory system. Neuroimage. 1996;3(02):79–88. doi: 10.1006/nimg.1996.0009. [DOI] [PubMed] [Google Scholar]
  • 76.Crottaz-Herbette S, Anagnoson R T, Menon V. Modality effects in verbal working memory: differential prefrontal and parietal responses to auditory and visual stimuli. Neuroimage. 2004;21(01):340–351. doi: 10.1016/j.neuroimage.2003.09.019. [DOI] [PubMed] [Google Scholar]
  • 77.Rodriguez-Jimenez R, Avila C, Garcia-Navarro C. Differential dorsolateral prefrontal cortex activation during a verbal n-back task according to sensory modality. Behav Brain Res. 2009;205(01):299–302. doi: 10.1016/j.bbr.2009.08.022. [DOI] [PubMed] [Google Scholar]
  • 78.Toronov V, Walker S, Gupta R. The roles of changes in deoxyhemoglobin concentration and regional cerebral blood volume in the fMRI BOLD signal. Neuroimage. 2003;19(04):1521–1531. doi: 10.1016/s1053-8119(03)00152-6. [DOI] [PubMed] [Google Scholar]
  • 79.Huppert T J, Hoge R D, Diamond S G, Franceschini M A, Boas D A. A temporal comparison of BOLD, ASL, and NIRS hemodynamic responses to motor stimuli in adult humans. Neuroimage. 2006;29(02):368–382. doi: 10.1016/j.neuroimage.2005.08.065. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Ogawa S, Menon R S, Tank D W. Functional brain mapping by blood oxygenation level-dependent contrast magnetic resonance imaging. A comparison of signal characteristics with a biophysical model. Biophys J. 1993;64(03):803–812. doi: 10.1016/S0006-3495(93)81441-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Kim S G, Richter W, Uğurbil K. Limitations of temporal resolution in functional MRI. Magn Reson Med. 1997;37(04):631–636. doi: 10.1002/mrm.1910370427. [DOI] [PubMed] [Google Scholar]
  • 82.Harrison S C, Hartley D E. Shedding light on the human auditory cortex: a review of the advances in near infrared spectroscopy (NIRS) Rep Med Ins. 2019;12:31–42. [Google Scholar]
  • 83.Hebden J C, Cooper R J, Gibson A, Everdell N, Austin T.Simultaneous EEG and diffuse optical imaging of seizure-related hemodynamic activity in the newborn infant brainIn: Popp J, Drexler W, Tuchin VV, Matthews DL, eds. Proceedings of SPIE. Published online June 1, 2012 10.1117/12.922990 [DOI]
  • 84.Leamy D J, Ward T E. A novel co-locational and concurrent fNIRS/EEG measurement system: design and initial results. Annu Int Conf IEEE Eng Med Biol Soc. 2010;2010:4230–4233. doi: 10.1109/IEMBS.2010.5627377. [DOI] [PubMed] [Google Scholar]
  • 85.Habermehl C, Holtze S, Steinbrink J. Somatosensory activation of two fingers can be discriminated with ultrahigh-density diffuse optical tomography. Neuroimage. 2012;59(04):3201–3211. doi: 10.1016/j.neuroimage.2011.11.062. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Toronov V, Webb A, Choi J H. Investigation of human brain hemodynamics by simultaneous near-infrared spectroscopy and functional magnetic resonance imaging. Med Phys. 2001;28(04):521–527. doi: 10.1118/1.1354627. [DOI] [PubMed] [Google Scholar]
  • 87.Zhang X, Toronov V Y, Fabiani M, Gratton G, Webb A G. The study of cerebral hemodynamic and neuronal response to visual stimulation using simultaneous NIR optical tomography and BOLD fMRI in humans. Proc SPIE Int Soc Opt Eng. 2005;5686(115):566–572. doi: 10.1117/12.593435. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88.Seki Y, Miyashita T, Kandori A, Maki A, Koizumi H. Simultaneous measurement of neuronal activity and cortical hemodynamics by unshielded magnetoencephalography and near-infrared spectroscopy. J Biomed Opt. 2012;17(10):107001. doi: 10.1117/1.JBO.17.10.107001. [DOI] [PubMed] [Google Scholar]
  • 89.Steinmetzger K, Shen Z, Riedel H, Rupp A. Auditory cortex activity measured using functional near-infrared spectroscopy (fNIRS) appears to be susceptible to masking by cortical blood stealing. Hear Res. 2020;396:108069. doi: 10.1016/j.heares.2020.108069. [DOI] [PubMed] [Google Scholar]
  • 90.Pringle J, Roberts C, Kohl M, Lekeux P. Near infrared spectroscopy in large animals: optical path length and influence of hair covering and epidermal pigmentation. Vet J. 1999;158(01):48–52. doi: 10.1053/tvjl.1998.0306. [DOI] [PubMed] [Google Scholar]
  • 91.Wassenaar E B, Van den Brand J GH. Reliability of near-infrared spectroscopy in people with dark skin pigmentation. J Clin Monit Comput. 2005;19(03):195–199. doi: 10.1007/s10877-005-1655-0. [DOI] [PubMed] [Google Scholar]
  • 92.Couch L, Roskosky M, Freedman B A, Shuler M S. Effect of skin pigmentation on near infrared spectroscopy. Am J Anal Chem. 2015;06(12):911–916. [Google Scholar]
  • 93.Papasideris M, Ayaz H, Hall P A. Medial prefrontal brain activity correlates with emerging symptoms of anxiety and depression in late adolescence: a fNIRS study. Dev Psychobiol. 2021;63(07):e22199. doi: 10.1002/dev.22199. [DOI] [PubMed] [Google Scholar]
  • 94.Ohnishi M, Kusakawa N, Masaki S. Measurement of hemodynamics of auditory cortex using magnetoencephalography and near infrared spectroscopy. Acta Otolaryngol Suppl. 1997;532:129–131. doi: 10.3109/00016489709126161. [DOI] [PubMed] [Google Scholar]
  • 95.Sakatani K, Chen S, Lichty W, Zuo H, Wang Y P. Cerebral blood oxygenation changes induced by auditory stimulation in newborn infants measured by near infrared spectroscopy. Early Hum Dev. 1999;55(03):229–236. doi: 10.1016/s0378-3782(99)00019-5. [DOI] [PubMed] [Google Scholar]
  • 96.Zaramella P, Freato F, Amigoni A. Brain auditory activation measured by near-infrared spectroscopy (NIRS) in neonates. Pediatr Res. 2001;49(02):213–219. doi: 10.1203/00006450-200102000-00014. [DOI] [PubMed] [Google Scholar]
  • 97.Blasi A, Lloyd-Fox S, Johnson M H, Elwell C. Test-retest reliability of functional near infrared spectroscopy in infants. Neurophotonics. 2014;1(02):25005. doi: 10.1117/1.NPh.1.2.025005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Hong K-S, Santosa H. Decoding four different sound-categories in the auditory cortex using functional near-infrared spectroscopy. Hear Res. 2016;333:157–166. doi: 10.1016/j.heares.2016.01.009. [DOI] [PubMed] [Google Scholar]
  • 99.Yoo S-H, Santosa H, Kim C-S, Hong K-S. Decoding multiple sound-categories in the auditory cortex by neural networks: an fNIRS study. Front Hum Neurosci. 2021;15:636191. doi: 10.3389/fnhum.2021.636191. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100.Oghalai J S. Functional neuroimaging of speech perception in cochlear implant recipients. Volta Review. 2017;116(01):95–102. [Google Scholar]
  • 101.Bortfeld H. Functional near-infrared spectroscopy as a tool for assessing speech and spoken language processing in pediatric and adult cochlear implant users. Dev Psychobiol. 2019;61(03):430–443. doi: 10.1002/dev.21818. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 102.Harrison S C, Lawrence R, Hoare D J, Wiggins I M, Hartley D EH. Use of functional near-infrared spectroscopy to predict and measure cochlear implant outcomes: a scoping review. Brain Sci. 2021;11(11):1439. doi: 10.3390/brainsci11111439. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 103.Sevy A BG, Bortfeld H, Huppert T J, Beauchamp M S, Tonini R E, Oghalai J S.Neuroimaging with near-infrared spectroscopy demonstrates speech-evoked activity in the auditory cortex of deaf children following cochlear implantation Hear Res 2010270(1-2):39–47. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 104.Bisconti S, Shulkin M, Hu X, Basura G J, Kileny P R, Kovelman I. Functional near-infrared spectroscopy brain imaging investigation of phonological awareness and passage comprehension abilities in adult recipients of cochlear implants. J Speech Lang Hear Res. 2016;59(02):239–253. doi: 10.1044/2015_JSLHR-L-14-0278. [DOI] [PubMed] [Google Scholar]
  • 105.Olds C, Pollonini L, Abaya H. Cortical activation patterns correlate with speech understanding after cochlear implantation. Ear Hear. 2016;37(03):e160–e172. doi: 10.1097/AUD.0000000000000258. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 106.Zhou X, Seghouane A-K, Shah A. Cortical speech processing in postlingually deaf adult cochlear implant users, as revealed by functional near-infrared spectroscopy. Trends Hear. 2018;22:2.33121651878685E15. doi: 10.1177/2331216518786850. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 107.Sherafati A, Dwyer N, Bajracharya A.Dorsolateral prefrontal cortex supports speech perception in listeners with cochlear implants bioRxiv(Preprint posted online October 16, 2021) 10.1101/2021.10.16.464654 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108.Rovetti J, Goy H, Pichora-Fuller M K, Russo F A. Functional near-infrared spectroscopy as a measure of listening effort in older adults who use hearing aids. Trends Hear. 2019;23:2.331216519886722E15. doi: 10.1177/2331216519886722. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.Bell L, Peng Z E, Pausch F. fNIRS assessment of speech comprehension in children with normal hearing and children with hearing aids in virtual acoustic environments: pilot data and practical recommendations. Children (Basel) 2020;7(11):219. doi: 10.3390/children7110219. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 110.Dewey R S, Hartley D EH. Cortical cross-modal plasticity following deafness measured using functional near-infrared spectroscopy. Hear Res. 2015;325:55–63. doi: 10.1016/j.heares.2015.03.007. [DOI] [PubMed] [Google Scholar]
  • 111.Anderson C A, Lazard D S, Hartley D EH. Plasticity in bilateral superior temporal cortex: effects of deafness and cochlear implantation on auditory and visual speech processing. Hear Res. 2017;343:138–149. doi: 10.1016/j.heares.2016.07.013. [DOI] [PubMed] [Google Scholar]
  • 112.Chen L-C, Sandmann P, Thorne J D, Bleichner M G, Debener S. Cross-modal functional reorganization of visual and auditory cortex in adult cochlear implant users identified with fNIRS. Neural Plast. 2016;2016:4.382656E6. doi: 10.1155/2016/4382656. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.Chen L-C, Stropahl M, Schönwiesner M, Debener S. Enhanced visual adaptation in cochlear implant users revealed by concurrent EEG-fNIRS. Neuroimage. 2017;146:600–608. doi: 10.1016/j.neuroimage.2016.09.033. [DOI] [PubMed] [Google Scholar]
  • 114.Anderson C A, Wiggins I M, Kitterick P T, Hartley D EH. Adaptive benefit of cross-modal plasticity following cochlear implantation in deaf adults. Proc Natl Acad Sci U S A. 2017;114(38):10256–10261. doi: 10.1073/pnas.1704785114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115.Anderson C A, Wiggins I M, Kitterick P T, Hartley D EH. Pre-operative brain imaging using functional near-infrared spectroscopy helps predict cochlear implant outcome in deaf adults. J Assoc Res Otolaryngol. 2019;20(05):511–528. doi: 10.1007/s10162-019-00729-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 116.Mushtaq F, Wiggins I M, Kitterick P T, Anderson C A, Hartley D EH. The benefit of cross-modal reorganization on speech perception in pediatric cochlear implant recipients revealed using functional near-infrared spectroscopy. Front Hum Neurosci. 2020;14:308. doi: 10.3389/fnhum.2020.00308. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 117.Pollonini L, Olds C, Abaya H, Bortfeld H, Beauchamp M S, Oghalai J S. Auditory cortex activation to natural speech and simulated cochlear implant speech measured with functional near-infrared spectroscopy. Hear Res. 2014;309:84–93. doi: 10.1016/j.heares.2013.11.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 118.Vannasing P, Florea O, González-Frankenberger B. Distinct hemispheric specializations for native and non-native languages in one-day-old newborns identified by fNIRS. Neuropsychologia. 2016;84:63–69. doi: 10.1016/j.neuropsychologia.2016.01.038. [DOI] [PubMed] [Google Scholar]
  • 119.Giordano V, Alexopoulos J, Spagna A. Accent discrimination abilities during the first days of life: an fNIRS study. Brain Lang. 2021;223:105039. doi: 10.1016/j.bandl.2021.105039. [DOI] [PubMed] [Google Scholar]
  • 120.van de Rijt L PH, van Opstal A J, Mylanus E AM. Temporal cortex activation to audiovisual speech in normal-hearing and cochlear implant users measured with functional near-infrared spectroscopy. Front Hum Neurosci. 2016;10:48. doi: 10.3389/fnhum.2016.00048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 121.Altvater-Mackensen N, Grossmann T. The role of left inferior frontal cortex during audiovisual speech perception in infants. Neuroimage. 2016;133:14–20. doi: 10.1016/j.neuroimage.2016.02.061. [DOI] [PubMed] [Google Scholar]
  • 122.Shader M J, Luke R, Gouailhardou N, McKay C M. The use of broad vs restricted regions of interest in functional near-infrared spectroscopy for measuring cortical activation to auditory-only and visual-only speech. Hear Res. 2021;406:108256. doi: 10.1016/j.heares.2021.108256. [DOI] [PubMed] [Google Scholar]
  • 123.Santosa H, Hong M J, Hong K-S. Lateralization of music processing with noises in the auditory cortex: an fNIRS study. Front Behav Neurosci. 2014;8:418. doi: 10.3389/fnbeh.2014.00418. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 124.Bigliassi M, León-Domínguez U, Altimari L R. How does the prefrontal cortex “listen” to classical and techno music? A functional near-infrared spectroscopy (fNIRS) study. Psychol Neurosci. 2015;8(02):246–256. [Google Scholar]
  • 125.Bicciato G, Keller E, Wolf M. Increase in low-frequency oscillations in fNIRS as cerebral response to auditory stimulation with familiar music. Brain Sci. 2021;12(01):42. doi: 10.3390/brainsci12010042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 126.Ferreri L, Aucouturier J-J, Muthalib M, Bigand E, Bugaiska A. Music improves verbal memory encoding while decreasing prefrontal cortex activity: an fNIRS study. Front Hum Neurosci. 2013;7:779. doi: 10.3389/fnhum.2013.00779. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 127.Ferreri L, Bigand E, Perrey S, Muthalib M, Bard P, Bugaiska A. Less effort, better results: how does music act on prefrontal cortex in older adults during verbal encoding? An fNIRS study. Front Hum Neurosci. 2014;8:301. doi: 10.3389/fnhum.2014.00301. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 128.Ferreri L, Bigand E, Bard P, Bugaiska A. The influence of music on prefrontal cortex during episodic encoding and retrieval of verbal information: a multichannel fNIRS study. Behav Neurol. 2015;2015:707625. doi: 10.1155/2015/707625. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 129.Plichta M M, Gerdes A BM, Alpers G W. Auditory cortex activation is modulated by emotion: a functional near-infrared spectroscopy (fNIRS) study. Neuroimage. 2011;55(03):1200–1207. doi: 10.1016/j.neuroimage.2011.01.011. [DOI] [PubMed] [Google Scholar]
  • 130.Gruber T, Debracque C, Ceravolo L. Human discrimination and categorization of emotions in voices: a functional near-infrared spectroscopy (fNIRS) study. Front Neurosci. 2020;14:570. doi: 10.3389/fnins.2020.00570. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131.Zhang D, Zhou Y, Yuan J. Speech prosodies of different emotional categories activate different brain regions in adult cortex: an fNIRS study. Sci Rep. 2018;8(01):218. doi: 10.1038/s41598-017-18683-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 132.Zhang D, Zhou Y, Hou X, Cui Y, Zhou C. Discrimination of emotional prosodies in human neonates: a pilot fNIRS study. Neurosci Lett. 2017;658:62–66. doi: 10.1016/j.neulet.2017.08.047. [DOI] [PubMed] [Google Scholar]
  • 133.Hirsch J.An fNIRS Approach to Two-person Neuroscience. Biophotonics Congress: Biomedical Optics 2020 (Translational, Microscopy, OCT, OTS, BRAIN)Published online 2020. 10.1364/brain.2020.bm4c.1 [DOI]
  • 134.Hirsch J, Adam Noah J, Zhang X, Dravida S, Ono Y. A cross-brain neural mechanism for human-to-human verbal communication. Soc Cogn Affect Neurosci. 2018;13(09):907–920. doi: 10.1093/scan/nsy070. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 135.Ono Y, Zhang X, Noah J A, Dravida S, Hirsch J. Bidirectional connectivity between Broca's area and Wernicke's area during interactive verbal communication. Brain Connect. 2022;12(03):210–222. doi: 10.1089/brain.2020.0790. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 136.Wijayasiri P, Hartley D EH, Wiggins I M. Brain activity underlying the recovery of meaning from degraded speech: a functional near-infrared spectroscopy (fNIRS) study. Hear Res. 2017;351:55–67. doi: 10.1016/j.heares.2017.05.010. [DOI] [PubMed] [Google Scholar]
  • 137.Shannon R V, Zeng F-G, Kamath V, Wygonski J, Ekelid M.Speech recognition with primarily temporal cues Science 1995270(5234):303–304. [DOI] [PubMed] [Google Scholar]
  • 138.Wild C J, Yusuf A, Wilson D E, Peelle J E, Davis M H, Johnsrude I S. Effortful listening: the processing of degraded speech depends critically on attention. J Neurosci. 2012;32(40):14010–14021. doi: 10.1523/JNEUROSCI.1528-12.2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 139.Lawrence R J, Wiggins I M, Anderson C A, Davies-Thompson J, Hartley D EH. Cortical correlates of speech intelligibility measured using functional near-infrared spectroscopy (fNIRS) Hear Res. 2018;370:53–64. doi: 10.1016/j.heares.2018.09.005. [DOI] [PubMed] [Google Scholar]
  • 140.Lawrence R J, Wiggins I M, Hodgson J C, Hartley D EH. Evaluating cortical responses to speech in children: a functional near-infrared spectroscopy (fNIRS) study. Hear Res. 2021;401:108155. doi: 10.1016/j.heares.2020.108155. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 141.Mushtaq F, Wiggins I M, Kitterick P T, Anderson C A, Hartley D EH. Investigating cortical responses to noise-vocoded speech in children with normal hearing using functional near-infrared spectroscopy (fNIRS) J Assoc Res Otolaryngol. 2021;22(06):703–717. doi: 10.1007/s10162-021-00817-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 142.Defenderfer J, Kerr-German A, Hedrick M, Buss A T. Investigating the role of temporal lobe activation in speech perception accuracy with normal hearing adults: an event-related fNIRS study. Neuropsychologia. 2017;106:31–41. doi: 10.1016/j.neuropsychologia.2017.09.004. [DOI] [PubMed] [Google Scholar]
  • 143.Defenderfer J, Forbes S, Wijeakumar S, Hedrick M, Plyler P, Buss A T. Frontotemporal activation differs between perception of simulated cochlear implant speech and speech in background noise: an image-based fNIRS study. Neuroimage. 2021;240:118385. doi: 10.1016/j.neuroimage.2021.118385. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144.Rowland S C, Hartley D EH, Wiggins I M. Listening in naturalistic scenes: What can functional near-infrared spectroscopy and intersubject correlation analysis tell us about the underlying brain activity? Trends Hear. 2018;22:2.331216518804116E15. doi: 10.1177/2331216518804116. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 145.Rovetti J, Goy H, Zara M, Russo F A. Reduced semantic context and signal-to-noise ratio increase listening effort as measured using functional near-infrared spectroscopy. Ear Hear. 2022;43(03):836–848. doi: 10.1097/AUD.0000000000001137. [DOI] [PubMed] [Google Scholar]
  • 146.White B E, Langdon C. The cortical organization of listening effort: new insight from functional near-infrared spectroscopy. Neuroimage. 2021;240:118324. doi: 10.1016/j.neuroimage.2021.118324. [DOI] [PubMed] [Google Scholar]
  • 147.Corbetta M, Patel G, Shulman G L. The reorienting system of the human brain: from environment to theory of mind. Neuron. 2008;58(03):306–324. doi: 10.1016/j.neuron.2008.04.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 148.Vossel S, Geng J J, Fink G R. Dorsal and ventral attention systems: distinct neural circuits but collaborative roles. Neuroscientist. 2014;20(02):150–159. doi: 10.1177/1073858413494269. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 149.Corbetta M, Shulman G L. Control of goal-directed and stimulus-driven attention in the brain. Nat Rev Neurosci. 2002;3(03):201–215. doi: 10.1038/nrn755. [DOI] [PubMed] [Google Scholar]
  • 150.Hickok G, Poeppel D. The cortical organization of speech processing. Nat Rev Neurosci. 2007;8(05):393–402. doi: 10.1038/nrn2113. [DOI] [PubMed] [Google Scholar]
  • 151.Strand J F, Ray L, Dillman-Hasso N H, Villanueva J, Brown V A. Understanding speech amid the jingle and jangle: recommendations for improving measurement practices in listening effort research. Audit Percept Cogn. 2020;3(04):169–188. doi: 10.1080/25742442.2021.1903293. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 152.Botvinik-Nezer R, Holzmeister F, Camerer C F.Variability in the analysis of a single neuroimaging dataset by many teams Nature 2020582(7810):84–88. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Seminars in Hearing are provided here courtesy of Thieme Medical Publishers

RESOURCES