I. INTRODUCTION
In crowded listening environments, selective attention enables information to be extracted from a talker of interest. However, in many cases it is desirable to retrieve information from a talker who is outside the immediate focus of attention (e.g., when two people talk at once). While some early studies showed that listeners with normal hearing perform poorly when asked to recall messages from unattended talkers (e.g. Cherry 1953), subsequent studies indicated that listeners are able to process unattended speech to some extent (e.g. Moray 1959; Conway et al. 2001; Rivenez et al. 2006) and can perform remarkably well at following two talkers when instructed to do so in advance (Best et al. 2006; Gallun et al. 2007; Ihlefeld and Shinn-Cunningham 2008).
A recent survey of listeners with hearing loss (Gatehouse and Noble 2004) revealed that the self-perception of communication handicap is strong in listening situations calling for divided or rapidly shifting attention, such as following two talkers at once, catching the beginning of what a new speaker says, or engaging in conversation with a group of people. Given these subjective reports, we hypothesized that hearing loss impairs one’s ability to deal effectively with simultaneous messages and that this difficulty would be revealed in objective dual-task listening experiments.
While several studies have examined divided listening in listeners with hearing-impairment (e.g. Strouse et al. 2000; Mackersie et al. 2001; Humes et al. 2006; Singh et al. 2008), the majority have used older listeners, making it difficult to factor out the differential contributions of age and hearing loss to the results. For this reason, we recruited a group of young listeners with sensorineural hearing loss and compared their performance to a similar group of young listeners with normal hearing.
We chose a task in which listeners were required to respond to two simultaneously presented messages (see Best et al. 2006; Ihlefeld and Shinn-Cunningham 2008). In order to focus on the task of processing two simultaneous messages, we made an effort to eliminate other factors that might interfere with the ability of hearing-impaired listeners to hear out the messages. First, the overall presentation level was adjusted on an individual basis to reduce audibility as a factor in the hearing-impaired group. Second, the two messages were presented dichotically (one to each ear) to avoid peripheral interference between the stimuli that would be likely to impair speech intelligibility in the hearing-impaired group (Duquesnoy 1983; Festen and Plomp 1990; Summers and Leek 1998). Note that this design is similar to the classic dichotic digits paradigm introduced by Broadbent (1954) and used or adapted by others since to study attention (e.g. Treisman 1971), hemispheric dominance effects (e.g. Kimura 1961; Bryden 1963; reviewed in Hugdahl 2003), and auditory processing disorders (e.g. Fifer et al. 1983; Jerger and Martin 2006).
Finally, the difficulty of speech reception was varied parametrically by adding noise to the messages at a range of signal-to-noise ratios. This allowed us to assess dual-task performance over various levels of listening difficulty within a group, but more importantly enabled us to compare the two groups at SNRs that roughly equate their speech reception performance.
II. MATERIALS AND METHODS
A. Participants
The participants were nine listeners with normal hearing (NH; 4 male and 5 female) and nine listeners with sensorineural hearing loss (HL; 3 male and 6 female). The NH group ranged in age from 18–29 years (mean 22), and the HL group ranged from 18–42 years (mean 27). The NH listeners were screened to ensure that they had pure-tone thresholds in the normal range (no greater than 20 dB HL) for octave frequencies from 250 Hz to 8000 Hz. The HL listeners had mild to moderately severe, bilateral, symmetric, sloping, sensorineural hearing losses. Seven of the nine were regular bilateral hearing-aid wearers, but participated in the experiment with their aids removed. Mean audiograms for both groups are shown in Fig. 1. All listeners were paid for their participation. The experimental protocols were approved by the Boston University Charles River Campus Institutional Review Board.
Figure 1.
Mean audiograms for listeners with normal hearing (NH) and sensorineural hearing loss (HL). Error bars indicate across-subject standard deviations.
B. Stimuli
Digital stimuli were generated on a PC using MATLAB software (MathWorks Inc.). The stimuli were D/A converted and attenuated using Tucker-Davis Technologies hardware (System II or III), and presented over Sennheiser headphones (HD 580 or HD 265 linear). Listeners were seated in a sound-treated booth fitted with a monitor and mouse. In all conditions they indicated their responses by clicking with the mouse on a graphical user interface.
Speech materials were taken from the Coordinate Response Measure corpus (Bolia et al. 2000), which consists of sentences of the form “Ready <call sign>, go to <color> <number> now.” Only sentences spoken by the four male talkers in the corpus were used. For monaural stimuli, one sentence with the call sign “Charlie” was selected at random. For dichotic stimuli, two sentences with call signs “Charlie” and “Baron” were selected randomly with the constraints that each was spoken by a different talker and that the colors and numbers in the two utterances differed. No attempt was made to time-align the keywords across pairs of sentences, although there is considerable overlap in the sentences by design.
The sentence(s) were presented at a level that was fixed for each listener (see below) and were presented either in quiet or with speech-shaped noise added. Speech-shaped noise was created by filtering randomly generated broadband noises with the average frequency spectrum of the set of sentences used in the experiment. For dichotic stimuli, the noise was independent in the two ears but equal in level. In blocks where noise was added, the level of the noise was selected randomly from trial to trial from a set of four levels chosen separately for each listener group. The resulting signal-to-noise ratios (SNRs) were -12, -9, -6, -3 dB and -9, -6, -3, 0 dB for the NH and HL groups, respectively. The choice of these ranges was based on pilot experiments conducted on the first listener from each group.
C. Presentation level
In an attempt to compensate for differences in the audibility of speech signals between the two listener groups, presentation levels were set to a fixed sensation level chosen separately for each listener. Levels were set by measuring quiet sentence identification thresholds for each listener and presenting the speech stimuli at a fixed level above this threshold. Note that while this approach equates overall sensation level, it does not ensure an equal sensation level across frequencies. An adaptive procedure consisting of 20 trials was used to measure thresholds. On each trial, a randomly drawn sentence was presented to one ear only. Listeners identified the color and number of each sentence, and were scored as correct only if both words were correct. The presentation level was varied using a one-up, one-down adaptive rule that tracks the 50% correct point on the psychometric function. The initial step-size was 4 dB; after three reversals the step-size was decreased to 2 dB. The threshold calculation was based on all reversals after the first three (or four, to give an even number of reversals). One threshold was collected for each ear and then these two thresholds were averaged. Where possible, stimuli were presented at 45 dB SL. This was possible in only four of the nine HL listeners. The remaining five HL listeners found this level uncomfortable; for these listeners the level was reduced to a comfortable level (the lowest final level was 35 dB SL). The absolute level of the speech stimuli delivered by the headphones ranged between 80 and 103 dB SPL (as measured in a KEMAR artificial head) in the HL group.
For five of the nine NH listeners, thresholds were collected in the same way and the speech stimuli were presented at 45 dB SL (absolute levels of 56–74 dB SPL). For the other four listeners in this group, a fixed level of 65 dB SPL was used1.
D. Procedures
Listeners performed three tasks, presented in different blocks of trials. In “control” trials, only one message (containing the call sign “Charlie”) was presented to one ear and listeners reported the color and number keywords of this message. The stimulated ear was chosen randomly on each trial, with each ear being chosen an equal number of times over the course of a block. In the “single-task” trials, two messages were presented (one to each ear) and listeners were asked to report the color and number keywords from the message containing the call sign “Charlie.” The ear receiving the “Charlie” message was randomly chosen on each trial and the other ear received a message containing the call sign “Baron.” The assignment of call-signs to ears was balanced such that each ear received the “Charlie” message 50% of the time. In “dual task” trials, dichotic messages (statistically identical to those presented in single-task trials) were presented and listeners were asked to report the color and number keywords from the “Charlie” message followed by the color and number keywords from the “Baron” message. These instructions encouraged listeners to prioritize the processing of the “Charlie” message over that of the “Baron” message.
Trials were organized into blocks of 80, with the task fixed within a block. Noise was added randomly on a trial-by-trial basis such that each block contained 20 repetitions at each of the four predefined SNRs (see above). One block of each task (in the order control, single, dual) comprised a session and took approximately 30 minutes. Four sessions were completed by each listener.
In addition to being tested over the range of SNRs described above, the nine HL listeners and five of the NH listeners were also tested in quiet. One 40-trial block of each task in quiet (in the order control, single, dual) was completed before the main experimental sessions.
III. RESULTS
A. Mean performance as a function of SNR
Mean performance across listeners in each listener group is plotted in Fig. 2 as a function of SNR for the control task, the single task, and for the messages reported first (M1) and second (M2) in the dual task. Note that the range of SNRs tested was different for the two listener groups. Scores are averaged across trials in which the “Charlie” message was presented to the left ear and trials in which it was presented to the right ear. A small effect of ear-of-presentation was observed, consistent with previous reports of “right-ear dominance.” Specifically, when the “Charlie” message was presented to the right ear, performance was slightly better in the single task and for M1 in the dual task compared to when “Charlie” was presented to the left ear. Similarly, performance for M2 in the dual task was slightly better when “Baron” was presented to the right ear. This effect of ear-of-presentation was not statistically significant, and was much smaller than the effect of report order, as expected based on previous studies showing that an imposed response order overrides ear dominance effects (Wilson et al. 1968).
Figure 2.
Performance for listeners with normal hearing (NH: left panel) and hearing loss (HL: right panel) as a function of SNR. The different lines in each panel show across-subject mean scores in the control task (diamonds), the single task (squares), and the dual task for M1 (circles) and M2 (triangles). For the NH group, the data point for the quiet condition represents the mean of only five of the nine listeners. Error bars indicate standard errors of the mean.
In the control (monaural) task, performance varied with SNR from 70% to 96% in the NH group (diamonds, left panel) and from 69% to 96% in the HL group (diamonds, right panel). In quiet, performance on the control task was near perfect in both groups (100% and 99%). In the single task, where there was an irrelevant sentence present in the ear opposite the target, scores in both listener groups dropped (squares), indicating that presenting a message in the unattended ear interfered with performance. The magnitude of this interference was 11 percentage points on average in the NH group when noise was present (5 percentage points in quiet), and 16 percentage points on average in the HL group when noise was present (10 percentage points in quiet). In the dual task, performance for M1 (circles) was consistently worse than single-task performance, dropping by 12 percentage points on average in the NH group when noise was present (19 percentage points in quiet) and 10 percentage points on average in the HL group when noise was present (6 percentage points in quiet). Performance for M2 (triangles) was consistently worse than performance on M1, by 13 percentage points on average in the NH group when noise was present (4 percentage points in quiet) and by 16 percentage points on average in the HL group (15 percentage points in quiet).
Separate two-way ANOVAs were conducted on the arcsin-transformed data for the two listener groups with factors of task/message (control, single task, dual task M1, dual task M2) and SNR2. These ANOVAs revealed significant main effects of task/message [NH: F(3,24)=53.4, p<0.001; HL: F(3,24)=144.7, p<0.001] and SNR [NH: F(3,24)=98.7, p<0.001; HL: F(3,24)=177.9, p<0.001]. Post hoc comparisons with a Bonferroni correction indicated that all task/message conditions were significantly different from one another for both NH and HI groups, Moreover, all SNRs differed from one other except for the -9 and -6 dB SNRs in the NH group. The interaction between task/message and SNR was significant for the HL group [F(9,72)=2.0, p<0.05] but not the NH group [F(9,72)=2.0, p=0.06].
B. Error patterns
Examination of responses revealed that many errors in the single task were due to reporting one or two keywords from the irrelevant ear, and many errors in the dual task were due to reporting the messages in the wrong order. A breakdown of the errors is provided in Fig. 3.
Figure 3.
Frequency of confusion errors (top row) and random errors (bottom row) for listeners with normal hearing (NH: left column) and hearing loss (HL: right column). The different lines in each panel show across-subject mean scores in the control task (diamonds), the single task (squares), and the dual task for M1 (circles) and M2 (triangles).
The top row of Fig. 3 shows the rate of “confusion” errors, in which there was confusion between the ears for one or both keywords. For example, if the required keywords were “red one” and the keywords from the other message were “blue two,” then responses of “blue two,” “blue one,” and “red two” would all constitute confusion errors. Confusion errors occurred on approximately 20% of the trials for both groups of listeners. This extent of across-ear interference was surprising given that many previous studies have found no measurable interference in dichotic listening tasks (Cherry 1953; Drullman and Bronkhorst 2000). These errors may have been more prevalent in the current study either because of the highly confusable nature of the speech materials in the corpus (c.f. Brungart 2001) or because the target ear varied from trial to trial in an unpredictable way and had to be selected on the basis of the call sign (Gallun et al. 2007). The added noise also influenced confusion errors, as shown by the tendency of these errors to decrease with increasing SNR. The bottom row of Fig. 3 shows the rate of “random” errors, in which one or both of the reported keywords were not from either message. In the above example, a response such as “red three” or “green three” would fall in this category. Random errors dropped off with increasing SNR, and occurred with similar frequency in the control and single tasks, as well as for M1 in the dual task. In the case of M2, random errors were far more frequent and more sensitive to SNR (compare slopes in bottom panels of Fig. 3).
C. Comparison of dual-task errors for NH and HL listeners
For those listeners who completed the initial “quiet” session (rightmost points in Fig. 3), two separate ANOVAs were conducted on the arcsin-transformed error rates in quiet in the dual task (with message, M1 or M2, as a within-subjects factor and listener group as a between-subjects factor). For confusion errors, the analysis revealed no significant effect of message [F(1,12)=0.1, p=0.7] or listener group [F(1,12)=0.02, p=0.9], and no interaction [F(1,12)=0.6, p=0.5]. For random errors, there was a significant effect of message [F(1,12)=6.4, p<0.05], but no significant effect of listener group (F(1,12)=0.1, p=0.8] and no significant interaction [F(1,12)=0.5, p=0.5]. Thus, there is no evidence that the listener groups differed with respect to their dual-task performance in quiet.
To compare performance for the two listener groups in the dual task when noise was present, Fig. 4 (left column) shows mean error rates replotted on the same axis for both groups (NH: filled symbols, HL: open symbols) and both messages (M1: circles, M2: triangles). The top left panel shows confusion errors and the bottom left panel shows random errors.
Figure 4.
Frequency of confusion errors (top row) and random errors (bottom row) in the dual task for listeners with normal hearing (NH: filled symbols) and hearing loss (HL: open symbols). The different lines in each panel show across-subject mean scores for M1 (circles) and M2 (triangles). The left column shows data replotted from Fig. 3 on the same SNR axis (error bars have been omitted to aid visualization but statistical comparisons are given in the text). The right column shows the same data but with the HL group shifted to the left by 5 dB.
For both groups, confusion errors were more common and more sensitive to SNR for M1 than M2. Confusion errors for the HL group were higher than the NH group for M1, but the groups were similar for M2. A repeated-measures ANOVA was conducted on the arcsin-transformed confusion error rates over the common range of SNRs, with message and SNR as within-subjects factors and listener group as a between-subjects factor. The analysis revealed significant main effects of message [F(1,16)=26.6, p<0.001] and SNR [F(2,32)=5.1, p<0.05], but no main effect of listener group (F(1,16)=1.8, p=0.2]. The two-way interaction between message and SNR was significant [F(2,32)=8.5, p<0.005], consistent with the observation that M1 confusion errors were more sensitive to the SNR than M2 confusion errors. The interaction between message and listener group was also significant [F(1,16)=5.2, p<0.05], supporting the observation that the HL group made more M1 errors than the NH group. No other interactions were significant.
Random errors in M2 for both groups were far more common than those in M1, and more sensitive to SNR. Random errors were more prevalent in the HL group than in the NH group, especially for M2. A repeated-measures ANOVA was conducted on the arcsin-transformed random error rates over the common range of SNRs revealed significant main effects of message [F(1,16)=148.3, p<0.001], SNR [F(2,32)=134.0, p<0.001], and listener group (F(1,16)=13.9, p<0.005]. The two-way interaction between message and SNR was significant [F(2,32)=9.1, p<0.005], indicating that random errors in M2 were more sensitive to the SNR than random errors in M1. The two-way interaction between message and listener group was also significant [F(1,16)=5.9, p<0.05], consistent with the observation that the HL group showed a greater deficit for M2. No other interactions were significant.
Listeners with hearing loss often perform as well as listeners with normal hearing on speech intelligibility tasks when given a more favorable SNR. As performance was examined at a range of SNRs in the current experiment, it was possible to examine whether a simple SNR increase would also explain differences in performance between groups on a dual task. A calculation was done to determine the shift (in dB) that best aligned the error functions for the two groups on the single task (minimized the absolute error between groups across the two error types3) and this shift was then applied to the error patterns for the dual task. The right column of Fig. 4 shows error data identical to that in the left column but with the HL data shifted along the SNR axis by this optimal shift (5 dB).
The shifted error functions line up well between the two groups. In fact the 5-dB shift, which minimized the mean absolute error between groups for the single task (mean absolute error across the two error types of 1.3 percentage points) also minimized the error between groups for the dual task (mean absolute error across the two error types and two messages of 1.6 percentage points). Note that this shift not only aligned error functions that were fairly similar between groups (all confusion errors, and M1 random errors) but also the error functions that differed greatly between groups (M2 random errors). The fact that M2 random errors are more sensitive to SNR means that a given improvement in SNR gives rise to a larger reduction in errors. Put differently, the effective drop in SNR caused by hearing loss has a larger impact on M2 random errors than on M1 random errors because M2 error functions are steeper.
IV. DISCUSSION
A. Performance for M1 and M2 in the dual task
In our dual task, performance was poorer for each message than for the one message reported in the single task. For M1, the difference was relatively small and was due to both an increase in confusion errors (where having to report both messages increased the chances of subjects interchanging the keywords) and an increase in random errors (which may be a consequence of processing load). For M2, the deficit relative to the single task was far greater due to a much larger occurrence of random errors.
More than 50 years ago, Broadbent (1954) proposed that simultaneous inputs to the auditory system are processed serially to some extent. He presented two sequences of digits simultaneously to the two ears and observed that, although listeners could recall all digits, responses were always made to one ear before the other. Broadbent postulated that simultaneous sensory inputs are stored temporarily via immediate auditory memory and then processed serially by a limited capacity mechanism (1957; 1958, chapter 9; see also Lachter et al. 2004). A consequence of such a scheme is that the secondary message in the pair must be stored whilst the primary message is processed. Our results, showing large differences in performance for the message reported first and second, are consistent with this idea (see also Ihlefeld and Shinn-Cunningham 2008).
One difficulty with the dual-response design, however, is that the responses themselves must be made sequentially. Specifically, it is possible that the poorer performance on M2 is related to the fact that it must be retained in memory longer than M1 during the response interval (Sperling 1960). However, the response method employed (one mouse click per message on a grid of color-coded/numbered buttons) minimized this time delay; thus we believe that the performance differences observed for M1 and M2 are primarily due to differences in the order in which they are processed (and/or by the mechanism underlying their processing). An alternative approach would have been to use a “partial report” procedure in which listeners are asked after stimulus presentation to report back just one of two messages (Sperling 1960; Darwin et al. 1972; Gallun et al. 2007), but this design has the disadvantage that performance on both M1 and M2 could not be evaluated within a single trial. Further experiments will be required to confirm that differential performance for M1 and M2 truly reflects different processing mechanisms.
B. The effect of noise on the processing of simultaneous messages
In both normal-hearing and hearing-impaired listeners, we found that increasing the noise level affected performance for M1 in the dual task in nearly the same way that it affected performance in the single task. In contrast, the ability to report M2 decreased more dramatically with increasing noise level due to a sharp rise in random errors. These results support the conclusion that the processing of simultaneous messages interacts with the quality of the inputs.
In the conceptual model described above, in which simultaneous inputs are processed serially, the input that is processed second is held in the form of a raw sensory representation that is volatile and degrades with time (Broadbent 1957; Brown 1958; Durlach and Braida 1969). If this were the case, it would explain why performance for M2 is particularly sensitive to the integrity of the acoustic input. A degraded input will degrade even further in this store, and may not even be useful by the time it is fully processed. In essence, there may be a trade-off between SNR and the time interval over which a sensory trace must be maintained. Note that the effect of noise on M2 was almost exclusively due to an increase in random errors; confusion errors were quite constant as a function of SNR (Figs. 3 and 4). This supports the idea that sensory degradation, not increased confusion between the streams, is responsible for the dramatic effect of noise on recall of M2. In related recent work, it has been suggested that reduced absolute signal level (in the absence of noise) can also disrupt echoic persistence and hence performance on a secondary task (Baldwin and Struckman-Johnson 2002; Baldwin 2007).
An alternative explanation of this result is that the increased difficulty of processing M1 in trials with a low SNR effectively drained a limited pool of processing resources, leaving fewer resources for processing M2. This rationale has been used previously to explain the effect of noise on the ability to store part of a single attended message for later recall (Rabbitt 1968; Pichora-Fuller et al. 1995).
C. The effect of hearing loss on the processing of simultaneous messages
The primary goal of this study was to assess the effect of sensorineural hearing loss on the processing of simultaneous messages. Given previous reports that situations involving divided or rapidly switching attention are difficult for listeners with hearing loss, we expected to find a larger deficit in responses to a secondary message than to a primary message (relative to listeners with normal hearing). In order to focus on the task of processing simultaneous messages without confounding factors related to peripheral resolvability of the messages, we used dichotic presentation and an increased overall presentation level in the hearing-impaired group.
Dual-task error rates in the quiet condition did not differ significantly between the two groups of listeners. Thus, it seems that hearing impairment in the absence of noise does not necessarily mimic the effects of added noise in listeners with normal hearing. Note that this may be a ceiling effect, as error rates were quite low in the quiet condition for this task (particularly random errors, for which a difference between groups might be expected). In a previous study that used a more demanding dichotic digits task (with recall of three digits per ear required), a measurable deficit in quiet was observed for listeners with milder hearing impairments than those of our listeners (Neijenhuis et al. 2004).
In the presence of added noise, listeners with hearing loss performed worse than listeners with normal hearing at the same SNR. For M1, there was a small increase in confusion errors and a small elevation of random errors. For M2, there was in fact no increase in confusion errors but there was a very large elevation in random errors. This indicates that the ability to extract information from a second, simultaneous message is particularly poor in listeners with hearing loss when compared to listeners with normal hearing at the same SNR. However, when the two groups of listeners were compared at different SNRs (higher in the HL group by 5 dB), error rates for both M1 and M2 were similar in the two groups. In other words, our results suggest that in a simultaneous listening task, just as in many selective listening tasks, listeners with sensorineural hearing loss perform like normally-hearing listeners given a poorer SNR. It appears that the addition of noise disrupts the processing of multiple sources and that the presence of a hearing impairment exacerbates this effect.
A key feature of the design of the current study was that listeners were asked to give multiple responses to simultaneous messages, allowing us to compare performance on the two messages. This approach has not been widely adopted in listeners with hearing loss. While several studies used a double-vowel paradigm, in which listeners are asked to identify both of a pair of simultaneously presented vowels (Summers and Leek 1998; Arehart et al. 2005; Rossi-Katz and Arehart 2005), very few have used sentence-length speech in which memory can be assumed to play a more significant role (but see Mackersie et al. 2001; Neijenhuis et al. 2004). Mackersie and colleagues (2001) employed a “simultaneous sentence test,” in which listeners reported back two simultaneous messages (spoken by one male and one female voice, presented monaurally). Although the authors were primarily concerned with performance on one message (the one reported first; they only required listeners to report the second message to make the task more difficult), their results show that fewer keywords were reported correctly for the secondary message than for the primary message. This finding is consistent with the results from our task in which two messages spoken by two male talkers were presented to separate ears. In this previous study, listeners with hearing loss were able to recall an average of only 10% of the keywords from the second message. In contrast, while we find that hearing-impaired listeners are worse at reporting M2 than M1, performance is still around 25%, even at the lowest SNRs tested, and 40% if confusion errors are excluded. Part of this difference may be that the memory load in our task was relatively low: there were only two key words to be recalled from each message, and the messages came from a small, closed response set. However, in addition, the listeners with hearing loss in the study of Mackersie et al. were older than their normal-hearing counterparts (mean ages 73 and 27 years respectively); thus, it is possible that performance of the HL listeners was lower due to age-related effects. While aging may well contribute to poor performance in divided listening tasks (see also Humes et al. 2006; Singh et al. 2008), the results of the current study suggest that hearing loss in otherwise-healthy, young adults interferes with the ability to recall a second message during divided listening in noise.
An effect of hearing loss on recall of speech stimuli has been reported previously (Rabbitt 1990; Pichora-Fuller et al. 1995; McCoy et al. 2005), often in tasks involving listening to and responding to a sentence or a sequence of numbers while at the same time storing a word from the sequence for later recall. Hearing loss appears to impair performance on the recall task, an effect that has been explained in terms of an “effort hypothesis.” According to this hypothesis, hearing loss makes the immediate speech task more demanding, leaving fewer processing resources for storing the to-be-recalled items. This hypothesis is also supported by studies that have used a secondary task that is non-auditory and thus does not depend directly on the quality of the auditory stimuli. For example, Rakerd et al. (1996) showed that young hearing-impaired listeners perform more poorly than young normal-hearing listeners on a secondary task involving memorization of visually-presented digits, when the primary task is to comprehend an ongoing speech passage.
For the task explored in the current study, namely the immediate recall of simultaneous messages, we propose that hearing loss may also have a direct effect on the processing of M2 by degrading its spectrotemporal representation in the auditory system. In other words, hearing loss may compromise a listener’s ability to process simultaneous messages in a similar way to added noise, by degrading the sensory trace that is used for the processing of a source outside the primary focus of attention (Shinn-Cunningham and Best 2008).
D. Implications of the results
We find that listeners with hearing loss show a larger deficit (relative to their normal-hearing counterparts) on the processing of a secondary message than on the processing of a primary message when listening at a given SNR. However, the performance of these listeners could be equated with that of listeners with normal hearing simply by improving the overall SNR. This suggests that technology, behavioral changes, or environmental modifications focused on improving the SNR should be very effective in aiding communication in complex environments for hearing-impaired listeners.
Our finding that improving the SNR has a larger impact on the processing of secondary talkers than on the processing of a primary talker may prove to be important in understanding the full extent of benefits available from bilateral hearing aids and bilateral cochlear implants. Until now, bilateral systems have been compared to unilateral configurations using a variety of selective listening and sound localization tasks. In these measures, the size of the “bilateral benefits” reported have varied substantially across listeners and studies (Brown and Balkany 2007; Ching et al. 2007; Boymans et al. 2008; Marrone et al. 2008). If bilateral systems were tested using a listening task in which listeners were required to extract information from two simultaneous messages, even larger benefits might be observed (see also Noble and Gatehouse 2006). Although the case tested in the current study represented an extreme example (in which one message was delivered to each ear), there are many natural situations in which different sounds in the environment are spatially distributed such that they have different “better ears.” In such cases, while good reception of a single sound requires a good SNR at one ear, the successful reception of multiple sounds may require a good SNR at both ears.
As a final note, it is possible that the task used in this study may underestimate the difficulty of simultaneous processing for listeners with hearing loss in more realistic situations. First, presenting the two simultaneous messages to separate ears alleviated the known difficulties that hearing-impaired listeners have with segregating simultaneous voices; this is evidenced by the fact that the drop in performance from the monaural control task to the dichotic single task was similar in the two listener groups. Second, the trial-based structure of the speech task meant that stimuli were always followed by a silent period in which listeners could make optimal use of temporarily “stored” sensory information. In more realistic situations, where conversations flow rapidly and continuously, such a catch-up strategy would be impossible and this may exacerbate the effects of a degraded sensory representation. Finally, the speech materials used in the current study (where each sentence had only two keywords from a closed set) gave rise to primary and secondary tasks with modest cognitive demands. The effects we observed might be greater/exaggerated for tasks involving longer, open-set sentences; in such conditions, the primary speech task would require more sustained attention and the memory demands of the secondary task would be increased. This might be expected to increase overall difficulty, particularly for older listeners with hearing loss, as there appears to be an interaction between the effects of memory load and age (Gordon-Salant and Fitzgibbons 1997; Wingfield et al. 2005).
Acknowledgments
This work was supported by grants from NIDCD and AFOSR. Virginia Best was also supported by a University of Sydney Postdoctoral Research Fellowship. Frederick Gallun was also supported by the Department of Veterans Affairs, Veterans Health Administration, Rehabilitation Research and Development Service. The authors would like to thank Desiree Balderrama for her help with data collection and Antje Ihlefeld for reading an earlier version of the manuscript. Portions of this work were presented at the International Congress on Acoustics in Madrid, Spain (September 2007) and the International Hearing Aid Research Conference in California, USA (August 2008).
Footnotes
Four of the NH subjects completed the study before the HL listeners began and the threshold measurement was introduced.
The data collected in quiet were not included in these ANOVAs because not all subjects in the NH group completed the quiet condition.
The error functions were first linearly interpolated to give a resolution of 0.5 dB. Then for each pair of error functions, the absolute error was calculated as the HL function was shifted to the left in 0.5 dB steps over the range 0 to 10 dB. The chosen shift value gave the smallest mean absolute error across the two error types in the single task.
References
- Arehart KH, Rossi-Katz J, Swensson-Prutsman J. Double-vowel perception in listeners with cochlear hearing loss: differences in fundamental frequency, ear of presentation, and relative amplitude. J Speech Lang Hear Res. 2005;48:236–252. doi: 10.1044/1092-4388(2005/017). [DOI] [PubMed] [Google Scholar]
- Baldwin CL, Struckman-Johnson D. Impact of speech presentation level on cognitive task performance: implications for auditory display design. Ergonomics. 2002;45:61–74. doi: 10.1080/00140130110115336. [DOI] [PubMed] [Google Scholar]
- Baldwin CL. Cognitive implications of facilitating echoic persistence. Memory and Cognition. 2007;35:774–780. doi: 10.3758/bf03193314. [DOI] [PubMed] [Google Scholar]
- Best V, Gallun FJ, Ihlefeld A, et al. The influence of spatial separation on divided listening. J Acoust Soc Am. 2006;120:1506–1516. doi: 10.1121/1.2234849. [DOI] [PubMed] [Google Scholar]
- Bolia RS, Nelson WT, Ericson MA, et al. A speech corpus for multitalker communications research. J Acoust Soc Am. 2000;107:1065–1066. doi: 10.1121/1.428288. [DOI] [PubMed] [Google Scholar]
- Boymans M, Goverts ST, Kramer SE, et al. A prospective multi-centre study of the benefits of bilateral hearing aids. Ear and Hearing. 2008;29:930–941. doi: 10.1097/aud.0b013e31818713a8. [DOI] [PubMed] [Google Scholar]
- Broadbent DE. The role of auditory localization in attention and memory span. J Exp Psychol. 1954;47:191–196. doi: 10.1037/h0054182. [DOI] [PubMed] [Google Scholar]
- Broadbent DE. Immediate memory and simultaneous stimuli. Q J Exp Psychol. 1957;9:1–11. [Google Scholar]
- Broadbent DE. Perception and Communication. London: Pergamon Press; 1958. [Google Scholar]
- Brown J. Some tests of the decay theory of immediate memory. Q J Exp Psychol. 1958;10:12–21. [Google Scholar]
- Brown K, Balkany T. Benefits of bilateral cochlear implantation: a review. Curr Opin Otolaryngol Head Neck Surg. 2007;15:315–318. doi: 10.1097/MOO.0b013e3282ef3d3e. [DOI] [PubMed] [Google Scholar]
- Brungart DS. Informational and energetic masking effects in the perception of two simultaneous talkers. J Acoust Soc Am. 2001;109:1101–1109. doi: 10.1121/1.1345696. [DOI] [PubMed] [Google Scholar]
- Bryden MP. Ear preference in auditory perception. J Exp Psych. 1963;65:103–105. doi: 10.1037/h0042773. [DOI] [PubMed] [Google Scholar]
- Cherry EC. Some experiments on the recognition of speech, with one and with two ears. J Acoust Soc Am. 1953;25:975–979. [Google Scholar]
- Ching T, van Wanrooy E, Dillon H. Binaural-bimodal fitting or bilateral implantation for managing severe to profound deafness: a review. Trends Amplif. 2007;11:161–192. doi: 10.1177/1084713807304357. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Conway ARA, Cowan N, Bunting MF. The cocktail party phenomenon revisited: the importance of working memory capacity. Psychonomic Bull Rev. 2001;8:331–335. doi: 10.3758/bf03196169. [DOI] [PubMed] [Google Scholar]
- Darwin CJ, Turvey MT, Crowder RG. An auditory analogue of the Sperling partial report procedure: Evidence for brief auditory storage. Cog Psych. 1972;3:255–267. [Google Scholar]
- Drullman R, Bronkhorst AW. Multichannel speech intelligibility and talker recognition using monaural, binaural, and three-dimensional auditory presentation. J Acoust Soc Am. 2000;107:2224–2235. doi: 10.1121/1.428503. [DOI] [PubMed] [Google Scholar]
- Duquesnoy AJ. Effect of a single interfering noise or speech source upon the binaural sentence intelligibility of aged persons. J Acoust Soc Am. 1983;74:739–743. doi: 10.1121/1.389859. [DOI] [PubMed] [Google Scholar]
- Durlach NI, Braida LD. Intensity perception. I Preliminary theory of intensity perception. J Acoust Soc Am. 1969;2:372–383. doi: 10.1121/1.1911699. [DOI] [PubMed] [Google Scholar]
- Festen JM, Plomp R. Effects of fluctuating noise and interfering speech on the speech-reception threshold for impaired and normal hearing. J Acoust Soc Am. 1990;88:1725–1736. doi: 10.1121/1.400247. [DOI] [PubMed] [Google Scholar]
- Fifer RC, Jerger JF, Berlin CI, et al. Development of a dichotic sentence identification test for hearing-impaired adults. Ear and Hearing. 1983;4:300–305. doi: 10.1097/00003446-198311000-00007. [DOI] [PubMed] [Google Scholar]
- Gallun FJ, Mason CR, Kidd G., Jr Task-dependent costs in processing two simultaneous auditory stimuli. Percept Psychophys. 2007;69:757–771. doi: 10.3758/bf03193777. [DOI] [PubMed] [Google Scholar]
- Gatehouse S, Noble W. The Speech, Spatial and Qualities of Hearing Scale (SSQ) Int J Audiol. 2004;43:85–99. doi: 10.1080/14992020400050014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gordon-Salant S, Fitzgibbons PJ. Selected cognitive factors and speech recognition performance among young and elderly listeners. J Speech Lang Hear Res. 1997;40:423–431. doi: 10.1044/jslhr.4002.423. [DOI] [PubMed] [Google Scholar]
- Hugdahl K. Dichotic listening in the study of auditory laterality. In: Hugdahl K, Davidson RJ, editors. The Asymmetrical Brain. Cambridge: MIT Press; 2003. [Google Scholar]
- Humes LE, Lee JH, Coughlin MP. Auditory measures of selective and divided attention in young and older adults using single-talker competition. J Acoust Soc Am. 2006;120:2926–2937. doi: 10.1121/1.2354070. [DOI] [PubMed] [Google Scholar]
- Ihlefeld A, Shinn-Cunningham BG. Spatial release from energetic and informational masking in a divided speech identification task. J Acoust Soc Am. 2008;123:4380–4392. doi: 10.1121/1.2904825. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Jerger J, Martin J. Dichotic listening tests in the assessment of auditory processing disorders. Audiological Medicine. 2006;4:25–34. [Google Scholar]
- Kimura D. Cerebral dominance and the perception of verbal stimuli. Can J Psychol. 1961;15:166–171. [Google Scholar]
- Lachter J, Forster KI, Ruthruff E. Forty-five years after Broadbent (1958): Still no identification without attention. Psychol Rev. 2004;111:880–913. doi: 10.1037/0033-295X.111.4.880. [DOI] [PubMed] [Google Scholar]
- Mackersie CL, Prida TL, Stiles D. The role of sequential stream segregation and frequency selectivity in the perception of simultaneous sentences by listeners with sensorineural hearing loss. J Speech Lang Hear Res. 2001;44:19–28. doi: 10.1044/1092-4388(2001/002). [DOI] [PubMed] [Google Scholar]
- Marrone N, Mason CR, Kidd G., Jr Evaluating the benefit of hearing aids in solving the cocktail party problem. Trends in Amplification. 2008;12:300–315. doi: 10.1177/1084713808325880. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McCoy SL, Tun PA, Cox LC, et al. Hearing loss and perceptual effort: Downstream effects on older adults’ memory for speech. Quart J Exp Psych. 2005;58A:22–33. doi: 10.1080/02724980443000151. [DOI] [PubMed] [Google Scholar]
- Moray N. Attention in dichotic listening: Affective cues and the influence of instructions. Q J Exp Psychol. 1959;11:55–60. [Google Scholar]
- Neijenhuis K, Tschur H, Snik A. The effect of mild hearing impairment on auditory processing tests. J Am Acad Audiol. 2004;15:6–16. doi: 10.3766/jaaa.15.1.3. [DOI] [PubMed] [Google Scholar]
- Noble W, Gatehouse S. Effects of bilateral versus unilateral hearing aid fitting on abilities measured by the Speech, Spatial, and Qualities of Hearing Scale (SSQ) Int J Audiol. 2006;45:172–181. doi: 10.1080/14992020500376933. [DOI] [PubMed] [Google Scholar]
- Pichora-Fuller MK, Schneider BA, Daneman M. How young and old adults listen to and remember speech in noise. J Acoust Soc Am. 1995;97:593–608. doi: 10.1121/1.412282. [DOI] [PubMed] [Google Scholar]
- Rabbitt PM. Channel capacity, intelligibility and immediate memory. Quart J Exp Psych. 1968;20:241–248. doi: 10.1080/14640746808400158. [DOI] [PubMed] [Google Scholar]
- Rabbitt PM. Mild hearing loss can cause apparent memory failures which increase with age and reduce with IQ. Acta Oto-laryngologica. 1990;476:167–176. doi: 10.3109/00016489109127274. [DOI] [PubMed] [Google Scholar]
- Rakerd B, Seitz PF, Whearty M. Assessing the cognitive demands of speech listening for people with hearing losses. Ear and Hearing. 1996;17:97–106. doi: 10.1097/00003446-199604000-00002. [DOI] [PubMed] [Google Scholar]
- Rivenez M, Darwin CJ, Guillaume A. Processing unattended speech. J Acoust Soc Am. 2006;119:4027–4040. doi: 10.1121/1.2190162. [DOI] [PubMed] [Google Scholar]
- Rossi-Katz JA, Arehart KH. Effects of cochlear hearing loss on perceptual grouping cues in competing-vowel perception. J Acoust Soc Am. 2005;118:2588–2598. doi: 10.1121/1.2031975. [DOI] [PubMed] [Google Scholar]
- Shinn-Cunningham BG, Best V. Selective attention and hearing impairment. Trends in Amplification. 2008;12:283–299. doi: 10.1177/1084713808325306. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Singh G, Pichora-Fuller MK, Schneider BA. Auditory spatial attention in conditions of real and simulated spatial separation by younger and older adults. J Acoust Soc Am. 2008:124. doi: 10.1121/1.2949399. [DOI] [PubMed] [Google Scholar]
- Sperling G. The information available in brief visual presentations. Psychological Monographs. 1960;74:1–29. [Google Scholar]
- Strouse A, Wilson RH, Brush N. Recognition of dichotic digits under pre-cued and post-cued response conditions in young and elderly listeners. Br J Audiol. 2000;34:141–151. doi: 10.3109/03005364000000124. [DOI] [PubMed] [Google Scholar]
- Summers V, Leek MR. F0 processing and the separation of competing speech signals by listeners with normal hearing and with hearing loss. J Speech Lang Hear Res. 1998;41:1294–1306. doi: 10.1044/jslhr.4106.1294. [DOI] [PubMed] [Google Scholar]
- Treisman AM. Shifting attention between the ears. Q J Exp Psychol. 1971;23:157–167. doi: 10.1080/14640747108400236. [DOI] [PubMed] [Google Scholar]
- Wilson RH, Dirks DD, Carterette EC. Effects of ear preference and order bias on the reception of verbal materials. J Speech Hear Res. 1968;11:509–522. doi: 10.1044/jshr.1103.509. [DOI] [PubMed] [Google Scholar]
- Wingfield A, Tun P, McCoy SL. Hearing loss in older adulthood. Curr Dir Psychol Sci. 2005;14:144–148. [Google Scholar]




