Highlights
-
•
We propose a highly automated procedure for extracting latencies and amplitudes of auditory brainstem responses (ABRs) based on curve registration through non-linear time warping.
-
•
We compare different registration conditions using an example ABR data set with a wide range of response latencies and signal-to-noise ratios.
-
•
We demonstrate that the best registration condition closely matched the performance of expert human observers.
Keywords: Supra-threshold auditory brainstem responses, Hidden hearing loss, Cochlear synaptopathy, Dynamic time warping, Continuous monotone registration
Abstract
Background and objectives
Animal results have suggested that auditory brainstem responses (ABRs) to transient sounds presented at supra-threshold levels may be useful for measuring hearing damage that is hidden to current audiometric tests. Evaluating such ABRs requires extracting the latencies and amplitudes of relevant deflections, or “waves”. Currently, this is mostly done by human observers manually picking the waves’ peaks and troughs in each individual response – a process that is both time-consuming and requiring of expert experience. Here, we propose a highly automated procedure for extracting individual ABR wave latencies and amplitudes based on the well-established methodology of non-linear curve registration.
Methods
First, the to-be-analysed individual ABRs are temporally aligned – either with one another or, if available, with a pre-existing template – by locally compressing or stretching their time axes with smooth and invertible time warping functions. Then, the individual latencies and amplitudes of relevant ABR waves are obtained by picking the latencies of the waves’ peaks and troughs on the common (aligned) time axis and combining these with the individual aligned responses and inverse time warping functions.
Results
Using an example ABR data set with a wide range of response latencies and signal-to-noise ratios (SNRs), we test different choices for fitting the time warping functions. We cross-validate the warping results using independent response replicates and compare automatically and manually extracted latencies and amplitudes for ABR waves I and V. Using a Bayesian approach, we show that, for the best registration condition, automatic and manual data were statistically similar.
Conclusions
Non-linear curve registration can be used to temporally align individual ABRs and extract their wave latencies and amplitudes in a way that closely matches results from manual picking.
1. Introduction
In humans, sound-evoked neuronal responses from subcortical, or brainstem, auditory structures can be measured non-invasively with electroencephalographic, or scalp electrodes. Currently, the clinical use of these “auditory brainstem responses” (ABRs) is mostly limited to hearing screening and objective estimation of audiometric thresholds. Animal evidence, however, has suggested that, in future, the clinical role of ABRs may expand to detecting hearing damage not thought to be associated with audiometric loss, such as damage to the cochlear synapses [1], [2], [3], [4], inner hair cells [5,6] or auditory nerve [7], which have been found to manifest in a reduction of auditory-nerve responses to supra-threshold sounds. This has prompted considerable efforts in exploring whether, or to what degree, audiometrically hidden hearing damage also occurs in humans [reviewed in 8,9], and developing suitable tests for its measurement and diagnosis [reviewed in 10]. Most of the tests considered so far have involved ABRs elicited by transient, supra-threshold sounds, such as loud clicks [reviewed in 11].
ABR-based measurement of audiometrically hidden hearing damage, such as synaptopathy, poses a fundamentally different analytical challenge than ABR-based estimation of audiometric thresholds. Whilst ABR-based threshold estimation requires to decide whether a measured response contains a ‘true’ ABR over and above the inherent noise, evaluation of supra-threshold ABRs requires characterization of the responses’ properties, such as wave latencies and amplitudes. Many of the previous studies of human synaptopathy have measured the amplitude of the earliest ABR wave, wave I [e.g., [12], [13], [14]], thought to arise from the auditory nerve, and some have also measured the amplitude of the prominent wave V [e.g., 15–17].
Currently, ABR wave latencies and amplitudes are mostly extracted through manual picking of the waves’ peaks and troughs by human observers. However, manual picking is time-consuming and requires expert experience, which is prohibitive when dealing with large data sets or performing a routine clinical test. This study was aimed at devising a procedure for extracting individual latencies and amplitudes of ABR waves that would require minimal human input. First, we review the strengths and limitations of relevant previous approaches. Based on this, we develop the rationale for the current approach, which involves the well-established methodology of non-linear curve registration [18,19]. Using an example ABR data set from 23 normal healthy subjects comprising 12 different experimental conditions, we examine a range of different registration conditions. For each condition, we compare the automatically picked latencies and amplitudes of waves I and V with corresponding manually picked values. Finally, we use a realistic set of simulated ABR data to explore the effect of signal-to-noise ratio (SNR).
2. Background
ABRs evoked by transient sounds, such as clicks or brief chirps, consist of a series of consecutive waves [20] arising from different levels of the subcortical auditory pathway [reviewed in 21]. Like other natural time series, ABRs exhibit both amplitude and time variability: features of individual subjects’ ABRs, such as wave peaks or troughs, vary in both amplitude and latency. Due to time variability, the amplitude of a given feature (e.g., wave peak) can only be meaningfully measured after first establishing the feature's latency in each individual response. Otherwise, amplitudes would be measured at different points within the feature's shape and would thus be confounded by variability in its latency. A human observer will pick an ABR wave's vertex-positive peak and the subsequent trough, and take the peak latency as a measure of the wave's latency, and the peak-to-trough amplitude difference as a measure of the wave's amplitude. As example, Fig. 1A shows manual picks of waves I and V in 22 individual ABRs to a broadband chirp stimulus (detailed information about the current data set is given in Section 3.2.1).
To obviate the need for manual picking, Elberling [22] proposed to cross-correlate individual ABRs with a “response template” (Fig. 1B, left panel) to obtain individual measures of overall response latency and response-to-template similarity. The template was generated by averaging the individual responses after time-shifting them to align their wave-V peaks (typically the largest peaks in the responses; see Fig. 1A) and normalizing them by their root-mean-square amplitudes. The template may be considered an estimate of the individual responses’ common underlying structure – a kind of “structural average” [23] response, which, unlike the conventional “cross-sectional” average (i.e., average of the unaligned responses), is less affected by the smoothing effects of time variability. However, as the approach only aligns the wave-V peaks, smoothing could continue to affect other waves, if their interpeak latencies relative to wave V vary across responses. Using the current data set as example, the right panel in Fig. 1B shows that, even in normal healthy subjects of similar age, the interpeak latency between waves I and V can vary significantly across subjects. Even greater variability is known to arise as a result of maturation and ageing [24] and in certain neurological conditions [25].
Two different approaches have been proposed to address this problem. The computationally simpler, but also less powerful approach is the “wave template” approach [[26], [27], [28], [29]], which fits separate templates to individual ABR waves (see Fig. 1C, for a simplified example). Unlike the response templates of Elberling [22], the wave templates are not derived from the measured responses, but implemented as synthetic functions, representing the waves’ idealized shapes. To fit them, each template is allowed to shift and scale in time and in amplitude. This assumes that ABR waves are “shape-invariant”, that is, that they retain the same basic shape across individuals and/or conditions, varying only in size and/or scale. This assumption, however, is unlikely to be generally valid, given that ABR waves represent a mixture of multiple activities generated by different sources [20,[30], [31], [32]], which may vary both in strength and in timing, and thus represents a limitation of the wave template approach.
In contrast, the dynamic time warping (DTW) approach proposed by Picton, Hunt, Mowrey, Rodriguez and Maru [33] makes no assumptions about response shapes – like the response template approach of Elberling [22], it is completely non-parametric. The main difference is that, rather than using linear time shifts to align individual responses, the DTW approach uses non-linear time transformations, locally stretching or compressing portions of the responses to align common features, such as waves (Fig. 1D). Even though the DTW approach was not specifically designed for extracting wave latencies and amplitudes, its basic idea of non-linearly aligning individual responses to form a structural average would seem to hold great promise for achieving this purpose. However, in its original implementation, the approach used the discrete form of time warping, developed previously by Sakoe and Chiba [34] for measuring similarity between speech tokens, rather than the continuous form developed only later [18,19]. As a result, the approach suffered from unnecessary conceptual and computational complexity.
Neither the template-based approaches nor the DTW approach have so far been applied in ABR studies of human synaptopathy. Instead, several of these studies have used a much simpler approach, picking ABR peaks and troughs by finding the corresponding zero-crossings in the temporal derivative of the response waveform [16,17,35,36]. Various derivative-based ABR analysis procedures have been proposed previously [[37], [38], [39], [40], [41], [42], [43]], and all were particularly focussed on separating the ‘true’ ABR from the inherent noise, as noise is disproportionately amplified by differentiation. In contrast, the human synaptopathy studies have dealt with the noise problem by limiting the search for derivative zero-crossings to predefined time windows around relevant wave peaks and troughs in the average response across individuals (Fig. 1E, left panel). Individual responses that did not contain a derivative zero-crossing within a given search window were assumed not to exhibit the corresponding wave and thus excluded from further analysis. This, however, is not necessarily true. Rather, at least a proportion of the excluded responses will have exhibited the relevant wave, but its latency will have lain outside of the respective search window (see insets in Fig. 1E, left panel), and so, exclusion will have created a selection bias towards the wave's group-average latency. If there was an association between the wave's latency and amplitude, the bias would have also affected the amplitudes, and thus potentially created a false null result. In the current data set, the wave-V latency and amplitude exhibited such association (Fig. 1E, right panel) – in this case, created by the effect of gender [44,45] rather than synaptopathy. As a result, imposing the same search windows as Prendergast, Guest, Munro, Kluk, Leger, Hall, Heinz and Plack [16] would have excluded significantly more females than males at the lower end of the latency range (18 vs 2; binomial test: p < 0.001), and significantly more males than females at the higher end of the range (22 vs 4; p < 0.001), and this would have reduced the wave-V amplitude difference between males and females by a substantial 21%.
3. Materials and methods
3.1. Design considerations
The previous discussion highlights the need for a new, more specific procedure for extracting ABR wave latencies and amplitudes. Like the response template [22] and DTW [33] approaches, the new procedure should align individual ABRs to create a time-variability-adjusted “structural average” [23] response and define a common “structural” time axis, comparable to the standardized spatial coordinate systems used for aligning medical imaging data [46]. Relevant waves could then be identified in the structural average response, and their peak and trough latencies determined on the common structural time axis (referred to as “structural latencies”). The individual wave amplitudes could then be determined by evaluating the aligned individual responses at the structural peak and through latencies (to yield individual peak and trough amplitudes, although alternative amplitude measures, such as the peak area, could also be derived), and the individual wave latencies could be determined by transforming the structural peak latencies back to the original (unaligned) individual time axes.
As mentioned in Section 2, the DTW approach [33] used a discrete time warping procedure. Such procedures are designed to align sequences of elements that are not amenable to averaging or interpolation, particularly, non-numerical elements, such as amino acids in proteins or nucleotides in DNA and RNA [47]. As a result, discrete time warping is an inherently pairwise procedure, and generalizing it to find the structural average of a set of multiple sequences, as would be required here, is not a trivial task [48]. Moreover, discrete time warping does not allow time-scaling of segments within the to-be-aligned sequences. Instead, single elements in one sequence are allowed to map to multiple elements in the other (Fig. 1D, left panel), and thus the warping path (Fig. 1D, right panel), defining the transformation from the original to the aligned sequences, is non-invertible. When applied to ABRs, this would preclude transforming the structural latencies (on the aligned time axis) back to the original (unaligned) time axes.
Therefore, we here propose to use a continuous implementation of dynamic time warping [18,19], also known as “(non-linear) curve registration”. Curve registration is part of functional data analysis [49] and deals with sequences of continuous numerical data that, like ABRs, can be regarded as discretizations of smooth functions, or “curves” (rather than sequences of discrete elements, such as proteins). In the continuous case, the warping functions (equivalent to the warping paths in the discrete case) are specifically designed to be not only smooth, but also strictly increasing (see Fig. 2A, upper and middle panels). This ensures that single points in the original curves are uniquely mapped to single points in the structural average curve and vice versa (i.e., the warping functions are invertible). To construct such functions, we here opted for the “continuous monotone registration” method proposed by Ramsay [50], which models the logarithm of the warping functions’ derivative as a linear combination of B-spline, or “basis-spline” functions (Fig. 2A, lower panel; see Section 3.3.1).
To test the approach, we applied it to extract the wave-I and -V latencies and amplitudes from an example ABR data set comprising 23 normal healthy subjects and 12 different experimental conditions (Fig. 2B). To create a gold-standard reference, the same latencies and amplitudes were also extracted manually by three expert observers (the three authors), who were blinded to subjects and conditions.
3.2. Data acquisition and pre-processing
3.2.1. ABR acquisition
ABRs were recorded from a total of 23 (14 female, 9 male) young (age range = 20–37 years), normal-hearing (thresholds in the test ear ≤ 25 dB HL at all octave frequencies between 0.125 and 8 kHz) subjects, none of whom reported any history of audiological or neurological disease, and who gave prior written informed consent. The study procedures complied with the Declaration of Helsinki guidelines (version 6, 2008) and were approved by the Ethics Committee of the University of Nottingham School of Medicine, but were not formally pre-registered as stipulated by the Declaration's 2014 amendment.
The evoking stimulus was either a click, or a chirp designed to compensate for cochlear dispersion. The click and chirp had the same energy level, corresponding to a peak-equivalent sound pressure level (SPL) of 90 dB for the click. They were generated by adding sinusoids at integer multiples of 10 Hz between 0.25 and 8 kHz (with rounded spectral edges to avoid edge tones) either in cosine phase (for the click) or with the phase delays proposed by Elberling and Don [51] (for the chirp). They were presented either in quiet (referred to as “broadband” condition), or in a background of continuous high-pass-filtered noise (referred to as “high-pass-masked” conditions). The noise was filtered from an 80-dB SPL white noise using one of five different octave-spaced cutoff frequencies (0.5, 1, 2, 4 or 8 kHz). The noise is intended to mask any response contributions above the cutoff frequency. Normally, high-pass masking is used to estimate response contributions from narrow frequency regions [so-called “derived-band” responses; 52]. Here, masking was merely used to generate responses with a wide range of latencies and SNRs. The stimuli were presented with alternating polarity at a rate of 20/s and repeated for a total of at least 6000 times for each condition.
The stimuli and the noise were generated digitally (24.4-kHz sampling rate) in Matlab (version R2019a, The Mathworks, Natick, MA, USA), digital-to-analogue converted (24-bit amplitude resolution) using Tucker Davis Technologies (Alachua, FL, USA) System 3 (RP2.1 and HB7) and presented monaurally to the left ear via ER-2 insert earphones (Etymotic Research Inc., Elk Grove Village, IL, USA).
The ABRs were acquired using a BioSemi ActiveTwo system (BioSemi B. V., Amsterdam, Netherlands) with ABR-type electrodes and a vertical montage, with reference at the vertex, active electrode on the ipsilateral mastoid, and ground on the forehead. They were sampled at a rate of 16.384 kHz.
3.2.2. ABR pre-processing
All pre-processing was performed using Matlab R2019a. First, the raw data were low-pass filtered at 2 kHz and high-pass filtered at 100 and 150 Hz (both the low- and high-pass filters were implemented as 4th-order Butterworth IIR filters), and divided into 45-ms epochs including a 5-ms pre-stimulus baseline. Then, the 6000 least artefactual epochs (in terms of overall maximum voltage) were submitted to a Bayesian weighted-averaging procedure [using 250-trial blocks; 53] to create an overall average response for each condition and subject, as well as two replicate responses comprising alternate halves (3000) of the trials. The Bayesian-averaged responses were then further averaged across subjects to create a “cross-sectional” average response for each condition.
Fig. 2B shows that, particularly for the click stimulus, the average responses for different high-pass-masking conditions differed substantially in overall latency. This is due to cochlear dispersion [54]. Such average condition-related latency differences should be eliminated by linear time shifting before trying to non-linearly align the individual responses [18]. Here, we advanced the high-pass-masked responses to coincide with the earlier broadband response (Fig. 2C). The shift delays were based on the average responses across subjects. They were determined by cross-correlating the responses and finding the lag that maximized their correlation coefficient for the time range between 0 and 12 ms.
Next, we combined the 100- and 150-Hz high-pass-filtered versions of the responses by cross-fading them with 2-ms linear transition ramps centred on 5 ms. As a result, the combined responses contained the 150-Hz-filtered version in the time range of wave I (≤ 4 ms), and the 100-Hz-filtered version in the time range of wave V (≥ 6 ms), as we found these filter settings to be optimal in pilot tests [see also 55]. Finally, each individual response was mean-corrected and normalized by its root-mean-square amplitude over the time range from 0 to 12 ms (compare scale bars in Figs 2B and 2C).
3.3. Automatic extraction of ABR wave latencies and amplitudes
3.3.1. Construction of time warping functions
Time warping was performed within the time range between 0 and 12 ms (grey highlight in Fig. 2C), which contained both the peaks and subsequent troughs of waves I and V in all pre-processed responses. For convenience, this was normalized to range from 0 to 1 (see Fig. 2A). To align a given individual response, xi(t) (where i indicates the individual and t indicates time), with a given target, y(t), a time warping function, hi(t), was constructed by modelling the natural logarithm of its first derivative, ln {hi′(t)}, as a linear combination, , of 2nd-order (linear) B-spline functions, Bk(t), with Kequally spaced knots, tk = (k − 1)/(K − 1) = (k − 1)Δt, where Δt = 1/(K − 1) is the knot spacing (see Fig. 2A, lower panel). K was set to 13, so that Δt corresponded to 1 ms. The B-splines were defined as:
Note that, for k = 1, Bk is not defined over [tk − 1 tk], and for k = K, Bk is not defined over [tk tk + 1]. In order to derive hi(t), wi(t) was submitted to the exponential function, integrated, and then scaled and shifted, so that hi(t) spanned the same range as t ([0 1]): . The B-spline coefficients, ci = [ci1, ci2, …, ciK], were fitted by minimizing a suitable cost function (Section 3.3.2) using the constrained non-linear problem solver fminconin Matlab with the default interior-point algorithm.
Time warping was performed on the discrete ABR time series using Matlab R2019a. Derivatives were approximated by Newton's difference quotient, and integrals by the rectangle rule. The warped responses, , were created by evaluating the original responses, xi(t), at the warped time points, hi(t), using linear interpolation.
3.3.2. Choice of fitting criterion
In its original version, the continuous monotone registration method fits the B-spline coefficients by minimizing a penalized squared difference criterion, . Here, the first term is the sum of squared differences between the warped and target curves, and y(t), and the second term is a warping roughness penalty, which shrinks the derivative, wi′(t), of the fitted spline function, wi(t) (see Section 3.3.1), according to a smoothing parameter, λ. Here, we tested five different values of λ, including λ = 0 (no roughness penalty) and four non-zero values, ranging from 0.001 to 1 in decade steps. As the spline function models the logarithm of the derivative of the warping function, ln {hi′(t)}, its derivative represents the warping function's relative curvature, hi′′(t)/hi′(t), [50].
The squared deviation shrinks the pointwise amplitude differences between the warped and target curves. If the curves, or some of their features, differ in amplitude, this can lead to shape distortion by shrinking time periods where amplitude differences are large, and expanding periods where amplitude differences are small. An example of shape distortion is shown in the left panel of Fig. 3A. The roughness penalty can ameliorate this problem, as shape distortion tends to be associated with high warping function curvature, or roughness (Fig. 3A, right panel), but rarely eliminates the problem entirely. Other means of reducing shape distortion include reducing amplitude differences between the to-be-aligned and target curves [19,56] – here, we normalized the individual ABRs by their root-mean-square amplitude (see Section 3.2.2) – and using a different fitting criterion [18]. Here, we tested a total of four different criteria, including the original PSD criterion, as well as three alternative criteria. The first alternative criterion, the penalized squared difference of the derivatives, or PSDD, criterion is similar to the PSD criterion, except that it minimizes the sum of squared differences between the curves’ derivatives (Fig. 3B), , rather than the curves themselves. The second alternative criterion, the penalized maximum correlation, , maximizes the Pearson correlation, ρ, between the warped and target curves, rather than minimizing their squared differences. As the Pearson correlation is independent of overall amplitude differences, correlation maximization may be more appropriate for problems like time warping, where a ‘perfect’ fit does not necessarily mean zero deviation between the warped and target curves [Fig. 3A middle panel; 57]. Finally, for completeness, we also tested the penalized maximum correlation of the derivatives, or PMCD, criterion, which is similar to the PMCcriterion except that it maximizes the correlation of the curves’ derivatives, . All four fitting criteria (PSD, PSDD, PMC, and PMCD) were paired with all five values of the smoothing parameter, λ (0, 0.001, 0.01, 0.1 and 1), yielding a total of 20 loss functions.
3.3.3. Choice of registration procedure
To align a set of individual curves, xi(t), i = 1, 2, …, N, and construct their structural average (average of aligned curves), , the continuous monotone registration method first warps the original curves (xi(t)) to the cross-sectional average curve, , and then iterates the process, using the structural average curve (x*(t)) as new warping target. The process is typically stopped after the first iteration, as the average of the first-warped curves usually differs little from that of the second-warped curves. This was also true for the current data (Fig. 3C, left panel). This is the so-called “average target” (at) registration procedure.
In addition, we also tested the “pairwise” (pw) registration procedure proposed by Tang and Müller [58], which aligns each curve, xi(t), with every other curve, xj ≠ i(t) (see Fig. 3C, right panel, for an example), and then takes the average of the resulting warping functions as the curve's overall warping function, . Tang and Müller suggested that pwregistration may be less prone to shape distortion than at registration. Both procedures were paired with all of the 20 loss functions to yield a total of 40 registration conditions.
3.3.4. Registration protocol and extraction of wave latencies and amplitudes
Registration was performed separately for each stimulus condition (click or chirp), but collectively for the different high-pass-masking conditions, using the broadband condition (with the highest SNR) as common warping target. For the at registration procedure, each individual ABR for each condition was first warped to the average broadband response, and then the resulting warped responses were warped again to average of the first-warped broadband responses. For the pw registration procedure, each individual response for each condition was warped to each of the other individuals’ responses for the broadband condition and the resulting warping functions averaged.
All registrations were repeated three times – once for the overall individual responses (based on all 6000 trials; see Section 3.2.2), and again for each of the two replicate responses (based on alternate sets of 3000 trials). The warping functions resulting from the replicate registrations, and , were used to create “same-warped” and “cross-warped” replicates by applying them either to the replicates used for fitting them (e.g., ), or to the respective other replicates (e.g., ). The registrations of the overall responses were used to extract the latencies and amplitudes of waves I and V – the waves used in the human synaptopathy literature (see Section 1). For that, we picked the latencies of the waves’ peaks and troughs, tpeakand ttrough, in the structural “grand-average” response for each stimulus condition (that is, the average of the structural average responses across all high-pass-masking conditions, Fig. 3D; alternatively, we could also have used the structural average responses for the broadband condition only). To estimate the waves’ amplitudes, we took the amplitudes of the aligned individual responses at these latencies, , and calculated their peak-to-trough differences. And to obtain the waves’ latencies, we evaluated the individual warping functions at the structural peak latencies (hi(tpeak)) using linear interpolation, and then transformed back to milliseconds (as tand hi(t) were normalized between 0 and 1) and added the correlation lags between the average high-pass-masked and broadband responses, which had been used for linearly pre-aligning the responses at the pre-processing stage (see Section 3.2.2).
3.4. Statistics
A linear mixed model analysis of the manually picked latency differences between waves I and V was performed using the nlme package [59] for R [60]. The model included a fixed condition factor (12 levels) and a random subject intercept. The significance of the subject effect was tested using the likelihood ratio method.
A Bayes factor analysis comparing the automatically and manually picked wave-I and -V latencies and amplitudes was performed using the BayesFactor package for R [61]. The analysis was applied separately to each stimulus condition (click/chirp) and each data type (latencies/amplitudes). For each condition and data type, the Bayes factor was computed between the most likely model that included only wave (I or V) and high-pass-masking condition (0.5, 1, 2, 4, 8 and inf kHz) as fixed factors and the most likely model that additionally included the factor of picking method (manual vs automatic). Subjects were included as random intercepts.
4. Results
4.1. Alignment strength
Fig. 4 shows example registration results for a representative subset of registration conditions, illustrated using the individual ABRs for the broadband chirp condition. Panel A shows the original (unaligned) responses, replotted from Fig. 1A. Panel B suggests that most registration conditions created at least some degree of response alignment, in that the aligned individual responses were generally more tightly clustered around their average than the original responses. However, the figure also suggests that the strength of the alignment effect varied across conditions, and that some conditions caused shape distortion. Specifically, for the PSD and PMC fitting criteria combined with the average-target (at) registration procedure and no roughness penalty (smoothing parameter, λ = 0), it is apparent that some response portions were unduly stretched, resulting in unnaturally flat plateaus, whilst some portions were unduly compressed, resulting in unnaturally sharp peaks or troughs.
To quantify alignment strength, we calculated the original and aligned responses’ pointwise standard deviations across subjects (Fig. 5A, left panel), after mean-correcting and root-mean-square amplitude normalizing the responses over the warping time range (0–12 ms). To create a single-valued measure for each warping condition, henceforth referred to as “response deviation” (Fig. 5A, right panel), we then averaged the pointwise deviations over time (0–12 ms) and across all experimental conditions. The response deviations of the aligned responses were consistently smaller than those of the original responses, confirming that all tested registration conditions created at least some degree of alignment. Generally, the fitting criteria based on the ABR waveforms (PSD and PMC) created stronger alignment (smaller response deviations) than those based on the ABR derivatives (PSDD and PMCD), and the fitting criteria involving squared-difference minimization (PSD and PSDD) created stronger alignment than those involving correlation maximization (PMC and PMCD). Response deviations were also generally smaller (indicating stronger alignment) for the average-target (at) than pairwise (pw) registration procedure. Finally, response deviations tended to increase with increasing smoothing parameter, λ, but only for those conditions that yielded small response deviations (strong alignment) when λ was zero, namely the PSD and PMC fitting criteria, particularly when combined with the at registration procedure. In contrast, the response deviations for the PSDD and PMCD fitting criteria depended little on λ.
4.2. Replicate cross-validation
The response deviations shown in Fig. 5A are based on the overall responses for each subject and condition (based on all 6000 trials; see Section 3.2.2). They measure overall alignment strength, but do not indicate whether the responses were over-aligned: when warping functions are fitted and applied to the same responses, it is possible that not only the ‘true’ ABRs are aligned, but also the inherent noise. To examine whether, or to what degree, this applied to the tested registration conditions, we also calculated the response deviations of the same- and cross-warped replicate responses (Fig. 5B, left panel; Section 3.3.4). Replicates contain the same ‘true’ ABR but independent noises [62]. Consequently, an over-aligning warping function should align both the ABR and the noise in the same-warped replicates, but only the ABR in the cross-warped replicates. Thus, the difference between cross- and same-warped response deviations should indicate the degree of noise alignment. Conversely, the difference between original and cross-warped response deviations should indicate the degree of alignment of the ‘true’ ABR.
Fig. 5B (right panel) suggests that those registration conditions that created the strongest alignment of the overall responses also created strong noise alignment of the replicates (large differences between the cross- and same-warped response deviations). This was particularly apparent for the PSD and PMC fitting criteria combined with the at registration procedure and no or little roughness penalty (λ near zero). In contrast, the conditions that created the weakest alignment of the overall responses created little noise alignment (small differences between cross- and same-warped response deviations), but also little ABR alignment (similarly small differences between original and cross-warped response deviations) of the replicates. This was true for the PSDD and PMCD fitting criteria combined with the pw registration procedure, irrespective of λ. Some conditions, however, which created intermediate alignment of the overall responses, seemed to create at least some degree of ABR alignment of the replicate responses, whilst, at the same time, not causing too much noise alignment. This applied especially to the PSD and PMC fitting criteria combined with the at registration procedure and a strong roughness penalty (λ close to unity) and, to a slightly lesser degree, the PSDD and PMCD fitting criteria combined with the at registration procedure, irrespective of λ.
4.3. Comparison between automatic and manual picking results
Here, we examine how well the tested registration conditions were able to replicate our manual picking results for the wave-I and -V latencies and amplitudes. To quantify the correspondence between automatically and manually picked data, we calculated their root-mean-square deviations across waves and experimental conditions (Fig. 6A). Both the latency and amplitude deviations showed considerable variation across registration conditions, yet, large latency deviations were not necessarily associated with large amplitude deviations and vice versa. Large latency deviations were particularly associated with registration conditions that created over-alignment (strong noise alignment), namely, the PSD and PMC fitting criteria combined with the at registration procedure and no or little roughness penalty (λ near zero). This is understandable, given that over-alignment is associated with over-warping of the individual time axes, and thus misestimation of individual latencies. In contrast, large amplitude deviations were particularly associated with registration conditions that caused under-alignment (little noise, but also little ABR alignment), namely, the PSDD and PMCD fitting criteria combined with the pw registration procedure, irrespective of λ. Again, this is understandable: in the extreme case where a registration condition fails to cause any alignment at all, individual amplitudes would be measured at a constant latency across subjects, and would thus deviate substantially from their actual values. The smallest combined latency and amplitude deviations were yielded by the registration conditions that created intermediate alignment effects (reasonable ABR alignment and not too much noise alignment), particularly the PSDD fitting criterion combined with the at registration procedure, across all values of λ. The overall smallest combined deviation was associated with the PSDD, at and λ = 0 condition.
Fig. 6B shows that, despite the large differences in amplitude, and thus SNR, between waves and experimental conditions, the match between automatically and manually picked latencies and amplitudes was remarkably good across both conditions and waves. The correspondence was formally tested with a Bayes Factor analysis comparing the most likely statistical models excluding and including the effect of picking method (manual vs automatic) for each stimulus (click/chirp) and data type (latencies/amplitudes; see Section 3.4). All Bayes factors were either greater than, or close to, unity, and thus did not provide compelling evidence against the null hypothesis that the manually and automatically picked latencies and amplitudes were not statistically different. This is consistent with the finding, depicted in Fig. 6D and E, that manually and automatically picked latencies and amplitudes were highly significantly correlated. At the same time, Bayes factors were generally larger (indicating a better match between automatically and manually picked data) for the chirp than for the click stimulus. Presumably, this is, because the chirp yielded larger responses and thus better response SNRs.
4.4. Simulated data
The current ABR data set exhibited a wide range of SNRs, from, on average, −8.6 dB for the click-evoked condition high-pass-masked at 0.5 kHz, to 9.8 dB for the broadband chirp-evoked condition. To investigate the effect of SNR more directly, we created a realistic simulated data set with a similar range of SNRs as the measured data. This enabled us to create individual responses with known actual warping functions, and thus directly evaluate the accuracy with which they were estimated.
To create the simulated data, we first stacked the click-evoked ABRs to create larger broadband responses, compensated for cochlear dispersion [63], and then aligned the stacked responses using the PSDD, at, λ = 0 registration condition that created the best match between manual and automatic picking results. Then, we averaged the aligned responses across subjects to create a prototypical, or structural response, S(t) (Fig. 7A), which we de-aligned using the inverse of the individual warping functions, , to create individual true ABRs, . Finally, we created individual noise traces by taking the plus-minus reference of the replicate responses, [62], and then added the true ABRs to the noise traces after multiplying them with factors 10Δsnr/20, where Δsnr ranged from −21 to 0 dB in 3-dB steps: . The SNRs of the resulting simulated responses, , ranged from −10.3 dB at Δsnr = −21 dB to 10.7 dB at Δsnr = 0 dB. Like the measured responses, the simulated responses were warped to the condition with the highest SNR (Δsnr = 0 dB) using the PSDD, at, λ = 0 registration condition. This yielded “estimated” warping functions, , which were compared with the actual warping functions, hi(t), by correlating with hi(t) − t (Fig. 7C).
Fig. 7B shows that the reduction in response deviation between the original and aligned simulated responses decreased with decreasing SNR. This is expected to the extent that the warping aligned the true ABRs rather than the noise traces. At the same time, the correspondence between the estimated and actual warping functions also decreased with decreasing SNR, from an average of 0.855 at SNR = 10.7 dB to an average of only 0.161 at SNR = −10.3 dB (Fig. 7C). This shows that alignment quality depends strongly on the quality of the to-be-aligned responses: the lower the response SNRs, the more the alignment will be affected by noise. This, however, will also be the case for manual picking, as human observers will pick waveform peaks and troughs, the location and size of which is influenced by noise. This may be why manual and automatic picking results showed good correspondence even for conditions that yielded poor SNRs.
4.5. Results summary
In comparing different registration conditions, we found that (i) average-target (at) registration generally yielded better results than pairwise (pw) registration, (ii) squared-difference minimization generally yielded better results than correlation maximization, and (iii) registering the response derivatives generally yielded better results than registering the responses themselves.
Overall, the at registration procedure, combined with squared-difference minimization of the response derivatives (PSDD) with no roughness penalty (λ = 0) yielded the best match between automatically and manually picked wave-I and -V latencies and amplitudes.
5. Discussion
The aim of this study was to develop a procedure for extracting individual latencies and amplitudes of ABR waves with minimal manual involvement. The procedure involves non-linearly aligning, or “warping”, individual ABRs using the methodology of curve registration. To construct smooth and invertible time warping functions, we adopted the computationally efficient “continuous monotone registration” approach proposed by Ramsay [50]. Once aligned, the individual ABRs are averaged to create a structural average response. This is used for picking the required waves’ structural peak and trough latencies, which, in turn, are used to derive their individual latencies and amplitudes.
In the current study, the procedure was applied to align a set of individual ABRs (from multiple subjects and conditions) with one another, and thus the structural average response was constructed from that set. Alternatively, individual ABRs could be aligned with a previously constructed structural average response, based on an independent (e.g., normative) data set, and the respective individual wave latencies and amplitudes could be derived from previously picked (normative) structural peak and trough latencies.
To minimize the danger of over-alignment, any differences in amplitude shift or scale between the to-be-aligned responses should be minimized by appropriate linear transformations. In the current study, individual ABRs were mean-centred and normalized by their root-mean-square amplitude. Similarly, any overall latency differences between the to-be-aligned responses should be accounted for by linear time shifts. In the current study, ABRs were shifted by the latency differences between the subject-average responses for different experimental conditions. In other situations, linear time shifts may be required to account for overall latency differences between different subjects or subject groups.
In the current study, the proposed procedure was applied to an example set of transient-evoked ABRs. There is no reason, however, why it could not also be applied to other types of ABRs, particularly those evoked by longer and more complex sounds, such as speech syllables or short trains of clicks or chirps [64,65]. Such responses typically consist of multiple peaks, which roughly follow the stimulus waveform and are thus commonly referred to as “frequency-” or “envelope-following responses” (FFRs or EFRs; see Fig. 7D, upper panel, for an example). F/EFRs are often evaluated using summary measures, such as the stimulus-to-response correlation or the response frequency spectrum [66], calculated over the entire response duration. Such measures, however, cannot reveal differences between individual response peaks, which may arise as a result of differences between eliciting stimulus portions, or in consequence of neural factors, such as adaptation or efferent feedback [67]. Nonlinear curve registration could be used to evaluate such inter-peak differences (see Fig. 7D, lower panel).
In the current study, all tested criteria for fitting the time warping functions were based on the deviation between the warped and target responses (or response derivatives), without any consideration of their variability. In contrast, warping paths for aligning speech tokens with templates in automatic speech recognition are fitted with consideration of the templates’ segment-wise variability [68]. A similar strategy could also be applied to ABRs, particularly when aligning individual responses with a normative target response. Specifically, the deviation between the warped and target responses could be weighted in inverse proportion to the target variance (see Fig. 7E).
6. Appendix: mode of availability of software
The current procedure involves (i) pre-processing of individual ABRs, (ii) alignment of the responses using continuous monotone time warping, and, finally (iii), picking of structural peak and trough latencies and derivation of individual wave latencies and amplitudes. All of these steps have been implemented as separate Matlab functions, preproc.m, nlcurvereg.m and xtractlatamp.m, respectively (written in Matlab version R2019a), which are available for download from https://github.com/mszkk3/Non-linear-ABR-registration-.git. All three functions provide the option of analysing data either from a single experimental condition or from multiple conditions. In addition, they also give different options for defining the warping target. When there is only a single experimental condition, the target is the average response across all individual ABRs. When there are multiple conditions, the user has the option of warping to the subject-average response for a specific condition (like the broadband condition in the current study), or the grand-average response across all subjects and conditions. Alternatively, the responses can also be aligned with a separate target response, which could be based on an independent data set. The pre-processing function, preproc.m, offers the choice of pre-aligning either the subject-average responses across conditions (in the case of multiple conditions), or the individual responses across individuals (in the case of a single condition). The warping function, nlcurvereg.m uses the non-linear problem solver, fmincon, which requires Matlab's Optimization toolbox.
In addition, we also provide the current chirp-evoked data in a .mat file (chirpdata.mat) as well as a matlab script with three examples of how the functions’ different options could be used.
Declaration of Competing Interest
None of the authors declares any conflict of interest.
Acknowledgments
Acknowledgments
We thank the Hearing Theme of the NIHR Nottingham Biomedical Research Centre for valuable intellectual input and help with subject recruitment.
Funding information
This work was supported by the Medical Research Council [grant numbers MR/S003320/1, MC_UU_00010/2].
Footnotes
Present address: Hearing Sciences, School of Medicine, University of Nottingham, University Park, Building 40, NG7 2RD Nottingham, UK
References
- 1.Kujawa S.G., Liberman M.C. Adding insult to injury: cochlear nerve degeneration after "temporary" noise-induced hearing loss. J. Neurosci. 2009;29:14077–14085. doi: 10.1523/JNEUROSCI.2845-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Sergeyenko Y., Lall K., Liberman M.C., Kujawa S.G. Age-related cochlear synaptopathy: an early-onset contributor to auditory functional decline. J. Neurosci. 2013;33:13686–13694. doi: 10.1523/JNEUROSCI.1783-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Kujawa S.G., Liberman M.C. Synaptopathy in the noise-exposed and aging cochlea: primary neural degeneration in acquired sensorineural hearing loss. Hear. Res. 2015;330:191–199. doi: 10.1016/j.heares.2015.02.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Liberman M.C., Kujawa S.G. Cochlear synaptopathy in acquired sensorineural hearing loss: manifestations and mechanisms. Hear. Res. 2017;349:138–147. doi: 10.1016/j.heares.2017.01.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Salvi R., Sun W., Ding D., Chen G.D., Lobarinas E., Wang J., Radziwon K., Auerbach B.D. Inner hair cell loss disrupts hearing and cochlear function leading to sensory deprivation and enhanced central auditory gain. Front. Neurosci. 2016;10:621. doi: 10.3389/fnins.2016.00621. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Lobarinas E., Salvi R., Ding D. Insensitivity of the audiogram to carboplatin induced inner hair cell loss in chinchillas. Hear. Res. 2013;302:113–120. doi: 10.1016/j.heares.2013.03.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Wong S.J., Abrams K.S., Amburgey K.N., Wang Y., Henry K.S. Effects of selective auditory-nerve damage on the behavioral audiogram and temporal integration in the budgerigar. Hear. Res. 2019;374:24–34. doi: 10.1016/j.heares.2019.01.019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Bramhall N., Beach E.F., Epp B., Le Prell C.G., Lopez-Poveda E.A., Plack C.J., Schaette R., Verhulst S., Canlon B. The search for noise-induced cochlear synaptopathy in humans: mission impossible? Hear. Res. 2019;377:88–103. doi: 10.1016/j.heares.2019.02.016. [DOI] [PubMed] [Google Scholar]
- 9.Le Prell C.G. Effects of noise exposure on auditory brainstem response and speech-in-noise tasks: a review of the literature. Int. J. Audiol. 2019;58:S3–S32. doi: 10.1080/14992027.2018.1534010. [DOI] [PubMed] [Google Scholar]
- 10.Plack C.J., Leger A., Prendergast G., Kluk K., Guest H., Munro K.J. Toward a diagnostic test for hidden hearing loss. Trends Hear. 2016;20:1–9. doi: 10.1177/2331216516657466. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Bharadwaj H.M., Mai A.R., Simpson J.M., Choi I., Heinz M.G., Shinn-Cunningham B.G. Non-invasive assays of cochlear synaptopathy - candidates and considerations. Neuroscience. 2019;407:53–66. doi: 10.1016/j.neuroscience.2019.02.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Stamper G.C., Johnson T.A. Auditory function in normal-hearing, noise-exposed human ears. Ear Hear. 2015;36:172–184. doi: 10.1097/AUD.0000000000000107. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Bramhall N.F., Konrad-Martin D., McMillan G.P., Griest S.E. Auditory brainstem response altered in humans with noise exposure despite normal outer hair cell function. Ear Hear. 2017;38:e1–e12. doi: 10.1097/AUD.0000000000000370. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Johannesen P.T., Buzo B.C., Lopez-Poveda E.A. Evidence for age-related cochlear synaptopathy in humans unconnected to speech-in-noise intelligibility deficits. Hear. Res. 2019;374:35–48. doi: 10.1016/j.heares.2019.01.017. [DOI] [PubMed] [Google Scholar]
- 15.Schaette R., McAlpine D. Tinnitus with a normal audiogram: physiological evidence for hidden hearing loss and computational model. J. Neurosci. 2011;31:13452–13457. doi: 10.1523/JNEUROSCI.2156-11.2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Prendergast G., Guest H., Munro K.J., Kluk K., Leger A., Hall D.A., Heinz M.G., Plack C.J. Effects of noise exposure on young adults with normal audiograms I: Electrophysiology. Hear. Res. 2017;344:68–81. doi: 10.1016/j.heares.2016.10.028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Guest H., Munro K.J., Prendergast G., Howe S., Plack C.J. Tinnitus with a normal audiogram: Relation to noise exposure but no evidence for cochlear synaptopathy. Hear. Res. 2017;344:265–274. doi: 10.1016/j.heares.2016.12.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Ramsay J.O., Li X. Curve registration. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 1998;60:351–363. [Google Scholar]
- 19.Wang K.M., Gasser T. Alignment of curves by dynamic time warping. Ann. Stat. 1997;25:1251–1276. [Google Scholar]
- 20.Jewett D.L., Romano M.N., Williston J.S. Human auditory evoked potentials: possible brain stem components detected on the scalp. Science. 1970;167:1517–1518. doi: 10.1126/science.167.3924.1517. [DOI] [PubMed] [Google Scholar]
- 21.Møller A.R. Neural generators for auditory brainstem evoked potentials. In: Burkard R.F., Eggermont J.J., Don M., editors. Auditory Evoked Potentials. Lippincott Williams & Wilkins; Baltimore, USA: 2007. [Google Scholar]
- 22.Elberling C. Auditory electrophysiology. The use of templates and cross correlation functions in the analysis of brain stem potentials. Scand. Audiol. 1979;8:187–190. doi: 10.3109/01050397909076320. [DOI] [PubMed] [Google Scholar]
- 23.Kneip A., Gasser T. Statistical tools to analyze data representing a sample of curves. Ann. Stat. 1992;20:1266–1305. [Google Scholar]
- 24.Issa A., Ross H.F. An improved procedure for assessing ABR latency in young subjects based on a new normative data set. Int. J. Pediatr. Otorhinolaryngol. 1995;32:35–47. doi: 10.1016/0165-5876(94)01110-j. [DOI] [PubMed] [Google Scholar]
- 25.Eggermont J.J. Auditory brainstem response. Handb. Clin. Neurol. 2019;160:451–464. doi: 10.1016/B978-0-444-64032-1.00030-8. [DOI] [PubMed] [Google Scholar]
- 26.Motsch J.F. Université Paris XII; Paris: 1987. La Dynamique Temporelle Du Tronc Cérébral. [Google Scholar]
- 27.Vannier E., Adam O., Karasinski P., Ohresser M., Motsch J.F. Computer-assisted ABR interpretation using the automatic construction of the latency-intensity curve. Audiology. 2001;40:191–201. [PubMed] [Google Scholar]
- 28.Vannier E., Adam O., Motsch J.F. Objective detection of brainstem auditory evoked potentials with a priori information from higher presentation levels. Artif. Intell. Med. 2002;25:283–301. doi: 10.1016/s0933-3657(02)00029-5. [DOI] [PubMed] [Google Scholar]
- 29.Valderrama J.T., de la Torre A., Alvarez I., Segura J.C., Thornton A.R., Sainz M., Vargas J.L. Automatic quality assessment and peak identification of auditory brainstem responses with fitted parametric peaks. Comput. Methods Programs Biomed. 2014;114:262–275. doi: 10.1016/j.cmpb.2014.02.015. [DOI] [PubMed] [Google Scholar]
- 30.Jewett D.L., Williston J.S. Auditory-evoked far fields averaged from the scalp of humans. Brain. 1971;94:681–696. doi: 10.1093/brain/94.4.681. [DOI] [PubMed] [Google Scholar]
- 31.Achor L.J., Starr A. Auditory brain stem responses in the cat. I. Intracranial and extracranial recordings. Electroencephalogr. Clin. Neurophysiol. 1980;48:154–173. doi: 10.1016/0013-4694(80)90301-6. [DOI] [PubMed] [Google Scholar]
- 32.Scherg M., von Cramon D. A new interpretation of the generators of BAEP waves I-V: results of a spatio-temporal dipole model. Electroencephalogr. Clin. Neurophysiol. 1985;62:290–299. doi: 10.1016/0168-5597(85)90006-1. [DOI] [PubMed] [Google Scholar]
- 33.Picton T., Hunt M., Mowrey R., Rodriguez R., Maru J. Evaluation of brain-stem auditory evoked potentials using dynamic time warping. Electroencephalogr. Clin. Neurophysiol. Evoked Potentials Sect. 1988;71:212–225. doi: 10.1016/0168-5597(88)90006-8. [DOI] [PubMed] [Google Scholar]
- 34.Sakoe H., Chiba S. Dynamic programming algorithm optimization for spoken word recognition. IEEE Trans. Acoust. Speech Signal Process. 1978;26:43–49. [Google Scholar]
- 35.Guest H., Munro K.J., Plack C.J. Tinnitus with a normal audiogram: role of high-frequency sensitivity and reanalysis of brainstem-response measures to avoid audiometric over-matching. Hear. Res. 2017;356:116–117. doi: 10.1016/j.heares.2017.10.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Guest H., Munro K.J., Prendergast G., Millman R.E., Plack C.J. Impaired speech perception in noise with a normal audiogram: No evidence for cochlear synaptopathy and no relation to lifetime noise exposure. Hear. Res. 2018;364:142–151. doi: 10.1016/j.heares.2018.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Gabriel S., Durrant J.D., Dickter A.E., Kephart J.E. Computer identification of waves in the auditory brain stem evoked potentials. Electroencephalogr. Clin. Neurophysiol. 1980;49:421–423. doi: 10.1016/0013-4694(80)90240-0. [DOI] [PubMed] [Google Scholar]
- 38.Fridman J., John E.R., Bergelson M., Kaiser J.B., Baird H.W. Application of digital filtering and automatic peak detection to brain stem auditory evoked potential. Electroencephalogr. Clin. Neurophysiol. 1982;53:405–416. doi: 10.1016/0013-4694(82)90005-0. [DOI] [PubMed] [Google Scholar]
- 39.Boston J.R. Automated interpretation of brainstem auditory evoked potentials: a prototype system. IEEE Trans. Biomed. Eng. 1989;36:528–532. doi: 10.1109/10.24254. [DOI] [PubMed] [Google Scholar]
- 40.Pool K.D., Finitzo T. Evaluation of a computer-automated program for clinical assessment of the auditory brain stem response. Ear Hear. 1989;10:304–310. doi: 10.1097/00003446-198910000-00006. [DOI] [PubMed] [Google Scholar]
- 41.Delgado R.E., Ozdamar O. Automated auditory brainstem response interpretation. IEEE Eng. Med. Biol. 1994;13:227–237. [Google Scholar]
- 42.Popescu M., Papadimitriou S., Karamitsos D., Bezerianos A. Adaptive denoising and multiscale detection of the V wave in brainstem auditory evoked potentials. Audiol. Neurootol. 1999;4:38–50. doi: 10.1159/000013818. [DOI] [PubMed] [Google Scholar]
- 43.Bradley A.P., Wilson W.J. Automated analysis of the auditory brainstem response using derivative estimation wavelets. Audiol. Neurootol. 2005;10:6–21. doi: 10.1159/000081544. [DOI] [PubMed] [Google Scholar]
- 44.Don M., Ponton C.W., Eggermont J.J., Masuda A. Gender differences in cochlear response time: an explanation for gender amplitude differences in the unmasked auditory brain-stem response. J. Acoust. Soc. Am. 1993;94:2135–2148. doi: 10.1121/1.407485. [DOI] [PubMed] [Google Scholar]
- 45.Don M., Ponton C.W., Eggermont J.J., Masuda A. Auditory brainstem response (ABR) peak amplitude variability reflects individual differences in cochlear response times. J. Acoust. Soc. Am. 1994;96:3476–3491. doi: 10.1121/1.410608. [DOI] [PubMed] [Google Scholar]
- 46.Oliveira F.P., Tavares J.M. Medical image registration: a review. Comput. Methods Biomech. Biomed. Eng. 2014;17:73–93. doi: 10.1080/10255842.2012.670855. [DOI] [PubMed] [Google Scholar]
- 47.Needleman S.B., Wunsch C.D. A general method applicable to the search for similarities in the amino acid sequence of two proteins. J. Mol. Biol. 1970;48:443–453. doi: 10.1016/0022-2836(70)90057-4. [DOI] [PubMed] [Google Scholar]
- 48.Petitjean F., Ketterlin A., Gançarski P. A global averaging method for dynamic time warping, with applications to clustering. Pattern Recognit. 2011;44:678–693. [Google Scholar]
- 49.Ramsay J.O., Silverman B.W. 2nd ed. Springer; New York: 2005. Functional Data Analysis. [Google Scholar]
- 50.Ramsay J.O. Estimating smooth monotone functions. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 1998;60:365–375. [Google Scholar]
- 51.Elberling C., Don M. Auditory brainstem responses to a chirp stimulus designed from derived-band latencies in normal-hearing subjects. J. Acoust. Soc. Am. 2008;124:3022–3037. doi: 10.1121/1.2990709. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Don M., Eggermont J.J. Analysis of the click-evoked brainstem potentials in man unsing high-pass noise masking. J. Acoust. Soc. Am. 1978;63:1084–1092. doi: 10.1121/1.381816. [DOI] [PubMed] [Google Scholar]
- 53.Elberling C., Wahlgreen O. Estimation of auditory brainstem response, ABR, by means of Bayesian inference. Scand. Audiol. 1985;14:89–96. doi: 10.3109/01050398509045928. [DOI] [PubMed] [Google Scholar]
- 54.Dau T., Wegner O., Mellert V., Kollmeier B. Auditory brainstem responses with optimized chirp signals compensating basilar-membrane dispersion. J. Acoust. Soc. Am. 2000;107:1530–1540. doi: 10.1121/1.428438. [DOI] [PubMed] [Google Scholar]
- 55.Mason S.M. Effects of high-pass filtering on the detection of the auditory brainstem response. Br. J. Audiol. 1984;18:155–161. doi: 10.3109/03005368409078942. [DOI] [PubMed] [Google Scholar]
- 56.Wang K.M., Gasser T. Synchronizing sample curves nonparametrically. Ann. Stat. 1999;27:439–460. [Google Scholar]
- 57.Livadiotis G., McComas D.J. Fitting method based on correlation maximization: applications in space physics. J. Geophys. Res. Space. 2013;118:2863–2875. [Google Scholar]
- 58.Tang R., Müller H.G. Pairwise curve synchronization for functional data. Biometrika. 2008;95:875–889. [Google Scholar]
- 59.J. Pinheiro, D. Bates, S. DebRoy, D. Sarkar, R Core Team, nlme: Linear and nonlinear mixed effects models, 2020, R package version 3.1-147, https://CRAN.R-project.org/package=nlme.
- 60.R Core Team . R Foundation for Statistical Computing; Vienna, Austria: 2020. R: A Language and Environment For Statistical Computing.https://www.R-project.org/ [Google Scholar]
- 61.D.M. Richard, J.N. Rouder, BayesFactor: Computation of Bayes Factors for Common Designs. , 2018, R package version 0.9.12-4.2, https://CRAN.R-project.org/package=BayesFactor.
- 62.Schimmel H. The (+/-) reference: accuracy of estimated mean components in average response studies. Science. 1967;157:92–94. doi: 10.1126/science.157.3784.92. [DOI] [PubMed] [Google Scholar]
- 63.Don M., Masuda A., Nelson R., Brackmann D. Successful detection of small acoustic tumors using the stacked derived-band auditory brain stem response amplitude. Am. J. Otol. 1997;18:608–621. ; discussion 682-605. [PubMed] [Google Scholar]
- 64.Worden F.G., Marsh J.T. Frequency-following (microphonic-like) neural responses evoked by sound. Electroencephalogr. Clin. Neurophysiol. 1968;25:42–52. doi: 10.1016/0013-4694(68)90085-0. [DOI] [PubMed] [Google Scholar]
- 65.Russo N., Nicol T., Musacchia G., Kraus N. Brainstem responses to speech syllables. Clin. Neurophysiol. 2004;115:2021–2030. doi: 10.1016/j.clinph.2004.04.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Skoe E., Kraus N. Auditory brain stem response to complex sounds: a tutorial. Ear Hear. 2010;31:302–324. doi: 10.1097/AUD.0b013e3181cdb272. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Junius D., Dau T. Influence of cochlear traveling wave and neural adaptation on auditory brainstem responses. Hear. Res. 2005;205:53–67. doi: 10.1016/j.heares.2005.03.001. [DOI] [PubMed] [Google Scholar]
- 68.Benzeghiba M., De Mori R., Deroo O., Dupont S., Erbes T., Jouvet D., Fissore L., Laface P., Mertins A., Ris C., Rose R., Tyagi V., Wellekens C. Automatic speech recognition and speech variability: A review. Speech Commun. 2007;49:763–786. [Google Scholar]