Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2021 Sep;69:None. doi: 10.1016/j.csl.2021.101221

NUVA: A Naming Utterance Verifier for Aphasia Treatment

David S Barbera a,, Mark Huckvale b, Victoria Fleming a, Emily Upton a, Henry Coley-Fisher a, Catherine Doogan a, Ian Shaw c, William Latham a, Alexander P Leff a, Jenny Crinion a
PMCID: PMC8117974  PMID: 34483474

Highlights

  • NUVA automatically assesses online word naming attempts in aphasia therapy.

  • Significantly more accurate and faster than leading commercial speech recognition.

  • Accuracies between 83.6% and 93.6% validate use in clinical research.

Index Terms: Speech disorders, Word naming, Aphasia, Anomia, Speech recognition, Dynamic Time Warping

Abstract

Anomia (word-finding difficulties) is the hallmark of aphasia, an acquired language disorder most commonly caused by stroke. Assessment of speech performance using picture naming tasks is a key method for both diagnosis and monitoring of responses to treatment interventions by people with aphasia (PWA). Currently, this assessment is conducted manually by speech and language therapists (SLT). Surprisingly, despite advancements in automatic speech recognition (ASR) and artificial intelligence with technologies like deep learning, research on developing automated systems for this task has been scarce. Here we present NUVA, an utterance verification system incorporating a deep learning element that classifies 'correct' versus' incorrect' naming attempts from aphasic stroke patients. When tested on eight native British-English speaking PWA the system's performance accuracy ranged between 83.6% to 93.6%, with a 10-fold cross-validation mean of 89.5%. This performance was not only significantly better than a baseline created for this study using one of the leading commercially available ASRs (Google speech-to-text service) but also comparable in some instances with two independent SLT ratings for the same dataset.

1. Introduction

Word retrieval difficulties(aka anomia) are one of the most pervasive symptoms of post-stroke aphasia (Matti. Laine, 2006). Recent data suggests there are around 350,000 people in the UK alone living with chronic aphasia (Stroke Association, 2018). Despite its' prevalence, few individuals receive a sufficient dose of speech and language therapy to recover maximally. For example, in the UK through the National Health Service, patients receive on average 8-12 hours (Code & Heron, 2003); however, reviews of speech and language intervention studies have shown superior outcomes for treatments that deliver around 100 hours of therapy (Bhogal et al., 2003; Brady et al., 2016). Assessment of patients' spoken picture naming abilities and then practising repetitively over time a range of vocabulary using spoken picture naming tasks, are both integral parts of impairment based anomia treatments (Whitworth et al., 2014).

Picture naming tasks (confrontation naming) are typically carried out in face-to-face settings, to allow SLTs to assess the individual's accuracy for each naming attempt, and, crucially, provide item-by-item feedback during therapy sessions. An Automated Speech Recognition system (ASR) that could reliably assess patients' speech performance on these picture naming tasks would offer two clear advantages. Firstly, increased consistency across naming attempts, and sensitivity to changes in patients' speech abilities over time, would enable therapeutic interventions can be better tracked. Secondly, ASR would allow patients to be able to perform these tasks independent of SLTs potentially remotely away from the clinic in the comfort of their own home. This would not only 'free-up' SLTs to deliver more complex interventions in their 'face-to-face' clinical time but also support more patients who are unable to travel into the clinic to use these tasks effectively, a need which has become more pressing in light of recent COVID-19 travel restrictions and social distancing.

1.1. ASR in disordered speech

To date, most ASRs applied to speech disorders have targeted people with dysarthria - a motor speech disorder. While this can also be caused by stroke and patients may manifest both dysarthria and aphasia concurrently, the nature and characteristics of the speech deficits in dysarthria and aphasia are quite different (Abad et al., 2013). Dysarthria is a speech disorder resulting from a weakness, paralysis, or incoordination of the speech musculature (Darley Frederic L. et al., 1969). In this disorder, patients know the target word, e.g., spoon and correctly retrieve the lexical item, but the sounds are distorted, e.g., /poon/. These articulation errors tend to be highly consistent in a pattern, e.g., a patient may have difficulties saying all words beginning with an /s/, irrespective of the type of word, due to impaired tongue movements. In contrast, aphasia is an acquired disorder of language that is caused by damage to certain areas of the brain, which are primarily responsible for language function (Akbarzadeh-T & Moshtagh-Khorasani, 2007). Patients with this disorder commonly report that they know what they want to say but just can't find the right words. They do not have a motor speech disorder (i.e. they can say the words and sounds correctly, repetition can be intact), but rather, have a word (lexical) retrieval problem, akin to the 'tip of the tongue' phenomenon experienced by non-aphasic individuals. Alongside this, patients can make a variety of errors when speaking; for example, they may want to say /tea/ but say /coffee/ instead, a semantically related word, or say /key/, a phonemically related word, or even say a non-word, e.g., /tife/. There is high variability in speech error types, both between aphasic individuals with the same disorder and within an individual over time, even within the same day. This variety in speech patterns is a challenge for ASR systems and perhaps explains why the use and study of ASR in aphasia has been surprisingly scarce to date.

1.2. Overview of ASR in aphasia

The earliest studies to explore ASR in aphasia adopted 'off-the-shelf' commercially available software developed for healthy speakers. They utilised these tools to provide automated speech performance feedback to aphasic patients during therapy with mixed results. Linebarger and colleagues (Linebarger et al., 2001) incorporated a speech recognition module as part of a computerised-based training for language production. They reported positive treatment outcomes for the patients but unfortunately not much detail about the utility of the ASR technology used. Wade and colleagues (J. Wade, 2001) investigated the use of speech recognition software in aphasia and reported ASR accuracy levels of 80% at single word level and at phrase level. The software required extensive training sessions to adapt to the aphasic speaker and utilised a limited vocabulary of 50 words and 24 sentences. Further improvements in the reliability and stability of the system were recommended before clinical implementation.

Having identified the limitations of the previous approaches and the large unmet clinical need, more recent research efforts have focused on developing new ASR technology that can be used more effectively in aphasia. The latest advances in the ASR field and a prominent use of deep neural networks (DNN) have enabled a small number of groups to achieve: (1) transcriptions of narrative speech by PWA in both English and Cantonese (Le & Provost, 2016; T. Lee et al., 2016); (2) preliminary detection of phonemic paraphasic speech errors in connected speech by PWA (Le et al., 2017); (3) quantitative analysis of spontaneous speech by PWA (Le et al., 2018); and, (4) assessment of speech impairment in Cantonese PWA (Qin et al., 2019). Nevertheless, these current ASR solutions remain susceptible to variations in the language spoken, e.g., dialects and accents. While large corpora of aphasic speech samples in openly available databases, e.g., AphasiaBank in USA (Forbes et al., 2012; MacWhinney et al., 2011) are being developed that will provide useful priors/training sets for new models in the future, it is likely that not all languages or accents will benefit equally from this approach.

To address the lack of globally relevant aphasic speech corpora, and to develop a system that is able to take account of the high variability of speech performance by PWA and the distinction between aphasic speech and dysarthric speech we aimed to develop an ASR system that could reliably assess spoken single word picture naming performance in patients with aphasia without motor speech impairment (dysarthria or apraxia).

1.3. ASR for single word naming performance by PWA

In contrast to the challenge of analysing spontaneous speech utterances, assessing spoken picture naming performance has the advantage that the target word is known. The challenge for ASR in this context is to verify that a certain target word is uttered in a given segment of speech (Abad et al., 2013). Within the context of an intervention, the ASR system, or utterance verifier system, must also process each utterance in quick succession so that a binary 'correct'/’incorrect' response can provide feedback to the PWA or the therapy governing algorithm on an item-by-item basis.

To the best of our knowledge, only two groups have used and assessed an ASR-based system of such type in aphasic speakers' single word picture naming performance. In the project Vithea (Pompili et al., 2011), researchers developed an app for the treatment of aphasia for Portuguese speakers. They first presented results of an in-house ASR-engine called AUDIMUS (Meinedo et al., 2003) assessing picture naming using a keyword spotting technique to score spoken naming attempts as 'correct'/’incorrect'. An updated version of this reported an average ASR accuracy of 82%, with ranges between 69% and 93% across patients (Abad et al., 2013). The second group (Ballard et al., 2019) evaluated a digitally delivered intervention using picture naming tasks in native Australian English speaking people with both apraxia and aphasia. They used the open-source ASR engine CMU PocketSphinx (Cmusphinx/Pocketsphinx, 2014/2020) to provide patients with 'correct'/’incorrect' feedback for each of their naming attempts during treatment. The ASR-engine recognized 124 words; each word was phonetically different from the others. They reported an overall system performance accuracy of 80% and a range of well-classified scores between 65.1% and 82.8% across patients depending on impairment severity. Both these systems provided good 'proof-of-concept 'data that ASR systems for assessing spoken word performance are feasible, but the high error rate and variable performance across aphasic patients meant their clinical utility remained low.

The aim of this project is to present and assess the feasibility and stability of NUVA, a tailor-made ASR system incorporating a deep learning element to assess word naming attempts in people with aphasia.

2. Method

2.1. Deep Learning and Recurrent Neural Networks

As a subfield of artificial intelligence, machine learning studies computer algorithms that are able to learn from data how to perform a certain task (Goodfellow et al., 2016). Deep learning is a specific kind of machine learning which allows computational models that are composed of multiple layers to learn representations of data with multiple levels of abstraction, and has dramatically improved the state-of-the-art in speech recognition, visual object recognition, and many other domains such as drug discovery and genomics (LeCun et al., 2015). In its supervised form, these models are trained using data that comes with the ground truth -or labels- of the specific task the algorithm aims to learn. For example, in a phone-recognition task, the algorithm learns from speech signals (the data) where phoneticians have placed manually the phone boundaries as time alignments (the labels). Deep learning models are also called deep neural networks (DNN) as their layers take the form of artificial neural networks. For tasks that involve sequential data, such as speech, it is often better to use recurrent neural networks (RNN): at every time-step, each element of a sequence is processed by a RNN keeping track in their hidden units a ‘state vector’ that implicitly contains information about the history of all past elements of the sequence (LeCun et al., 2015). RNNs can be configured to take into consideration all the future elements of a sequence in addition to the past ones, in which case they are called bidirectional RNNs. However, the recurrent nature of this type of neural network can cause its hidden units to ‘blow up’ or vanish due to error signals flowing backward in time. To overcome such problem, a variation of RNNs was introduced called long short-term memory (LSTM) in which memory cells and gate units are used to construct an architecture that allows for constant error flow (Hochreiter & Schmidhuber, 1997). More recently, a variation of LSTMs was proposed using gate units without memory cells, called gated recurrent units (GRUs) showing comparable performance on speech signal modelling (Cho et al., 2014; Chung et al., 2014).

2.2. NUVA: A Naming Utterance Verifier for Aphasia

Given the scarcity of speech corpora in aphasia, we used a template-based system for picture naming verification. Key to our approach was utilising the framework developed by Ann Lee and James Glass (A. Lee & Glass, 2012) to detect word-level mispronunciations in non-native speech. It uses posteriorgram based pattern matching via a dynamic time warping (DTW) algorithm to compare a word uttered by a native speaker, (teacher), with the same word uttered by a non-native speaker (student). It was designed to be language agnostic. To generate posteriorgrams, our system NUVA replaces their Gaussian Mixture Model trained on unlabelled speech corpora with an acoustic model to yield phone-based posteriorgrams using a deep neural network (DNN) trained on English corpora from healthy speakers. Then, similar to Lee's teacher-versus-student framework, we compare healthy-versus-aphasic utterances. We defined a posteriorgram as a vector of posterior probabilities over phoneme classes in the English language for which we employed the ARPAbet system as used in the BEEP dictionary (Robinson, 1996) consisting of 45 symbols: 44 ARPAbet symbols plus silence. To enable future clinical utility of NUVA, we developed it to run embedded on mobile devices without computationally sophisticated model compression techniques.

DNNs together with DTW has been recently used to solve the task of query-by-example, where a system aims to find the closest matching phrase (Anguera et al., 2015; Ram et al., 2020; Yuan et al., 2020). However, in our task a system just needs to verify that a phrase is what it claims to be.

The following sections describe our system, how speech is processed, how our models were trained, and how our 'utterance-verifier' system NUVA was assembled. We then report the performance of NUVA at classifying eight aphasic patients' spoken picture naming attempts as 'correct' or 'incorrect' compared to SLT classifications of the same responses. SLT classification is seen as the clinical 'gold-standard' approach, and our aim was to develop an automated utterance verification system whose performance could match the 'gold-standard'.

2.3. Signal pre-processing and acoustic modelling

Speech recordings were pre-processed in overlapping frames of 30 milliseconds every 10 milliseconds, and a fast Fourier transform size of 64 milliseconds after a pre-emphasis filter of 0.95 to obtain a vector of 26 acoustic features per frame: 12 Mel-frequency cepstral coefficients (with a final liftering step with a coefficient of 23 applied to them), energy and 13 deltas. See step 1 and 2 in Figure 1.

Figure 1.

Figure 1

From signal to posterior probabilities. Bottom to top: speech signal is fragmented into frames every 10 milliseconds of a window size of 30 milliseconds (1), from each frame a vector of acoustic features is extracted (2) then each vector is fed to a Deep Neural Network (3) which outputs a vector of posterior probabilities or posteriorgram (4).

To train our acoustic model, we used a corpus of British healthy English speakers WSJCAM0 (Robinson et al., 1995). WSJCAM offers phone-level transcriptions using the ARPAbet phone set for British English. We then used Keras deep learning framework (Chollet & others, 2015) with TensorFlow (Martín Abadi et al., 2015) as the back-end. All our models used batch normalisation, a dropout rate of 0.5 and a categorical cross-entropy over 45 classes as the loss function. The training lasted until there was no improvement in accuracy for 50 epochs. We explored several types and configurations of recurrent neural networks and chose our final model as the one with the lowest Phone Error Rate (PER) on the WSJCAM0 test set. Our winning model was a bidirectional Gated Recurrent Unit (Chung et al., 2014) of 128 units and seven layers of depth trained with the Adam optimizer (Kingma & Ba, 2014) resulting in around 2 million parameters and achieving a phone error rate (PER) of 15.85%. See step 3 in Figure 1.

2.4. Comparison of utterances

NUVA uses two recordings from healthy native speakers for each target word, which are transformed into posteriorgrams offline via our DNN, as shown in Figure 1 (steps 1-4). Each naming attempt by an aphasic speaker is transformed into posteriorgrams online using our DNN and then compared to each of the posteriorgrams from the two healthy speakers via the DTW algorithm, see Figure 2. Adapting Lee's notation, (Ann Lee & Glass, 2012) given a sequence of posteriorgrams for the healthy speaker H=(ph1,ph2,,phn,) and the aphasic speaker A=(pa1,pa2,,pam,), a n×m distance matrix can be defined using the following inner product:

φha(i,j)=log(phi*paj) (1)

Figure 2.

Figure 2

NUVA: An utterance verification system for assessing naming attempts in anomia treatment. Given a naming attempt, e.g. target word /tree/, the aphasic's utterance is recorded and processed through our DNN) to generate posteriorgrams (1). The system keeps posteriorgrams of previously recorded healthy speakers' utterances for each target word, (2a and 2b). Posteriorgrams are compared using the DTW) algorithm yielding a distance number between 0 and + (3a and 3b). The minimum of both distances is selected (4) and compared to a set threshold (5) calibrated per speaker, in this example 0.675. If the distance is less than the threshold, then the decision is that the aphasic speaker has said the target word correctly, otherwise it is classified as incorrect.

For such a distance matrix, DTW will search for the path from (1,1) to (n,m) that minimizes the accumulated distance. Different from Lee's work, we used the minimum of the DTW accumulated distances for all comparisons with the two healthy speakers to make a final decision.

Many biometric systems of verification utilise a threshold value to make a decision (Malik et al., 2014). In NUVA if the value is smaller than the set threshold then the aphasic's naming attempt is deemed similar enough to that of the healthy speakers and is considered 'correct'; otherwise, it is classified as 'incorrect', see Figure 2. The existence of such threshold is functionally analogous to the β penalizing parameter in Abad (Abad et al., 2013): calibration of both controls the rate of false positives and false negatives for each patient. However, they are intrinsically different within each system.

The reason for using the minimum of DTW distances between the healthy speakers can intuitively be justified as choosing the healthy speaker's utterance that is closest aligned to the aphasic speaker. The number of healthy speakers could be increased to cater for a broader range of accents, offering a closer regional match to each aphasic speaker. Figure 3 illustrates this rationale quantitatively using Receiver Operating Characteristic (ROC) curves and Area Under the Roc Curve (AUC) scores. Choosing the minimum of the DTW distances between the healthy speakers yields the best fit and highest AUC score to the scorings provided by a speech and language therapist (SLT) used in this study. We discuss this further in the next section.

Figure 3.

Figure 3

ROC curves for SLT scoring. Each curve represents the best fit to SLT scorings for a given criteria of combining DTW distances: (i) choosing one healthy speaker, (ii) the minimum distance between healthy speakers, (iii) the maximum and (iv) the average.

2.5. Threshold Calibration

NUVA requires a set threshold to make a decision to determine whether an attempt is ‘correct’ or ‘incorrect’. Previous naming data from a PWA as assessed by their clinical SLT could be used for such purposes. In real life, a PWA is first assessed by a SLT to determine their diagnosis and level of speech and language impairment before being referred to a specific treatment program. Speech performance on picture naming tasks is a core part of the standard clinical SLT aphasia assessment, so it is reasonable to assume the existence of previously labelled (by a SLT) speech data may be available to calibrate NUVA's threshold. Two methods used to calibrate the most adequate threshold given previously held data are:

2.5.1. Adapted per patient

The threshold is calibrated using previously held data from a certain patient. In this case, every new patient will need to have the threshold calibrated according to the data used in a previous assessment.

2.5.2. Fixed

The threshold is calibrated using previously held data from various patients and then is fixed for any new patient.

3. Experiment and data

3.1. Healthy speakers corpus: WSJCAM0

The underlying DNN of our system was trained using speech from healthy English speakers. We used WSJCAM0, a corpus of spoken British English derived from the Wall Street Journal text corpus specifically designed for the construction and evaluation of speaker-independent speech recognition systems (Robinson et al., 1995). The database has recordings of 140 speakers each speaking about 110 utterances. The corpus is partitioned into 92 training speakers, 20 development test speakers and two sets of 14 evaluation test speakers. It offers word and phone based time alignments on verified utterance transcriptions and employs ARPAbet as a phone set. WSJCAM0 is currently distributed by the Linguistic Data Consortium.

3.2. Participants

Eight native English speakers, six male, with chronic post-stroke anomia, were recruited. Demographics are shown in Table 1 below. Inclusion criteria were aphasia in the absence of speech apraxia (severe motor speech impairment) as evidenced by: (i) impaired naming ability on the object naming subtest of the Comprehensive Aphasia Test, CAT (Swinburn et al, 2004); scores below < 38 are classified as impaired; (ii) good single word repetition on the CAT subtest; normative cut-off>12. All patients gave written consent, and data was processed in accordance with current GDPR guidelines. Ethical approval was granted by NRES Committee East of England – Cambridge, 18/EE/228.

Table 1.

Demographic and clinical data for the patient.

Patient ID Sex Age Months post-stroke CAT Object naming CAT Repetition
P1 M 65 108 32 19
P2 M 58 90 19 22
P3 M 70 91 10 28
P4 F 62 21 28 24
P5 M 64 14 6 25
P6 M 59 98 30 31
P7 M 57 109 27 24
P8 F 82 38 29 23

Mean(IQR) 65(8) 71(67) 23(13) 25(3)

Max score possible (/48) (/32)
Cut-off used <38 >12

3.3. Stimuli

Picture naming stimuli consisted of 220 coloured drawings. They were selected from the top 2000 most frequent words using the Zipf index of the SUBTLEX-UK corpus (van Heuven et al., 2014) keeping the same distribution of parts of speech for nouns, verbs, adverbs and adjectives. See Appendix C for a list of the items selected.

3.4. Dataset Collection

We used a tailor-made gamified picture naming treatment App developed in Unity on an Android tablet Samsung SM-T820 to deliver the picture stimuli and record the patients' speech responses. Patients' speech recordings were collected using a Sennheiser headset SC 665 USB at 16 kHz which were then stored in a compliant WAVE-formatted file using a 16 bit PCM encoding.

Patients were instructed to name each item presented on screen as quickly and accurately as possible using a single word response. They were given up to 6 seconds to complete each picture naming attempt. The item presentation order was randomised across patients. A SLT was present throughout the assessment and scored the naming responses online in a separate file without giving the patient any performance feedback. A total of 1760 speech recordings (220 words x 8 patients) were acquired.

3.5. Procedure

A SLT classified all naming attempts online while patients were in session with the treatment App. After a further revision, six categories were used to classify the attempts using the transcriptions and notes taken during the online session: "Correct", "No Response", "Filler", "Phonological Error", "Circumlocution" and "Other", in addition to “Correct” or “Incorrect” as a final decision. These six categories were used a posteriori to manually detect speech sections within each recording that were used by the SLT to make a decision. When patients produced multiple utterances in a naming attempt, only one utterance representative of the online verdict by the SLT of such naming attempt was segmented to construct a single-utterance recording. For example, when a patient's response was scored as 'Filler', and the corresponding recording comprised of multiple 'um', 'ah', 'eh', only one of those utterances was selected and segmented to create a single-utterance naming attempt per item. These single-utterance recordings were the data used to evaluate NUVA and the baseline. Each naming attempt was then relabelled as 'correct' or 'incorrect' accordingly, and this last classification was used as the ground truth to evaluate NUVA's performance and baseline. Figure 4 describes the dataset and each of the patient's naming performance. Performances were calculated using 'Scikit-learn' (Pedregosa et al., 2011) and significance testing was implemented using 'MLxtend' (Raschka, 2018) and 'scikit-posthocs' (Terpilowski, 2019) packages.

Figure 4.

Figure 4

Each patient's naming performance on the 220 items of the picture naming test, as classified by a speech and language therapist (SLT).

3.5.1. Inter-rater reliability and agreement

A ground truth is always required to evaluate the performance of an automatic system. In this study, the ground truth was defined as a native English, speaking speech and language therapist's (SLT) online scoring of the PWAs’ naming attempts. The goal being for the automatic system to be as good as the SLT expert. However, as with all human systems scores might not be identical across individual SLTs. For example, SLTs might vary in determining whether derivatives of the target word, such as singularisation, pluralisation, nominalisation or verbalisation, are classified as correct responses. For this reason, a second native English speaking SLT (SLT2), independently scored all patients' naming attempts offline in addition to the primary SLT who scored them online. In this case, agreement overall between SLT experts is expected to be high, with different scores occurring only on a small number of responses reflecting variance in scoring criteria rather than errors per se. To obtain a measure of inter-rater reliability and account for such measurement noise we used Gwet's AC1 agreement coefficient (Gwet, 2014) i.e., an agreement coefficient corrected by chance.

We then combined this with McHugh's benchmark range for a more stringent interpretation of agreement coefficients as recommended for applications in health care and clinical research (McHugh, 2012) i.e., <.20: No agreement; .21-.39: Minimal; .40-.59: Weak; .60-.79: Moderate; .80-.90: Strong; Above .90: Almost perfect agreement. This was calculated in a series of steps detailed in the Appendix B.2. First, we calculated the probability that the statistic AC1 falls into each of the intervals of the benchmark levels resulting in a membership probability for each interval. Second, we computed the cumulative probability for each interval starting from the highest level of agreement down to the lowest. Third, we took the interval associated with the smallest cumulative probability that exceeds 95% in descending order of agreement as the final benchmark level. The advantage of using this method is that it gives us 95% certainty of the level of agreement reached among the SLT raters. Final level of inter-rater reliability and agreement between both SLTs is summarised in Table 2.

Table 2.

Inter-rater reliability information for the two SLTs engaged in this study. Showing the Gwet's AC1 agreement coefficient statistic, its 95% confidence interval, level of agreement reached and percentage agreement between the SLTs.

Patient PAa Gwet's AC1 StdErrb 95% CIc Level of Agreement
P1 99.55% 0.992 0.008 (0.976, 1.008) Almost Perfect
P2 95.91% 0.925 0.025 (0.877, 0.973) Strong
P3 97.73% 0.972 0.012 (0.948, 0.996) Almost Perfect
P4 97.27% 0.950 0.020 (0.911, 0.990) Almost Perfect
P5 97.27% 0.963 0.015 (0.933, 0.992) Almost Perfect
P6 95.45% 0.913 0.027 (0.860, 0.966) Strong
P7 92.27% 0.850 0.035 (0.781, 0.918) Moderate
P8 96.82% 0.943 0.021 (0.902, 0.985) Almost Perfect
All 96.53% 0.938 0.008 (0.923, 0.953) Almost Perfect
a

PA = percentage agreement

b

StdErr = standard error

c

CI = confidence interval

As predicted Inter-rater reliability between SLTs was high for both Gwet's AC1 (mean: 0.938; range 0.85-0.992) and percentage agreement (mean: 96.5%; range 92.3-99.5), with strong to almost perfect agreement.

Since both, percentage agreement and Gwet's AC1 are very high, we report results using the primary SLT who scored the attempts online as the ‘gold-standard’ and ground truth for all subsequent analyses with subsequent offline systems. Performance of all offline systems, i.e., the second SLT (SLT2), NUVA and the baseline ASR system were calculated across all reported metrics (Accuracy, False Positives, False Negatives, F1-score and Pearson's r).

3.5.2. ASR Baseline

Since there are no existing baselines that we could use to compare the performance of our system, we used an off-the-shelf standard speech recognition service. The rationale in doing so, is that it would be the fastest way for a clinical lab to implement automatic digital therapy for word naming.

We used Google Cloud Platform speech-to-text service configured with British English (the date used: 24/3/20), and adapted it for our task to create a baseline. As the service provides transcriptions, if the target word was found in the transcript, then the attempt was classified as 'correct', otherwise 'incorrect'. For each aphasic patient's naming attempt, the same recording to test our system was used to test the baseline.

4. Results

4.1. System Performance

As indicated in section 2.5 Threshold Calibration, NUVA utilises a set threshold to make a final decision on marking a patient's naming attempt as either 'correct' or 'incorrect' and, such set threshold can be calibrated using two methods: ‘fixed’ and ‘adapted’, using previously held data. In our study, we used all the available data as held data per each calibration method to have a notion of the best possible performance achievable by NUVA in ideal conditions. With the ‘fixed’ method, we set the same threshold for all patients to estimate the best possible performance across all patients involved in the study. With the ‘adapted’ method we set a different threshold for each patient to estimate the best possible performance for each patient. Figure 5 illustrates the performance for all patients across a continuum set of thresholds using Pearson's r as a metric. Figure 6 shows the same performance measure across the same continuum set of thresholds but for each patient. The resulting best thresholds are summarised in Table 3. Thresholds used for each version of NUVA (i) 'fixed' and (ii) 'adapted'.

Figure 5.

Figure 5

Performance (y-axis) per threshold (x-axis) across all patients (as a group) using Pearson's r as a metric. Y-axis = closer to 1, the better the performance.

Figure 6.

Figure 6

Performance (y-axis) per threshold (x-axis) per patient individually using Pearson's r as a metric. Y-axis = closer to 1 the better the performance

Table 3.

Thresholds used for each version of NUVA (i) 'fixed' and (ii) 'adapted'.

System Version fixed adapted
Data All Patients P1 P2 P3 P4 P5 P6 P7 P8
Best Threshold 0.680 0.670 0.920 0.680 0.420 0.860 0.840 0.670 0.690

4.2. Performance across all patients

Overall performance of NUVA for both versions, 'fixed' and 'adapted', are compared to a commercial baseline (Google STT) and a second independent SLT scorer, SLT2. Results are presented in Table 4.

Table 4.

Performance across all patients of (i) baseline (Google STT service), (ii) 'fixed' version of NUVA (iii) 'adapted' version of NUVA, (iv) a second independent SLT, SLT2, against the primary SLT scorer used in this study.

System Accuracy False Positives False Negatives F1-Score Pearson's r
baseline 0.882 0.020 0.098 0.795 0.727
fixed 0.905 0.049 0.046 0.855 0.784
adapted 0.913 0.057 0.031 0.871 0.807
SLT2 0.965 0.016 0.018 0.947 0.921

NUVA yields better performance than the baseline for all metrics, except for False Positives rates. These differences are significant with respect to the baseline, as shown in Table 5. Notice that we did not find a significant difference in performance between the 'fixed' and 'adapted' version of NUVA.

Table 5.

Significance testing across all patients (as a group) using Dunn's post hoc test (with Holm correction) between systems’ predictions and scores of a second independent SLT, SLT2.

Pairs All Patients
baseline-fixed *
baseline-adapted ⁎⁎
baseline-SLT2 ⁎⁎⁎
fixed-adapted NS
fixed-SLT2 ⁎⁎⁎
adapted-SLT2 ⁎⁎⁎

Notation:

p < 0.05,

⁎⁎

p < 0.01,

⁎⁎⁎

p < 0.001 and NS, non-significant.

In general, since scoring criteria between SLTs, which conforms the 'gold-standard', might differ slightly, we included the performance of a second independent SLT scorer (SLT2) with respect to the primary SLT scorer used in this study to illustrate the difference between the performance of all three computerised systems, and human SLT scores. Both baseline and NUVA, are significantly inferior to the performance of the second independent SLT scorer.

We ran Cochran's Q test (Cochran, 1950) to determine if the binary predictions of baseline, 'fixed' and 'adapted' versions of NUVA and also the SLT2 binary scores were statistically significantly different across all patients (p<0.001). We then performed pairwise comparisons using Dunn's procedure (Dunn, 1964) with Holm correction (Holm, 1979) for multiple comparisons, see Table 5.

4.3. Performance per Patient

Each system's performance per patient is illustrated in Figure 7, together with a second independent SLT scorer, SLT2 (see Appendix A for details). Significance testing follows the same methodology from the previous section but applied to each patient. See significance testing in Table 6.

Figure 7.

Figure 7

Comparison of performance between (i) a commercial baseline (ii) the 'fixed' version of NUVA, (iii) the 'adapted' version of NUVA (iv) a second independent SLT, SLT2, against the primary SLT used in this study. Y-axis = closer to 1, the better the performance.

Table 6.

Significance testing per patient using Dunn's post hoc test (with Holm correction) between systems’ predictions and scores of a second independent SLT, SLT2.

Pairs P1 P2 P3 P4 P5 P6 P7 P8
baseline-fixed NS NS NS NS NS * * ⁎⁎
baseline-adapted NS NS NS NS NS ⁎⁎ * ⁎⁎
baseline-SLT2 ⁎⁎ ⁎⁎ NS NS NS ⁎⁎ ⁎⁎⁎ ⁎⁎⁎
fixed-adapted NS NS NS NS NS NS NS NS
fixed-SLT2 NS ⁎⁎ * NS ⁎⁎ NS NS NS
adapted-SLT2 NS ⁎⁎ * NS * NS NS NS

Notation:

p < 0.05,

⁎⁎

p < 0.01,

⁎⁎⁎

p < 0.001 and NS, non-significant.

At least one version of NUVA had better performance than the baseline system for 7/8 patients, but not patient P4. These differences were statistically significant for 3/8 patients, see P6, P7 and P8. There was no significant difference between the 'fixed' and 'adapted' versions of NUVA. Performance of NUVA and SLT2 were not significantly different for 5/8 patients. NUVA performed worse for patients P2, P3 and P5.

In terms of false positives and false negatives, NUVA tended to have less false negatives than false positives (Figure 8). The baseline system, in contrast, had less false positives and more false negatives.

Figure 8.

Figure 8

False Positives and False Positives for (i) baseline, (ii) 'fixed' version of NUVA, (iii) 'adapted' version of NUVA and (iv) a second independent SLT, SLT2. Y-axis = less is better.

4.4. Extent of agreement between systems and SLTs

To ease the visualisation of results we used a numerical coding system to report the extent of agreement with 95% certainty using Gwet's benchmarking method (Gwet, 2014) using the range proposed by McHugh for healthcare and clinical research (McHugh, 2012), see Table 7 and Table 8. Calculations to obtain Table 8 can be found in Appendix B.2 using inter-rater reliability measures from Appendix B.1.

Table 7.

Benchmark range proposed by McHugh as agreement interpretation for applications in health care and clinical research. Code column is used as a short reference to the agreement description.

Benchmark Range Description Code
above .90 Almost Perfect 6
.80-.90 Strong 5
.60-.80 Moderate 4
.40-.59 Weak 3
.21-.39 Minimal 2
below .20 None 1

Table 8.

Extend of agreement between systems and each SLT and, also between SLTs themselves at the bottom, calculated with a 95% certainty using Gwet's benchmarking method (Gwet, 2014) .

Agreement between P1 P2 P3 P4 P5 P6 P7 P8 All
Baseline & SLT1 5 4 5 5 5 4 3 4 4
Baseline & SLT2 5 4 5 5 5 4 3 4 4
Fixed & SLT1 5 4 5 5 5 5 4 5 5
Fixed & SLT2 5 4 5 5 4 5 4 5 4
Adapted & SLT1 5 4 5 4 4 5 4 5 5
Adapted & SLT2 5 4 5 5 5 5 4 5 4
SLT1 & SLT2 6 5 6 6 6 5 4 6 6

Across patients, none of the ASR systems reached human SLT level of agreement. However, ‘fixed’ and ‘adapted’ versions of NUVA reached the closest i.e., just one level below SLT1, the ‘gold-standard’ in this study, and two levels below SLT2. The baseline ASR system reached two levels below SLT expert regardless the SLT used.

In general, the ‘fixed’ version of NUVA reached closer agreement levels to SLTs per patient using the ‘gold-standard’ SLT1 and, the ‘adapted’ version using SLT2. However, this difference is only evident in patients 4 and 5.4.5 System Cross-validation

To assess the predictive performance of NUVA and to judge how it performs outside the sample to a new unseen dataset, we used cross-validation. The assumption, in this case, was that previously collected speech samples from patients could be used to optimise the system's classifying threshold. This assumption is consistent with current practices as a patient always undertakes an initial assessment with an SLT before starting therapy; this assessment could provide the speech samples needed to calibrate NUVA's threshold. For each patient, a 10-fold cross-validation procedure was applied, and the average performance across folds is reported, together with minimum, maximum and range, see Table 9. Accuracies for all patients was high, above 84% with a range of 10% and a group average of 89.5%

Table 9.

Results for a 10-fold cross-validation for each patient of the 'adapted' version of NUVA. For each patient, the average across all folds is reported as Mean (±SD).

Patient Accuracy False Positives False Negatives F1-Score Pearson's r
P1 0.93(±0.068) 0.04(±0.045) 0.03(±0.046) 0.89(±0.106) 0.85(±0.149)
P2 0.84(±0.082) 0.11(±0.074) 0.05(±0.052) 0.78(±0.116) 0.67(±0.162)
P3 0.88(±0.055) 0.08(±0.029) 0.04(±0.038) 0.51(±0.247) 0.46(±0.278)
P4 0.94(±0.055) 0.03(±0.030) 0.04(±0.040) 0.89(±0.088) 0.85(±0.123)
P5 0.87(±0.060) 0.09(±0.059) 0.04(±0.032) 0.61(±0.247) 0.56(±0.261)
P6 0.93(±0.071) 0.05(±0.057) 0.03(±0.030) 0.91(±0.104) 0.85(±0.150)
P7 0.87(±0.081) 0.10(±0.059) 0.03(±0.041) 0.90(±0.065) 0.72(±0.183)
P8 0.90(±0.038) 0.05(±0.041) 0.05(±0.043) 0.85(±0.067) 0.79(±0.087)
Mean(SD) 0.895(0.03) 0.066(0.03) 0.039(0.01) 0.790(0.14) 0.718(0.14)
Min 0.836 0.027 0.027 0.506 0.462
Max 0.936 0.114 0.05 0.905 0.852
Range 0.1 0.086 0.023 0.399 0.389

4.5. Latency in systems' response

We assessed the baseline and NUVA's performance in terms of speed to label 'correct'/‘incorrect' a given naming attempt and speed in processing a second of speech. We used a Samsung tablet, SM-T820 model, running on Android 7 (SDK 24) to test NUVA's performance, as it was adapted to work embedded on a mobile platform. The baseline (Google Cloud Platform Speech-to-Text service) used a client-server model with gRPC protocol. Figure 9 and Table 10 illustrate for both systems (a) the speed to process 1760 naming recordings and (b) the average time to process one second of speech. In total, the baseline system had 37 outliers above 2.5 seconds. These have been clamped to 2.5 seconds in the visualisation but not in the calculations.

Figure 9.

Figure 9

Timing measurements (latency) in providing labels' correct'/’incorrect' for the baseline (Google STT service) and NUVA (embedded). (a) Time in seconds to process each of the naming attempts in our test. (b) Average time in seconds to process a second of speech. Y=Time in seconds, less is better.

Table 10.

Statistics of timing measurements (latency) of (i) baseline and (ii) NUVA in providing feedback. (a) Time to process each of the naming attempts in our test. (b) Average time to process a second of speech. The less, the better.

(a) (b)
Time (s) baseline NUVA Time (s) baseline NUVA
Mean(±SD) 0.825(±0.362) 0.605(±0.170) Mean(±SD) 0.571(±0.255) 0.404(±0.058)
Min 0.481 0.227 Min 0.247 0.252
Max 10.659 2.207 Max 6.513 0.851
Range 10.178 1.980 Range 6.266 0.599

Using a paired Wilcoxon Signed Rank Test, performance speed was statistically significantly different between the two systems (p<0.001). On average NUVA was 1.42 times faster than the baseline system in processing each naming attempt. Since naming attempts in the test set varied in length, this translated in NUVA taking 0.404 seconds on average to process a second of speech while the baseline took 0.571 seconds.

5. Discussion

This study aimed to evaluate the effectiveness of NUVA, a tailor-made ASR system to assess spoken picture naming performance in a group of eight chronic aphasic stroke patients. Results show that NUVA had within sample performance accuracies above 85% across all patients, irrespective of the thresholding approach used, and reached comparable agreement with the therapist's ratings in two of the patients. A cross-validation analysis found accuracy remained high, above 84% for all patients, with an average of 89.5% overall. This demonstrates the feasibility of NUVA to reliably discriminate 'correct' from 'incorrect' spoken naming attempts in aphasic speech.

5.1. Baseline vs. NUVA

To offer the reader a benchmark as to what level of performance can be achieved with aphasic speech data using standard, ‘off-the-shelf’, publicly available STT ASR systems we compared the performance of our tailor-made ASR system NUVA to a commercial baseline (Google Cloud Platform STT service).

NUVA performed significantly better overall in terms of both accuracy and speed. Furthermore, only the NUVA system reached human expert SLT level of performance agreement, albeit only in two patients. Taken together these data demonstrate that a tailor-made ASR technology for aphasic speech such as NUVA is feasible and may be an effective solution for accurate, reliable and fast automatic, digitally delivered speech feedback system for aphasia interventions targeting spoken word naming performance.

5.1.1. Accuracy

In terms of accuracy, NUVA performed well overall when compared to the SLT 'gold standard’. The three patients (2, 3 and 5) where NUVA performed significantly worse with respect SLT2 scorings (Table 9), were the more severely speech impaired, i.e., the lowest performers on the Object Naming subtest of the Comprehensive Aphasia Test (Swinburn et al., 2004). These results show that NUVA while promising, needs further improvement to reach a comparable human/SLT level in more severely impaired aphasic speakers. Similar performance decay is displayed by the baseline in patients 7 and 8 in addition to 2, 3 and 5, see Appendix A.

The same recordings were used by the baseline and NUVA systems and, both systems used British English corpora for training, however only the NUVA system was designed to accommodate the speech profile of PWA. This highlights both the need and potential to use tailor-made ASR technology to reach expert human level performance, (SLT), when processing speech data from clinical populations with aphasia.

5.1.2. Agreement between systems and SLTs

Percentage agreement and inter-rater reliability, as indexed by Gwet's AC1, was consistently high among SLTs reaching “Almost Perfect” levels of agreement for 5/8 patients. While patients 2, 6 and 7, yielded “Strong”, “Strong” and “Moderate” agreement levels, respectively.

The NUVA system, ‘fixed’ or ‘adapted’, reached the same level of agreement as the SLTs for patients 6 and 7 with 95% certainty. In contrast the baseline system did not reach human SLT agreement levels for any patient. Taken together these results, albeit in a small patient sample suggest creating tailor-made ASR technology specifically for aphasic speech has the potential to reach human expert SLT level of performance on naming tasks.

5.1.3. Latency on Feedback

An essential feature of an automated ASR system for verifying spoken naming attempts is its ability to provide a timely response online, akin to an online SLT rater. When such as system is to be used in conjunction with a treatment intervention e.g., an anomia treatment program for PWA the speed in providing an automatic answer has direct implications in the efficiency of the treatment design. For example, the faster the ASR performance and feedback to the PWA the greater the number of items that can be treated within a given time window allowing in turn more fine-grained experimental investigation of the impact of dose and intensity on speech production recovery.

We assessed latency response times for both the baseline and NUVA systems to illustrate what level of performance is currently possible with ‘off-the-shelf’ ASR software and our tailor-made tool. In this regard there are significant differences between both systems: the baseline relies on a client-server model with lags due to network communications, NUVA is embedded and its performance depends on the hardware it is deployed into. Nevertheless, NUVA the utterance verifier in a mobile device was significantly faster in providing feedback for each naming attempt than the client-server model used by Google STT service powered by its high-performance RPC protocol, gRPC. On average, NUVA processed each naming attempt 1.42 times faster than the baseline system. NUVA had an average latency of 0.404 seconds for a second of speech resulting in latencies of 0.808, 1.01 and 1.212 seconds to process naming attempts of lengths 2, 2.5 and 3 seconds long respectively; the baseline, 1.14, 1.425 and 1.71 seconds respectively. Future work will investigate whether (i) these NUVA latency results are acceptable when deployed as a real-time performance feedback system in PWA and (ii) performance speed can be optimised further using quantization techniques for training and inference or simply acquiring more powerful hardware.

5.2. NUVA Performance Accuracy

In the literature, to date, two key studies have evaluated ASR systems' performance on aphasic speakers' word naming attempts (Abad et al., 2013; Ballard et al., 2019). The level of heterogeneity between the three studies (including ours) is high (different languages spoken, types of aphasia, level of impairment, vocabulary assessed) nevertheless, NUVA does appear to offer both a more accurate and less variable performance. Abad and Ballard report a mean accuracy performance of their systems of 82% (range: 24%) and 75% (range: 18%) respectively. In contrast, NUVA reached a performance of 89.5% with a range of just 10% across patients (see Table 11 for details). However, Ballard's results are arguably more reliable as their system's performance is assessed online.

Table 11.

Comparison of NUVA results with two key studies in the literature that have used ASR to assess spoken naming attempts in aphasia.

Study Abad et al. 2013 Ballard et al. 2019 NUVA
Language Portuguese Australian English British English
Population Aphasia Aphasia/Apraxia Aphasia
N 8 5 8
Stimuli 103 124 220
Word Productions 824 7069 1760
Interrater Reliability Agreement (method) >0.85 (point-to-point agreement) 0.96 (point-to-point agreement) 0.938 (Gwet's AC1)
Accuracy 0.820 0.75 0.895
Min 0.690 0.651 0.836
Max 0.930 0.828 0.936
Range 0.240 0.177 0.100
Evaluation Method 10-Fold CV Online 10-Fold CV

5.3. Limitations

Our results, including the high cross-validation performance data, while encouraging, are perhaps best seen as a proof of concept at this time. The patient sample is small, and further validation in a larger sample of patients is required. To address this, NUVA is being used within a clinical trial of a novel anomia treatment App - iTALKbetter. The aim here is for NUVA to assess online 'live' naming attempts so that aphasic patients' speech performance can be used to drive therapy progression, e.g., when a patient names an item reliably three times, they no longer need to practice it, and they progress to practising new words and/or the ASR gives live feedback on a trial by trial basis to the patient and encourages their ongoing participation. In this context, any mistake of our system, i.e., a misclassification of speech attempts, could affect a patient's recovery. False positives could reinforce error-full responses when actual 'incorrect' attempts are understood as 'correct', while false negatives could frustrate patients when their actual 'correct' attempts are misclassified as 'incorrect', possibly leading them to abandon the intervention. For NUVA, while 'adapted' version gave the best accuracy performance, the 'fixed' version had on average a lower rate of false positives. As such, this is the solution we have deployed in the clinical trial of iTALKbetter. Nevertheless, we are mindful that for individual patients, there was variance in the range of false negatives (e.g., P2 and P5 was 9%) and false positives (e.g., P7 -10%). Further qualitative testing will investigate if such misclassification rates are acceptable or not to people with aphasia and clinicians.

5.4. Next steps

A significant benefit in the approach we used to construct NUVA is that it relies on recordings of the target word by healthy subjects. Healthy speech is considerably easier to obtain than aphasic speech meaning vocabulary of target words (as long as we can find recordings for them) could be enlarged indefinitely without the need for retraining models. We also predict the system's performance can be improved by expanding the variety of healthy voices for each target word, e.g., to include different accents.

Future steps will be oriented towards designing and developing a system that can offer a higher degree of information to a clinician other than just ‘correct’ or ‘incorrect’ performance. For example, if a PWA's naming attempt was classified as incorrect, it may be helpful for the clinician to understand if there was a pattern to the PWA's type of naming errors. Such as, did the PWA make a spoken naming attempt at all or were their errors mostly phonemically related to the target word e.g., /cap/ instead of /cat/ or semantically related e.g., /dog/.

6. Conclusions

We present here NUVA, a tailor-made utterance verification system based on a deep learning architecture to assess word naming attempts for people with aphasia automatically. In a sample of eight aphasic patients' 1760 naming attempts, NUVA performed significantly better than the commercial baseline (Google STT service). Given the scarcity of aphasic speech corpora, this represents a significant step towards creating a reliable and automatic spoken word assessment system for aphasic speakers. It also offers clinicians a deployable solution for gathering big data, optimisation of similar systems and further research.

Future work will focus on optimising the effectiveness of NUVA to deliver live (accurate and fast) speech performance feedback within a clinical trial using a digitally delivered anomia treatment App: iTALKbetter. NUVA is available open-source to encourage reproducibility and further development in this field; we welcome new insights and collaborations. Despite being initially trained in English, given the language-agnostic framework of our system, it would be interesting to see if it can be used in other languages. This would offer an invaluable tool for aphasic speakers of under-researched languages.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgements

DB and the ASR technical development is funded by a Medical Research Council iCASE PhD studentship, award number 1803748, JC is supported by a Wellcome Trust Senior Research Fellowship in Clinical Science (106161/Z/14/Z), and APL, VF, EU, HCF and CD by an NIHR Research Professorship.

Appendix A

Performance per patient of (i) baseline (Google STT), (ii) 'fixed' version of NUVA, (iii) 'adapted' version of NUVA and (iv) a second independent SLT scorer, SLT2, against the primary SLT scorer used in this study.

Image 15

Appendix B.1

Measures of inter-rater reliability, percentag agreement (PA) and Gwet's AC1 agreement coefficient, together with standard error (StdErr) and 95% confidence interval (CI) between systems and each speech and language therapist (SLT1 and SLT2).

.

Appendix B.2

Agreement between systems and each speech and language therapist (SLT1 and SLT2) using Gwet's benchmarking method (Gwet, 2014) on McHugh's benchmark range (McHugh, 2012).

Interval Membership Probability is defined as IMP=P(KGbSEZKGaSE), where KGis Gwet's AC1, SE is the standard error and interval (a,b) determine an agreement level in a benchmark range. See Gwet's AC1 and standard errors between systems and each SLT in Appendix B.1.

.
.

Appendix C

List of the 220 words selected as picture naming stimuli for this study based on the top 2000 most frequent words using the Zipf index of the SUBTLEX-UK corpus (van Heuven et al., 2014) keeping the same distribution of parts of speech for nouns, verbs, adverbs and adjectives. Included are the main descriptive attributes for each word: British English phonetic transcription based on the International Phonetic Alphabet (British Pronunciation) (Cambridge Dictionary| English Dictionary, Translations & Thesaurus, 2020), number of characters (Ch), number of phonemes in UK English (Ph), number of syllables (Sy), the Zipf index of the SUBTLEX-UK corpus (Zipf) (van Heuven et al., 2014), Part of Speech (PoS), Age of Acquisition (AoA) (Kuperman et al., 2012) and Concreteness (Con.) (Brysbaert et al., 2014).

.
.
.
.

References

  1. Abad A., Pompili A., Costa A., Trancoso I., Fonseca J., Leal G., Farrajota L., Martins I.P. Automatic word naming recognition for an on-line aphasia treatment system. Computer Speech & Language. 2013;27(6):1235–1248. doi: 10.1016/j.csl.2012.10.003. https://doi.org/ [DOI] [Google Scholar]
  2. Akbarzadeh-T M.-R., Moshtagh-Khorasani M. A hierarchical fuzzy rule-based approach to aphasia diagnosis. Journal of Biomedical Informatics. 2007;40(5):465–475. doi: 10.1016/j.jbi.2006.12.005. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  3. Anguera X., Rodriguez-Fuentes L.-J., Buzo A., Metze F., Szöke I., Penagarikano M. 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2015. QUESST2014: Evaluating Query-by-Example Speech Search in a zero-resource setting with real-life queries; pp. 5833–5837. https://doi.org/ [DOI] [Google Scholar]
  4. Ballard K.J., Etter N.M., Shen S., Monroe P., Tan C.T. Feasibility of Automatic Speech Recognition for Providing Feedback During Tablet-Based Treatment for Apraxia of Speech Plus Aphasia. American Journal of Speech - Language Pathology (Online); Rockville. 2019;28(2S):818–834. doi: 10.1044/2018_AJSLP-MSC18-18-0109. http://dx.doi.org. libproxy.ucl.ac.uk/10.m44/2018_AJSLP-MSC18-18-0m9. [DOI] [PubMed] [Google Scholar]
  5. Bhogal S.K., Teasell R.W., Foley N.C., Speechley M.R. Rehabilitation of Aphasia: More Is Better. Topics in Stroke Rehabilitation. 2003;10(2):66–76. doi: 10.1310/RCM8-5TUL-NC5D-BX58. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  6. Brady M.C., Kelly H., Godwin J., Enderby P., Campbell P. Speech and language therapy for aphasia following stroke. Cochrane Database of Systematic Reviews. 2016;6 doi: 10.1002/14651858.CD000425.pub4. https://doi.org/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Brysbaert M., Warriner A.B., Kuperman V. Concreteness ratings for 40 thousand generally known English word lemmas. Behavior Research Methods. 2014;46(3):904–911. doi: 10.3758/s13428-013-0403-5. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  8. Cambridge Dictionary | English Dictionary, Translations & Thesaurus. (2020). https://dictionary.cambridge.org/
  9. Cho, K., van Merrienboer, B., Bahdanau, D., & Bengio, Y. (2014). On the Properties of Neural Machine Translation: Encoder-Decoder Approaches. ArXiv:1409.1259 [Cs, Stat]. http://arxiv.org/abs/1409.1259
  10. Chollet, F., & others. (2015). Keras. https://keras.io
  11. Chung, J., Gulcehre, C., Cho, K., & Bengio, Y. (2014). Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. ArXiv:1412.3555 [Cs]. http://arxiv.org/abs/1412.3555
  12. Cmusphinx/pocketsphinx. (2020). [C]. cmusphinx. https://github.com/cmusphinx/pocketsphinx (Original work published 2014)
  13. Cochran W.G. The Comparison of Percentages in Matched Samples. Biometrika. 1950;37(3/4):256–266. doi: 10.2307/2332378. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  14. Code C., Heron C. Services for aphasia, other acquired adult neurogenic communication and swallowing disorders in the United Kingdom, 2000. Disability and Rehabilitation. 2003;25(21):1231–1237. doi: 10.1080/09638280310001599961. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  15. Darley Frederic L., Aronson Arnold E., Brown Joe R. Differential Diagnostic Patterns of Dysarthria. Journal of Speech and Hearing Research. 1969;12(2):246–269. doi: 10.1044/jshr.1202.246. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  16. Dunn O.J. Multiple Comparisons Using Rank Sums. Technometrics. 1964;6(3):241–252. doi: 10.2307/1266041. https://doi.org/ [DOI] [Google Scholar]
  17. Forbes M.M., Fromm D., MacWhinney B. AphasiaBank: A Resource for Clinicians. Seminars in Speech and Language. 2012;33(03):217–222. doi: 10.1055/s-0032-1320041. https://doi.org/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Goodfellow I., Bengio Y., Courville A. MIT Press; 2016. Deep Learning. [Google Scholar]
  19. Gwet K.L. Handbook of Inter-Rater Reliability, 4th Edition: The Definitive Guide to Measuring The Extent of Agreement Among Raters. Advanced Analytics. 2014:LLC. [Google Scholar]
  20. Hochreiter S., Schmidhuber J. Long Short-Term Memory. Neural Computation. 1997;9(8):1735–1780. doi: 10.1162/neco.1997.9.8.1735. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  21. Holm S. A Simple Sequentially Rejective Multiple Test Procedure. Scandinavian Journal of Statistics. 1979;6(2):65–70. [Google Scholar]
  22. Wade J., C. R., Petheram B. Voice recognition and aphasia: Can computers understand aphasic speech? Disability and Rehabilitation. 2001;23(14):604–613. doi: 10.1080/09638280110044932. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  23. Kingma, D. P., & Ba, J. (2014). Adam: A Method for Stochastic Optimization. ArXiv:1412.6980 [Cs]. http://arxiv.org/abs/1412.6980
  24. Kuperman V., Stadthagen-Gonzalez H., Brysbaert M. Age-of-acquisition ratings for 30,000 English words. Behavior Research Methods. 2012;44(4):978–990. doi: 10.3758/s13428-012-0210-4. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  25. Le D., Licata K., Mower Provost E. Automatic quantitative analysis of spontaneous aphasic speech. Speech Communication. 2018;100:1–12. doi: 10.1016/j.specom.2018.04.001. https://doi.org/ [DOI] [Google Scholar]
  26. Le D., Licata K., Provost E.M. Automatic Paraphasia Detection from Aphasic Speech: A Preliminary Study. Interspeech. 2017;2017:294–298. doi: 10.21437/Interspeech.2017-626. https://doi.org/ [DOI] [Google Scholar]
  27. Le, D., & Provost, E. M. (2016). Improving Automatic Recognition of Aphasic Speech with AphasiaBank. 2681–2685. https://doi.org/ 10.21437/Interspeech.2016-213
  28. LeCun Y., Bengio Y., Hinton G. Deep learning. Nature. 2015;521(7553):436–444. doi: 10.1038/nature14539. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  29. Lee A., Glass J. 2012 IEEE Spoken Language Technology Workshop (SLT) 2012. A comparison-based approach to mispronunciation detection; pp. 382–387. https://doi.org/ [DOI] [Google Scholar]
  30. Lee Ann, Glass J. 2012 IEEE Spoken Language Technology Workshop (SLT) 2012. A comparison-based approach to mispronunciation detection; pp. 382–387. https://doi.org/ [DOI] [Google Scholar]
  31. Lee T., Liu Y., Huang P.-W., Chien J.-T., Lam W.K., Yeung Y.T., Law T.K.T., Lee K.Y.S., Kong A.P.-H., Law S.-P. 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2016. Automatic speech recognition for acoustical analysis and assessment of cantonese pathological voice and speech; pp. 6475–6479. https://doi.org/ [DOI] [Google Scholar]
  32. Linebarger M.C., Schwartz M.F., Kohn S.E. Computer-based training of language production: An exploratory study. Neuropsychological Rehabilitation. 2001;11(1):57–96. doi: 10.1080/09602010042000178. https://doi.org/ [DOI] [Google Scholar]
  33. MacWhinney B., Fromm D., Forbes M., Holland A. AphasiaBank: Methods for studying discourse. Aphasiology. 2011;25(11):1286–1307. doi: 10.1080/02687038.2011.589893. https://doi.org/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Malik J., Girdhar D., Dahiya R., Sainarayanan G. Reference Threshold Calculation for Biometric Authentication. International Journal of Image, Graphics and Signal Processing. 2014;6(2):46–53. doi: 10.5815/ijigsp.2014.02.06. https://doi.org/ [DOI] [Google Scholar]
  35. Martín Abadi, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro, Greg S. Corrado, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Ian Goodfellow, Andrew Harp, Geoffrey Irving, Michael Isard, Jia, Y., Rafal Jozefowicz, Lukasz Kaiser, Manjunath Kudlur, … Xiaoqiang Zheng. (2015). TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. https://www.tensorflow.org/
  36. Laine Matti. Anomia: Theoretical and clinical aspects. Psychology. 2006 [Google Scholar]
  37. McHugh M.L. Interrater reliability: The kappa statistic. Biochemia Medica. 2012:276–282. doi: 10.11613/BM.2012.031. https://doi.org/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Meinedo H., Caseiro D., Neto J., Trancoso I. AUDIMUS.MEDIA: A Broadcast News Speech Recognition System for the European Portuguese Language. In: Mamede N.J., Trancoso I., Baptista J., das Graças Volpe Nunes M., editors. Computational Processing of the Portuguese Language. Springer; Berlin Heidelberg: 2003. pp. 9–17. (Vol. 2721https://doi.org/ [DOI] [Google Scholar]
  39. Pedregosa F., Varoquaux G., Gramfort A., Michel V., Thirion B., Grisel O., Blondel M., Prettenhofer P., Weiss R., Dubourg V. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research. 2011;12(Oct):2825–2830. [Google Scholar]
  40. Pompili A., Abad A., Trancoso I., Fonseca J., Martins I.P., Leal G., Farrajota L. Proceedings of the Second Workshop on Speech and Language Processing for Assistive Technologies. 2011. An on-line system for remote treatment of aphasia; pp. 1–10.http://dl.acm.org/citation.cfm?id=2140501 [Google Scholar]
  41. Qin Y., Lee T., Kong A.P.H. Automatic Assessment of Speech Impairment in Cantonese-speaking People with Aphasia. IEEE Journal of Selected Topics in Signal Processing. 2019 doi: 10.1109/JSTSP.2019.2956371. 1–1https://doi.org/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  42. Ram D., Miculicich L., Bourlard H. IEEE/ACM Transactions on Audio, Speech, and Language Processing. Vol. 28. 2020. Neural Network Based End-to-End Query by Example Spoken Term Detection; pp. 1416–1427. https://doi.org/ [DOI] [Google Scholar]
  43. Raschka S. MLxtend: Providing machine learning and data science utilities and extensions to Python's scientific computing stack. Journal of Open Source Software. 2018;3(24):638. doi: 10.21105/joss.00638. https://doi.org/ [DOI] [Google Scholar]
  44. Robinson, T. (1996). BEEP dictionary. BEEP Dictionary. http://svr-www.eng.cam.ac.uk/comp.speech/Section1/Lexical/beep.html
  45. Robinson T., Fransen J., Pye D., Foote J., Renals S. Proc. ICASSP. Vol. 95. 1995. Wsjcam0: A British English Speech Corpus For Large Vocabulary Continuous Speech Recognition; pp. 81–84. [Google Scholar]
  46. Stroke Association Stroke Association. 2018 https://www.stroke.org.uk/ [Google Scholar]
  47. Swinburn K., Porter G., Howard D. Gillian Porter and David Howard. Psychology Press; 2004. Comprehensive aphasia test: CAT /Kate Swinburn. [Google Scholar]
  48. Terpilowski M. scikit-posthocs: Pairwise multiple comparison tests in Python. Journal of Open Source Software. 2019;4(36):1169. doi: 10.21105/joss.01169. https://doi.org/ [DOI] [Google Scholar]
  49. van Heuven W.J.B., Mandera P., Keuleers E., Brysbaert M. Subtlex-UK: A New and Improved Word Frequency Database for British English. Quarterly Journal of Experimental Psychology. 2014;67(6):1176–1190. doi: 10.1080/17470218.2013.850521. https://doi.org/ [DOI] [PubMed] [Google Scholar]
  50. Whitworth A., Webster J., Howard D. Second edition. Psychology Press; 2014. A cognitive neuropsychological approach to assessment and intervention in aphasia: A clinician's guide. [Google Scholar]
  51. Yuan Y., Xie L., Leung C.-C., Chen H., Ma B. Fast Query-by-Example Speech Search Using Attention-Based Deep Binary Embeddings. IEEE/ACM Transactions on Audio, Speech, and Language Processing. 2020;28:1988–2000. doi: 10.1109/TASLP.2020.2998277. https://doi.org/ [DOI] [Google Scholar]

RESOURCES