Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2023 May 11;18(5):e0285629. doi: 10.1371/journal.pone.0285629

Causal speech enhancement using dynamical-weighted loss and attention encoder-decoder recurrent neural network

Fahad Khalil Peracha 1, Muhammad Irfan Khattak 1, Nema Salem 2,*, Nasir Saleem 3
Editor: Abdullah M Mutawa4
PMCID: PMC10174555  PMID: 37167227

Abstract

Speech enhancement (SE) reduces background noise signals in target speech and is applied at the front end in various real-world applications, including robust ASRs and real-time processing in mobile phone communications. SE systems are commonly integrated into mobile phones to increase quality and intelligibility. As a result, a low-latency system is required to operate in real-world applications. On the other hand, these systems need efficient optimization. This research focuses on the single-microphone SE operating in real-time systems with better optimization. We propose a causal data-driven model that uses attention encoder-decoder long short-term memory (LSTM) to estimate the time-frequency mask from a noisy speech in order to make a clean speech for real-time applications that need low-latency causal processing. The encoder-decoder LSTM and a causal attention mechanism are used in the proposed model. Furthermore, a dynamical-weighted (DW) loss function is proposed to improve model learning by varying the weight loss values. Experiments demonstrated that the proposed model consistently improves voice quality, intelligibility, and noise suppression. In the causal processing mode, the LSTM-based estimated suppression time-frequency mask outperforms the baseline model for unseen noise types. The proposed SE improved the STOI by 2.64% (baseline LSTM-IRM), 6.6% (LSTM-KF), 4.18% (DeepXi-KF), and 3.58% (DeepResGRU-KF). In addition, we examine word error rates (WERs) using Google’s Automatic Speech Recognition (ASR). The ASR results show that error rates decreased from 46.33% (noisy signals) to 13.11% (proposed) 15.73% (LSTM), and 14.97% (LSTM-KF).

1 Introduction

Speech signals in many real-world situations are degraded by noise signals. A degraded signal severely influences the performance of many speech-related applications, such as automatic speech recognition [1], speaker identification [2], and hearing aid devices [3]. A speech enhancement system is mainly involved in restoring the quality and improving the intelligibility of the signals degraded by the noise and is used at the front end of many speech applications to enhance their performance in noisy situations where they are altered. Background noises, competing speakers, and room reverberation are mainly the major sources of variations and distortions. An SE algorithm ideally ought to work passably in various acoustic situations, a broad-spectrum algorithm that is capable of performing well with little complexity and latency in every noisy situation is a challenging technical task.

The conventional approaches include spectral subtraction [4], Wiener filtering [5], statistical models [68], and hybrid SE models [9, 10] They show better performance in many stationary noises but face difficulties in handling nonstationary noises. In the recent past, deep learning has been developed into the mainstream for speech enhancement [11]. Given a speech dataset of the clean-noisy pairs, the neural networks can learn to transform the noisy magnitude spectra to their clean counterparts (mapping based) [1214] or estimate the time-frequency masks (masking-based) such as ideal binary mask (IBM) [15, 16], ideal ratio mask (IRM) [17, 18], and spectral magnitude mask (SMM) [19]. Fully connected networks (FCN) [19], feedforward neural networks (FDNN) with Kalman filtering [20], recurrent neural networks (RNN) [2123], and convolutional neural networks (CNN) [24, 25] are important deep learning approaches in SE.

A fully connected feedforward neural network showed that a DNN trained for a large number of background noises with a single speaker generalized better to untrained noise types [26]. Such a network, however, shows the difficulty in generalizing to both untrained speakers and noises when trained with a large number of speakers and noises. The RNN with LSTM is used to design a noise- and speaker-independent network for speech enhancement. A four-layered RNN network was used to train speech utterances belonging to 77 different speakers combined with 10,000 different noise types [26]. Recently, SE has aimed to improve the performance of the speaker and noise-independent networks. In [25], a CNN with gated and dilated convolutions is proposed for the magnitude enhancement. A recent trend is the use of attention mechanisms to improve the quality and intelligibility of noisy speech signals. In [27] a speech enhancement approach is proposed and used with an attention LSTM by replacing the forgetting gate with an attention gate. In [24] a dense CNN with a self-attention is proposed to assist feature extraction using a combination of feature reuse. In [28] a dual-path self-attention RNN is proposed to improve the long sequence of speech frames. A number of deep learning studies based on the attention mechanism for SE are successfully proposed with novel results [2934].

In most of the deep learning approaches, mean-square error (MSE) is used as the loss function [1519]. Other loss functions include Huber and mean absolute error (MAE). The gradient of the MAE remains invariant during training when loss approaches zero, resulting in missing the minima. Moreover, the Huber loss needs hyperparameter tuning. This brings further complexity when a loss function is dynamically weighted. A large error indicates poor learning on a particular instance in the dataset. A dynamically weighted loss function is used to alter the learning process by augmenting the weighted values corresponding to the learning errors. Through such an amendment, the loss function focuses on the large learning errors and improves the network performance.

In this paper, an attention encoder-decoder LSTM network for sequence-to-sequence learning is proposed. The motivation behind this research is the recent success of the attention mechanism in speech emotion recognition [33] and speech recognition [34]. Deep learning approaches can be regression or prediction tasks [35, 36]. It is useful to employ the attention process in speech enhancement since a human can focus on a certain part of a speech stream with more attention, such as target speech, whereas they perceive the surrounding noise with less attention. We have used an attention process on the encoder-decoder LSTM network that has been shown to perform better in modeling vital sequential information. LSTM [3739] can learn the weights of the past input features perfectly and predict the enhanced frames. The attention process determines the correlations between the previous frames and the current frames be enhanced and assigns weights to the previous frames. Experiments have shown that the proposed network consistently performed better in terms of speech quality and intelligibility. The overall structure of the proposed speech enhancement algorithm is depicted in Fig 1. We have summarized the main contributions of this study.

Fig 1.

Fig 1

  • For sequential learning to handle real-time speech applications that need low-latency causal processing, a causal speech enhancement based on attention encoder-decoder LSTM network is proposed.

  • By adding weighted values for large learning errors, a dynamically weighted loss function is used to improve the learning process. The loss function focuses on the large learning errors to further improve the network performance.

  • Automatic speech recognition is evaluated using estimated magnitude, thereby notably improving the word error rate in noisy situations.

The remainder of this paper is organized as follows. In Section 2, we explain the proposed speech enhancement algorithm. The dynamically-weighted loss is presented in Section 3. The experimentation is presented in Section 4. The results and discussions are presented in Section 5. Finally, the conclusions are drawn in Section 6.

2 Proposed speech enhancement

For a given clean speech signal xt and noise signal dt, the noisy speech signal yt is formed by the additive mixing as follows:

yt=xt+dt (1)

where {x,y,d}RN×1 and N shows number of the speech frames. A SE algorithm aims to recover a close estimate xt^ of the clean speech xt given yt. The inputs to the LSTM Encoder-Decoder are Y = [y1, .., yt, .., yN], where yt indicates the spectral magnitudes of the noisy speech at frame t. The high-level features h are extracted by the encoder from the input speech frames:

hK,hQ=Encoder(yt) (2)

where hK and hQ stand for the key and query, respectively. In this study, unidirectional LSTM is used as an encoder which shows a strong ability to model the sequential data leading to the improved performance of the speech enhancement [39]. The attention process is fed with key and query as the input to create fixed-length context vectors:

Ct=Attention(hK,hQ) (3)

The decoder output wt is the recovered enhanced speech signal xt^ which takes the context vectors Ct, the output of the encoder hQ, and the noisy speech yt, respectively.

w(t)=decoder(Ct,htQ,yt) (4)

The proposed attention encoder-decoder LSTM is depicted in Fig 2.

Fig 2.

Fig 2

2.1 Unidirectional LSTM encoder

The LSTM encoder extracts the high-level feature representations from the input speech frames. The input features are first fed into a fully-connected layer. The yt is the input to the LSTM cell as:

htK=f(yt) (5)

where f(⋅) is LSTM function whereas htK is LSTM output, respectively. The htQ can be computed as:

htQ=f(htK) (6)

2.2 Attention process

The attention process is fed with information about the key and query as inputs to create fixed-length context vectors. An attention process can use both previous and future speech frames. But, SE is a causal problem and uses previous speech frames to avoid processing latency. We have used casual dynamic and causal local attention approaches. To enhance a speech frame in causal dynamic attention, Y = [y1, .., yt] is used to compute the attention weights which means that all the previous speech frames are used to enhance the current frames. If the duration of the speech utterance is long, the attention weights of several previous speech frames can nearly be zero. Therefore, in casual local attention process, Y = [y1, .., yt] is used to compute the attention weights. The z is set to a constant. The normalized attention weight κ can be learned as:

κtk=exp(hkK,htQ)k=ltexp(hkK,htQ) (7)

l = 1 for causal dynamic attention whereas l = (tz) is used for the causal local attention. According to correlation computation, we have:

exp(hkK,htQ)=hkKTWhtQ (8)

The context vector with attention weights is given as:

Ct=n=lt(κtkhkK) (9)

With an attention-weighted context vector, the model decides the attention process.

2.2.1 Unidirectional LSTM decoder

The decoder recovers the output-enhanced speech by using the input features, encoder output, and context vector, respectively. The enhanced vector Et is learned from context vectors and features as:

Et=tanh(WE[Ct;htQ]+bE) (10)

where [Ct;htQ] shows the concatenation of the context and feature vectors, respectively. The ideal ratio mask (IRM) is finally estimated from the feature vectors. The time-frequency IRM(f, t) is given as:

IRM(f,t)=|X(f,t)|2|X(f,t)|2+|D(f,t)|2 (11)

where |X (f,t)| and |D (f,t)| show the magnitude spectra of clean speech and noise signals, respectively. The enhanced vectors are multiplied with the noisy features to recover the enhanced speech by taking the inverse Short-time Fourier Transform (STFT) as:

x^t=ytIRM(f,t) (12)

2.3 Weighted loss function

In masking-based deep learning methods for SE, a loss function presents a divergence between the predefined and the estimated mask. A loss function aims to reduce the errors produced during training. Mostly, the MSE (mean square error) is used as the basic loss function, given as:

loss(m,n)=1Lj=0L((m-n))2 (13)

where m, and n denote the input and the predicted value, respectively. Eq (13) can be represented in terms of the time-frequency mask as:

loss(Mx,M^x)=1Lj=0L((Mx-M^x))2 (14)

where M^(x) and M(x) denote the estimated and predefined IRM masks, respectively. The dynamical-weight loss function is used to adjust the network learning by multiplying weighted values corresponding to the learning errors. Thus, the loss function focuses on large errors to improve performance. The MSE loss function is multiplied by a weighted variable O to get the weighted MSE as:

loss(m,n)=O*(1Lj=0L((Mx-M^x))2) (15)

To emphasize the instances with large errors, the weight variable O in Eq (15) is updated according. The weight selection is done according to the following condition:

O*Loss(Mx,M^x)={|Mx-M^x|2,where|Mx-M^x|<B|(Mx-M^x)|,where|Mx-M^x|B (16)

Where |.| indicates the magnitude of ground truth and estimated masks. The weighting became halved when the absolute divergence is less than constant B which is set to 10 since it has been observed that the performance of the model at this instance was better.

3 Experiments

3.1 Data generation

In experiments, we have used IEEE dataset [40]. Two IEEE datasets are used which are composed of male speakers and female speakers, respectively. The noise sources are selected from AURORA [41]. For a given speech dataset D, we have Mtr and Mte as the training and testing speech utterances. The training and testing speech utterances in the dataset are denoted by Dtr and Dte, respectively. The noisy utterances are generated by adding the noise signals to Dtr and Dte:

ytri=xtri+dtri,i=1,2,3,...,Mtr (17)
ytej=xtej+dtej,j=1,2,3,...,Mtr (18)

3.2 Feature extraction

The input pairs y, x, d are transformed from the time to frequency domain using the STFT as:

Y=STFT(y);X=STFT(x);D=STFT(d) (19)

Where {X,Y,D}ZT×F, T and F show frame number and frequency bin number. We have used STFT magnitude |Y| as the input features.

3.3 Experimental setup

We have used speech utterances with a 16 kHz sampling rate. A 512 points Hanning window with 75% overlapping is used. We used the noisy phase during waveform reconstruction. The network consists of an input layer, three unidirectional LSTMs with 256 memory units followed by a fully connected output layer with 257 sigmoidal units. The number of epochs and the learning rate is set to 160 and 0.001, respectively. The weights are randomly initialized and trained with 32 sequences mini-batches by back-propagation through time with Adam optimizer. The three-layered LSTM network architecture with (128/256/256/256/257) memory cells is used. The details of hyperparameters are given in Table 1. To create the noisy utterances, three SNR levels are used (-5dB to +5dB) with a 5dB step size. For network training, IEEE speech utterances from the male and female speakers are duplicated three times for each SNR level and mixed with all noise types. Therefore, a total of 21600 (approx. 18 hours) speech utterances are used in the training process. We also used half-speech utterances during testing in the matched and mismatched conditions. During testing, each noise type is tested with a different set of utterances.

Table 1. Hyperparameters of all deep learning networks.

Hyperparameters Atten-LSTM LSTM LSTM-KF DNN
No of layers 3 3 3 3
Layer 1 Neurons 256 256 256 1024
Layer 2 Neurons 256 256 256 1024
Layer 3 Neurons 256 256 256 1024
Learning Rate 0.0001 0.0001 0.0001 0.0001
No of Epochs 160 160 160 160
Momentum 0.9 0.9 0.9 0.9

To evaluate the performance of the proposed approach, we used three objective measures: Short-time objective intelligibility (STOI) [42], Perceptual evaluation of speech quality (PESQ) [43], and Source-to-distortion ratio (SDR). To compare performance of the proposed method, baseline LSTM [26], DNN [18], LMMSE [6], OM-LSA [7], FDNN-KF [20], LSTM-KF [44] DeepXi-KF [45] and DeepResGRU-KF [46]. are selected as the competing methods. The competing deep learning methods can be represented as LSTM-IRM: LSTM with IRM as a training target. DNN-IRM: Feedforward DNN with IRM as a training target. DWAtten-LSTM-IRM: Proposed method with IRM as a training target.

4 Results and discussions

To signify the performance of the proposed speech enhancement method, we have compared the results with baseline LSTM-IRM [26], DNN-IRM [18], LMMSE [6], OM-LSA [7], FDNN-KF [20], LSTM-KF [44] DeepXi-KF [45] and DeepResGRU-KF [46]. For matched and unmatched conditions, the average STOI, PESQ, and SDR test values across several noise sources and SNRs are given in Tables 25. Note that, in contrast to the competing deep-learning methods, the proposed DWAtten-LSTM presents the highest performance in terms of the STOI, PESQ, and SDR values in noisy situations. Two conditions including Matched and Mismatched are considered in experiments. During match conditions, the speakers and utterances remain the same in training and testing sets whereas, in mismatched conditions, the speakers and utterances in the training set are different from the testing set.

Table 2. STOI (in%) in matched test scores.

Noise Algorithm -5dB 0dB 5dB Avg
Airport Noise Noisy (UnP) 62.4 73.7 83.9 73.3
DNN-IRM 80.2 86.5 90.7 85.8
LSTM-IRM 82.9 88.6 92.3 87.9
LSTM-KF 74.2 80.3 82.8 79.1
FDNN-KF 72.2 79.8 81.3 77.8
DeepXi-KF 75.9 81.2 84.7 80.6
DeepResGRU-KF 76.6 82.3 85.5 81.1
Proposed 85.6 90.5 94.0 90.0
Babble Noise Noisy (UnP) 56.7 68.1 79.6 68.1
DNN-IRM 72.0 79.3 85.9 79.0
LSTM-IRM 75.9 82.6 87.0 81.8
LSTM-KF 74.1 80.1 82.7 79.2
FDNN-KF 72.3 79.6 81.1 78.0
DeepXi-KF 75.8 81.3 84.8 80.7
DeepResGRU-KF 76.5 82.4 85.6 81.2
Proposed 79.8 85.2 89.2 84.7
Car Noise Noisy (UnP) 58.8 68.9 79.6 69.1
DNN-IRM 73.7 81.0 86.0 80.2
LSTM-IRM 78.5 84.7 89.1 84.1
LSTM-KF 78.3 83.2 87.6 83.0
FDNN-KF 75.4 80.4 83.3 79.7
DeepXi-KF 80.1 85.6 89.0 84.9
DeepResGRU-KF 82.2 86.8 91.2 86.7
Proposed 83.4 88.5 92.6 88.2
Factory Noise Noisy (UnP) 56.4 67.4 78.8 67.5
DNN-IRM 70.2 78.7 85.2 78.0
LSTM-IRM 74.4 80.2 88.1 80.9
LSTM-KF 74.1 78.5 83.6 78.7
FDNN-KF 70.7 76.2 81.4 76.1
DeepXi-KF 77.2 79.8 85.4 80.8
DeepResGRU-KF 79.1 81.2 86.5 82.2
Proposed 78.6 82.1 90.0 83.6
Street Noise Noisy (UnP) 63.0 74.0 84.0 73.7
DNN-IRM 78.4 85.3 90.3 84.6
LSTM-IRM 82.0 86.8 91.8 87.9
LSTM-KF 74.8 82.9 86.6 83.0
FDNN-KF 70.8 79.8 83.3 79.7
DeepXi-KF 78.3 84.9 89.0 84.9
DeepResGRU-KF 80.1 85.6 91.2 86.7
Proposed 84.9 89.7 93.4 89.3

Table 5. Unmatched test scores in five noise sources at all SNRs.

The results are averaged over all testing utterances.

Noise Algorithm STOI (in%) PESQ SDR
Airport Noise Noisy (UnP) 71.6 1.83 0.14
DNN-IRM 82.6 2.27 6.30
LSTM-IRM 85.0 2.37 6.84
LSTM-KF 85.9 2.40 6.96
FDNN-KF 83.4 2.31 6.41
DeepXi-KF 86.7 2.45 7.21
DeepResGRU-KF 86.9 2.46 7.31
Proposed 87.6 2.48 7.39
Babble Noise Noisy (UnP) 65.9 1.72 0.19
DNN-IRM 76.7 2.12 6.13
LSTM-IRM 78.5 2.23 6.31
LSTM-KF 79.1 2.28 6.35
FDNN-KF 77.2 2.15 6.18
DeepXi-KF 80.2 2.30 6.45
DeepResGRU-KF 80.8 2.32 6.51
Proposed 81.0 2.35 6.62
Car Noise Noisy (UnP) 67.5 1.65 0.13
DNN-IRM 79.6 2.17 6.38
LSTM-IRM 82.8 2.33 6.92
LSTM-KF 83.1 2.40 7.01
FDNN-KF 80.1 2.21 6.45
DeepXi-KF 84.2 2.45 7.12
DeepResGRU-KF 84.8 2.48 7.28
Proposed 86.0 2.49 7.45
Factory Noise Noisy (UnP) 66.1 1.61 0.18
DNN-IRM 75.7 2.08 5.92
LSTM-IRM 78.6 2.24 6.39
LSTM-KF 78.8 2.29 6.45
FDNN-KF 76.0 2.10 8.98
DeepXi-KF 79.5 2.35 6.57
DeepResGRU-KF 80.2 2.38 6.71
Proposed 81.3 2.39 6.87
Street Noise Noisy (UnP) 66.8 1.73 0.22
DNN-IRM 77.8 2.18 6.06
LSTM-IRM 81.5 2.31 6.34
LSTM-KF 81.9 2.35 6.40
FDNN-KF 78.0 2.21 6.10
DeepXi-KF 82.6 2.38 6.45
DeepResGRU-KF 84.7 2.40 6.58
Proposed 85.1 2.42 6.63

Table 3. PESQ in matched test scores.

Noise Algorithm -5dB 0dB 5dB Avg
Noisy (UnP) 1.53 1.86 2.14 1.84
Airport Noise DNN-IRM 1.93 2.41 2.79 2.37
LSTM-IRM 2.11 2.54 2.87 2.51
LSTM-KF 1.88 2.13 2.56 2.14
FDNN-KF 1.79 2.03 2.45 2.07
DeepXi-KF 1.99 2.21 2.58 2.23
DeepResGRU-KF 2.06 2.33 2.71 2.34
Proposed 2.29 2.67 2.98 2.65
Babble Noise Noisy (UnP) 1.52 1.75 2.07 1.78
DNN-IRM 1.95 2.34 2.66 2.32
LSTM-IRM 2.07 2.43 2.70 2.40
LSTM-KF 1.93 2.18 2.56 2.19
FDNN-KF 1.84 2.13 2.45 2.12
DeepXi-KF 2.04 2.26 2.60 2.28
DeepResGRU-KF 2.11 2.38 2.71 2.40
Proposed 2.19 2.56 2.78 2.51
Car Noise Noisy (UnP) 1.37 1.62 1.92 1.64
DNN-IRM 1.76 2.17 2.58 2.17
LSTM-IRM 2.01 2.41 2.78 2.40
LSTM-KF 2.03 2.33 2.67 2.41
FDNN-KF 1.94 2.25 2.60 2.33
DeepXi-KF 2.08 2.45 2.73 2.48
DeepResGRU-KF 2.21 2.55 2.86 2.52
Proposed 2.27 2.65 2.99 2.63
Factory Noise Noisy (UnP) 1.31 1.61 1.92 1.61
DNN-IRM 1.72 2.16 2.60 2.16
LSTM-IRM 1.94 2.32 2.71 2.32
LSTM-KF 1.81 2.05 2.36 2.07
FDNN-KF 1.72 2.01 2.28 2.01
DeepXi-KF 1.92 2.12 2.40 2.15
DeepResGRU-KF 1.98 2.26 2.55 2.26
Proposed 2.16 2.57 2.87 2.53
Street Noise Noisy (UnP) 1.47 1.86 2.01 1.78
DNN-IRM 1.94 2.44 2.65 2.34
LSTM-IRM 2.11 2.53 2.80 2.48
LSTM-KF 1.84 2.12 2.39 2.12
FDNN-KF 1.75 2.04 2.31 2.03
DeepXi-KF 1.93 2.21 2.45 2.20
DeepResGRU-KF 2.01 2.31 2.65 2.32
Proposed 2.29 2.66 2.99 2.65

In the matched conditions (Tables 24), the proposed DWAtten-LSTM approach achieved the best STOI, PESQ, and SDR values with airport noise, street noise, and car noise at SNR≥5dB, that is, STOI≥94%, PESQ≥2.99, and SDR≥10.8dB. The STOI with the babble noise is improved from 68.1% with noisy speech signals to 85.2% with DWAtten-LSTM and achieves 17.1% improvement in STOI at 0dB SNR. Similarly, the PESQ with airport noise is improved from 1.93 with DNN-IRM to 2.29 with proposed DWAtten-LSTM and improved the PESQ by factor 0.36 (18.75%) at -5dB SNR. Moreover, the SDR value with car noise is improved from 4.20dB with LSTM-IRM to 4.56dB with DWAtten-LSTM and achieved 0.36dB (8.57%) improvement at -5dB SNR level. Note from Table 2 (matched condition) that DWAtten-LSTM presents the highest STOI, PESQ, and SDR values in all noisy situations as compared to baseline LSTM with the same network architecture.

Table 4. SDR in matched test scores.

Noise Algorithm -5dB 0dB 5dB Avg
Airport Noise Noisy (UnP) -4.78 0.11 5.07 0.13
DNN-IRM 3.98 6.86 8.38 6.41
LSTM-IRM 4.09 7.10 9.54 6.91
LSTM-KF 4.05 7.13 9.55 6.92
FDNN-KF 3.99 7.01 9.42 6.80
DeepXi-KF 4.09 7.15 9.84 7.02
DeepResGRU-KF 4.16 7.22 10.1 7.16
Proposed 4.21 7.33 10.7 7.41
Babble Noise Noisy (UnP) -4.73 0.13 5.08 0.16
DNN-IRM 3.82 6.31 8.74 6.29
LSTM-IRM 3.95 6.58 9.05 6.52
LSTM-KF 4.03 7.21 9.25 6.83
FDNN-KF 3.95 7.10 9.12 6.72
DeepXi-KF 4.11 7.26 9.24 6.87
DeepResGRU-KF 4.19 7.31 9.29 6.93
Proposed 4.28 7.40 9.36 7.01
Car Noise Noisy (UnP) -4.85 0.08 5.05 0.09
DNN-IRM 3.81 6.42 8.83 6.35
LSTM-IRM 4.20 7.21 9.92 7.11
LSTM-KF 4.28 7.29 10.1 7.20
FDNN-KF 3.92 6.52 9.01 6.46
DeepXi-KF 4.34 7.35 10.2 7.45
DeepResGRU-KF 4.48 7.44 10.4 7.58
Proposed 4.56 7.59 10.8 7.65
Factory Noise Noisy (UnP) -4.69 0.12 5.07 0.17
DNN-IRM 3.66 5.34 8.72 6.24
LSTM-IRM 3.85 5.53 9.52 6.30
LSTM-KF 3.83 5.78 9.63 6.41
FDNN-KF 3.75 5.55 8.87 6.05
DeepXi-KF 3.91 6.12 9.96 6.66
DeepResGRU-KF 3.89 6.25 10.2 6.78
Proposed 4.01 6.69 10.3 6.99
Street Noise Noisy (UnP) -4.76 0.11 4.99 0.11
DNN-IRM 4.01 6.83 8.34 6.39
LSTM-IRM 4.05 7.32 9.22 6.86
LSTM-KF 4.08 7.43 9.33 6.94
FDNN-KF 4.02 6.98 8.40 6.46
DeepXi-KF 4.11 7.86 9.66 7.21
DeepResGRU-KF 4.18 7.73 9.89 7.26
Proposed 4.19 7.82 10.1 7.37

The average STOI, PESQ, and SDR test values across all noise sources and SNRs are given in Table 5 for unmatched conditions where the proposed DWAtten-LSTM achieved the best STOI, PESQ, and SDR values at airport noise at SNR≥5dB, that is, STOI≥92.3%, PESQ≥2.86, and SDR≥10.7dB. The STOI with factory noise is improved from 55.2% with noisy speech signals to 77.0% with DWAtten-LSTM and achieves 21.8% STOI gain at -5dB SNR. Fig 3 shows STOI, PESQ, and SDR improvements in various background noises where we can see the performance of the proposed DWAtten-LSTM in individual noise at -5dB, 0dB, and 5dB SNRs.

Fig 3.

Fig 3

We also compared DWAtten-LSTM to non-deep learning-based LMMSE and OM-LSA. Table 6 shows STOI, PESQ, and SDR values obtained with DWAtten-LSTM, LMMSE, and OM-LSA, respectively. STOI is improved from 70.22% and 71.50% with LLMSE and OM-LSA to 85.60% with DWAtten-LSTM and achieves 15.38% and 14.10% STOI gain at 0dB. Similarly, PESQ is improved from 2.22 and 2.20 with LMMSE, and OM-LSA to 2.85 with proposed DWAtten-LSTM and achieved 0.63 and 0.65 PESQ gain at 5dB. The PESQi and STOIi are demonstrated in Fig 4.

Table 6. Comparison against non-deep learning methods.

Test Scores are averaged over five noise sources at all SNRs.

STOI (in %) PESQ SDR
Algorithm -5dB 0dB 5dB Avg -5dB 0dB 5dB Avg -5dB 0dB 5dB Avg
Noisy (UnP) 58.02 68.87 80.05 68.98 1.43 1.73 2.01 1.72 -4.73 0.13 5.07 0.13
DNN-IRM 73.94 80.06 85.39 79.79 1.83 2.23 2.61 2.22 3.84 6.40 8.50 6.25
LSTM-IRM 77.39 83.15 87.89 82.81 2.02 2.36 2.72 2.37 4.00 6.61 9.35 6.65
LLMSE 59.21 70.22 81.80 70.41 1.49 1.75 2.22 1.82 -3.43 0.18 5.32 0.69
OM-LSA 59.49 71.50 82.14 71.04 1.52 1.81 2.20 1.84 -3.88 0.21 5.51 0.61
Proposed-M 82.46 87.20 91.84 87.16 2.24 2.62 2.92 2.59 4.19 7.17 10.3 7.20
Proposed-UM 79.06 84.08 89.48 84.20 2.09 2.40 2.79 2.43 4.09 6.72 10.0 6.97
Proposed-Avg 80.76 85.64 90.66 85.68 2.17 2.51 2.85 2.51 4.14 6.95 10.2 7.09

Fig 4.

Fig 4

The overall average STOI, PESQ, and SDR values for all noise sources are given in Table 6 for matched (denoted as Proposed-M), unmatched (denoted by Proposed-UM) conditions, and the average of both matched and unmatched (denoted as Proposed-Avg). PESQ and STOI values are calculated for the causal local attention process, and the value of z is varied from 4 to 12 with an increment of 4. Table 7 shows the results. It is noticed that values greater than 12 for z result in no further improvements and the best performance is achieved for z = 4. As compared to causal dynamic attention, causal local attention showed better results. The observations verified the inference that extensive previous information is not required in speech enhancement. This inference is logical since noisy situations, both types and SNRs, change over time. The observations are valid for the attention networks since the attention LSTM outperformed the baseline LSTM.

Table 7. Unmatched test scores in five noise sources at all SNRs.

The results are averaged over all testing utterances.

z STOI (in%) PESQ
-5dB 0dB 5dB -5dB 0dB 5dB
4 80.92 86.46 90.88 2.163 2.503 2.81
12 80.90 86.42 90.82 2.160 2.501 2.80
24 80.70 86.12 90.01 2.145 2.489 2.77

Table 8 shows the comparison between the loss errors and predicted results of the DWAtten-LSTM with and without the DW loss function. The DW loss function improved the predicted scores (STOI and PESQ). The errors are reduced by weighted MSE (3.42 × 10−4) as compared to a non-weighted MSE (3.54 × 10−4).

Table 8. Dynamical-Weight vs. Non-Dynamical-Weight loss.

Algorithm Objective Measure Errors
Proposed+MSE STOI: 83%, PESQ: 2.42 3.54 × 10−4
Proposed+DW-MSE STOI: 85%, PESQ: 2.50 3.42 × 10−4
Improvements STOIi: 2.1%, PESQi: 3.3% -3.51%

To understand the attention process, the attention maps are illustrated in Fig 5. The x-axis denotes hK and the y-axis denotes hQ. The points, (x;y) denote the attention weights. The attention-based network assigns different attention levels (weights) to the contextual frames. The top spectrogram shows noisy speech, and the other spectrogram shows clean speech, respectively.

Fig 5.

Fig 5

In experiments, time-varying spectral analysis is conducted to showcase the performance of DWAtten-LSTM. Fig 6 demonstrates the sample spectrogram analysis. A clean speech utterance is mixed with babble noise at 5 dB. The spectrogram of DWAtten-LSTM is plotted in Fig 6(F). The harmonic structures of the vowel and the formant peaks are well retained. Moreover, the spectrogram showed excellent structure during speech activity. During the speech pause, DWAtten-LSTM removed the residual noise signals. The weak harmonic structures in the high-frequency sub-bands are well maintained. Thus, a better speech quality of the enhanced speech is achieved by DWAtten-LSTM. The weak energy in the speech utterance is also well retained and yields less speech distortion. Therefore, the intelligibility of noisy speech is improved. The residual noise signals are evident in the spectrograms of LMMSE and OM-LSA, plotted in Fig 6(C) and 6(D).

Fig 6.

Fig 6

The complexity and convergence analysis are also given. The complexity of a deep learning algorithm revolves around the number of training parameters; LSTM networks have 1.2 million parameters. This is clearly a fewer number as compared to other networks used for speech enhancement, for example, 10 million parameters are used by the residual LSTM [47]. This also significantly reduces the training time and speeds up the process. DWAtten-LSTM took less time per epoch compared to the Residual LSTM (using an NVIDIA GTX 950 Ti GPU). Next, we observed the convergence of Weighted-MSE between the estimated and true values for the training and testing data sets of DWAtten-LSTM. The MSE has been reduced after every epoch until converging at around epoch 155.

According to STOI, PESQ, and SDR, the following inferences are drawn. Under various noisy situations, PESQ, STOI, and SDR values indicate that DWAtten-LSTM achieved the best improvements in quality (PESQi), intelligibility (STOIi), and speech distortion (SDRi) as compared to the competing deep learning and non-deep learning methods. The proposed DWAtten-LSTM method improved the quality without degrading speech intelligibility in noisy situations. All deep-learning methods showed repeated improvements in STOI and SDR values, which suggests the potential of deep learning for speech enhancement tasks.

The ASR systems use a magnitude spectrum of speech signals, and one would expect that deep learning approaches would certainly improve ASR performance in noisy situations. For ASR systems, SE algorithms operate at the front end. We have used Google ASR [48] to examine the ASR performance in terms of the WERs. The average WERs are given in Table 9, which shows that LSTM and DWAtten-LSTM boosted the ASR performance. The error rates decreased from 46.33% (noisy signals) to 13.11% (DWAtten-LSTM) and 15.73% (LSTM), respectively. The ASR gradually decreases as the SNR increases, partly because the noise becomes smaller. The ASR experiments aim to show the potential of the proposed RNNs and DNNs instead of achieving state-of-the-art (SOTA) results.

Table 9. WERs for different SE algorithms.

Noisy Speech DWAtten-LSTM LSTM LSTM-KF LMMSE
46.33% 13.11% 15.73% 14.97% 27.68%

4.1 Subjective evaluation

In addition, we have conducted subjective listening tests to assess the perceptual quality of enhanced speech. The enhanced speech utterances are randomly chosen from various noise sources (airport, babble, factory, and restaurant) using three SNRs, which are -5 dB, 0 dB, and 5 dB. In total, 300 speech utterances are used to assess DNN, LSTM, and the proposed SE. The participants are requested to assign a score (from 0 to 5) according to perceived speech quality. During experiments, no speech utterance is repeated. The listening tests are conducted in an isolated room using high-quality headphones. The data of the listeners who participated in the subjective listening tests for speech quality are given in Table 10. Prior training sessions are arranged to educate the listeners about the procedures.

Table 10. The subjective listener’s biodata.

Listeners L1 L2 L3 L4 L5 L6 L7 L8 L9
Age 27 29 33 36 40 40 45 48 48
Gender M M F M F F M M M

Fig 7 shows the subjective listening tests in terms of MOS. The proposed DWAtten-LSTM showed better MOS performance. The average MOS result at negative SNRs is higher than 2.80 (MOS≥2.86 at -5 dB), which indicates significant performance. At SNR≥0dB, the DWAtten-LSTM exceeded the average MOS score beyond 3.0 (MOS≥3.0 at SNR≥0dB). The MOS results for all listeners in the tests are averaged. The ANOVA tests for MOS at -5 dB, 0 dB, and 5 dB are [F (2, 9) = 43.4, p<0.0001], [F (2, 9) = 34.7, p<0.0001] and [F (2, 9) = 28.3, p<0.0001] which indicate the statistical significance achieved by DWAtten-LSTM in terms of the MOS scores. The other models (DNN and LSTM) also performed better since deep learning is able to produce better speech quality.

Fig 7.

Fig 7

4.2 Speech dereverberation

This section examined the dereverberation performance of the proposed SE. To train the SE model, three reverberation times (0.4 sec, 0.6 sec, and 0.8 sec) are considered. A total of 100 anechoic speech utterances from the IEEE dataset [40] are used to create the training dataset. The testing dataset contains 40 reverberant speech utterances. There is no overlap between the speech utterances used during model training and testing. The proposed method with reverberant speech utterances is compared and examined for dereverberation. The results are compared with the study of Wu and Wang [49], where estimated inverse filters and spectral subtraction are used to reduce reverberation. Table 11 shows the results using STOI and PESQ. The proposed method delivered the best STOI and PESQ scores, i.e., STOI≥78.3%, and PESQ≥2.45 at RT≥4 sec. The spectrograms are provided in Fig 8, where the smearing energy produced by reverberation is considerably reduced, showing that the reverberation performance of the proposed method is improved.

Table 11. Performance analysis of proposed SE in reverberant situations.

Metric Method Reverberation Time
0.4 sec 0.6 sec 0.8 sec
STOI Noisy Reverb 53.1 48.5 40.2
Wu and Wang [49] 65.3 59.8 55.1
Proposed 78.3 70.1 68.4
PESQ Noisy Reverb 2.11 1.98 1.81
Wu and Wang [49] 2.32 2.00 1.92
Proposed 2.45 2.19 2.04

Fig 8.

Fig 8

5 Conclusions

In this paper, we have proposed a monaural SE based on the attention LSTM encoder-decoder model with a novel loss function. The proposed DWAtten-LSTM estimated the magnitude spectrum from the noisy speech signals using an ideal ratio mask. We have compared this model to the baseline and competing for deep learning and non-deep-learning methods for speech intelligibility and quality assessment. The objective assessments are accomplished in various noisy situations using three input SNR levels. The PESQ and SDR values indicated that the proposed DWAtten-LSTM achieved significant gains of 0.79 (45.93%) and 6.96dB over noisy speech. Similarly, STOI indicated that DWAtten-LSTM kept intelligibility in all noisy situations and STOI achieved a large gain of 16.70% over the noisy speech. The subjective analysis confirmed the success of the proposed model in terms of speech quality. The results and analysis concluded that we achieved better results in terms of speech quality and intelligibility with the proposed DWAtten-LSTM. The attention process observations verified the inference that extensive previous information is not vital in speech enhancement. The proposed loss function significantly improved the model learning. Although deep learning for speech enhancement outperformed the conventional methods with their complex network architectures, yet required less computationally complex and efficient network architectures for improved performance. The proposed DWAtten-LSTM SE algorithm has demonstrated considerable performance gain as compared to the baseline LSTM and FDNN and achieved higher performance gains when compared to the conventional SE.

Our future research will focus on further improving the quality and intelligibility by proposing computationally less complex network architectures in intense unseen noises and speakers. Moreover, phase estimation will also be included to increase the speech quality. This study used STFT as a transformation tool for frequency domain representation; however, several transformations are available in the literature. In future studies, these transformations [5053] will be used for more in-depth analysis.

Supporting information

S1 File

(ZIP)

S2 File

(ZIP)

S3 File

(ZIP)

Data Availability

All relevant data are within the manuscript and its Supporting information files.

Funding Statement

Enter: The author(s) received no specific funding for this work.

References

  • 1. Lee W, Seong JJ, Ozlu B, Shim BS, Marakhimov A, Lee S. Biosignal sensors and deep learning-based speech recognition: A review. Sensors. 2021. Feb 17;21(4):1399. doi: 10.3390/s21041399 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Benzeghiba M, De Mori R, Deroo O, Dupont S, Erbes T, Jouvet D, et al. Automatic speech recognition and speech variability: A review. Speech communication. 2007. Oct 1;49(10–11):763–86. [Google Scholar]
  • 3. Kavalekalam MS, Nielsen JK, Boldt JB, Christensen MG. Model-based speech enhancement for intelligibility improvement in binaural hearing aids. IEEE/ACM Transactions on Audio, Speech, and Language Processing. 2018. Sep 26;27(1):99–113. doi: 10.1109/TASLP.2018.2872128 [DOI] [Google Scholar]
  • 4. Boll S. Suppression of acoustic noise in speech using spectral subtraction. IEEE Transactions on acoustics, speech, and signal processing. 1979. Apr;27(2):113–20. doi: 10.1109/TASSP.1979.1163209 [DOI] [Google Scholar]
  • 5. Saleem N, Irfan M. Noise reduction based on soft masks by incorporating SNR uncertainty in frequency domain. Circuits, Systems, and Signal Processing. 2018. Jun;37:2591–612. doi: 10.1007/s00034-017-0684-5 [DOI] [Google Scholar]
  • 6. Ephraim Y, Malah D. Speech enhancement using a minimum mean-square error log-spectral amplitude estimator. IEEE transactions on acoustics, speech, and signal processing. 1985. Apr;33(2):443–5. doi: 10.1109/TASSP.1985.1164550 [DOI] [Google Scholar]
  • 7. Cohen I. Optimal speech enhancement under signal presence uncertainty using log-spectral amplitude estimator. IEEE Signal processing letters. 2002. Apr;9(4):113–6. doi: 10.1109/97.1001645 [DOI] [Google Scholar]
  • 8. Mahmmod BM, Baker T, Al-Obeidat F, Abdulhussain SH, Jassim WA. Speech enhancement algorithm based on super-Gaussian modeling and orthogonal polynomials. IEEE Access. 2019. Jul 18;7:103485–504. doi: 10.1109/ACCESS.2019.2929864 [DOI] [Google Scholar]
  • 9.Mahmmod BM, Abdulhussain SH, Naser MA, Alsabah M, Mustafina J. Speech enhancement algorithm based on a hybrid estimator. InIOP Conference Series: Materials Science and Engineering 2021 Mar 1 (Vol. 1090, No. 1, p. 012102). IOP Publishing.
  • 10. Gnanamanickam J, Natarajan Y, KR SP. A hybrid speech enhancement algorithm for voice assistance application. Sensors. 2021. Oct 23;21(21):7025. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Wang D, Chen J. Supervised speech separation based on deep learning: An overview. IEEE/ACM Transactions on Audio, Speech, and Language Processing. 2018. May 30;26(10):1702–26. doi: 10.1109/TASLP.2018.2842159 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Wang Y, Wang D. Towards scaling up classification-based speech separation. IEEE Transactions on Audio, Speech, and Language Processing. 2013. Mar 7;21(7):1381–90. doi: 10.1109/TASL.2013.2250961 [DOI] [Google Scholar]
  • 13. Xu Y, Du J, Dai LR, Lee CH. A regression approach to speech enhancement based on deep neural networks. IEEE/ACM Transactions on Audio, Speech, and Language Processing. 2014. Oct 21;23(1):7–19. doi: 10.1109/TASLP.2014.2364452 [DOI] [Google Scholar]
  • 14. Xu Y, Du J, Dai LR, Lee CH. An experimental study on speech enhancement based on deep neural networks. IEEE Signal processing letters. 2013. Nov 14;21(1):65–8. doi: 10.1109/LSP.2013.2291240 [DOI] [Google Scholar]
  • 15. Saleem N, Khattak MI. Deep neural networks based binary classification for single channel speaker independent multi-talker speech separation. Applied Acoustics. 2020. Oct 1;167:107385. doi: 10.1016/j.apacoust.2020.107385 [DOI] [Google Scholar]
  • 16. Saleem N, Khattak MI. Multi-scale decomposition based supervised single channel deep speech enhancement. Applied Soft Computing. 2020. Oct 1;95:106666. doi: 10.1016/j.asoc.2020.106666 [DOI] [Google Scholar]
  • 17. Saleem N, Khattak MI. Deep Neural Networks for Speech Enhancement in Complex-Noisy Environments. International Journal of Interactive Multimedia and Artificial Intelligence. 2020. Mar 1;6(1):84–91. doi: 10.9781/ijimai.2019.06.001 [DOI] [Google Scholar]
  • 18. Wang Y, Narayanan A, Wang D. On training targets for supervised speech separation. IEEE/ACM transactions on audio, speech, and language processing. 2014. Aug 28;22(12):1849–58. doi: 10.1109/TASLP.2014.2352935 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Saleem N, Khattak MI, Qazi AB. Supervised speech enhancement based on deep neural network. Journal of Intelligent. Fuzzy Systems. 2019. Jan 1;37(4):5187–201. doi: 10.3233/JIFS-190047 [DOI] [Google Scholar]
  • 20. Xia Y, Wang J. Low-dimensional recurrent neural network-based Kalman filter for speech enhancement. Neural Networks. 2015. Jul 1;67:131–9. doi: 10.1016/j.neunet.2015.03.008 [DOI] [PubMed] [Google Scholar]
  • 21. Huang Y, Yu J, Leng J, Liu B, Yi Z. Continuous Recurrent Neural Networks Based on Function Satlins: Coexistence of Multiple Continuous Attractors. Neural Processing Letters. 2022. Apr 1:1–23. [Google Scholar]
  • 22.Zhao H, Zarar S, Tashev I, Lee CH. Convolutional-recurrent neural networks for speech enhancement. In2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2018 Apr 15 (pp. 2401–2405). IEEE.
  • 23. Saleem N, Khattak MI, Al-Hasan MA, Jan A. Multi-objective long-short term memory recurrent neural networks for speech enhancement. Journal of Ambient Intelligence and Humanized Computing. 2021. Oct;12(10):9037–52. doi: 10.1007/s12652-020-02598-4 [DOI] [Google Scholar]
  • 24. Pandey A, Wang D. A new framework for CNN-based speech enhancement in the time domain. IEEE/ACM Transactions on Audio, Speech, and Language Processing. 2019. Apr 29;27(7):1179–88. doi: 10.1109/taslp.2019.2913512 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Tan K, Chen J, Wang D. Gated residual networks with dilated convolutions for monaural speech enhancement. IEEE/ACM transactions on audio, speech, and language processing. 2018. Oct 14;27(1):189–98. doi: 10.1109/TASLP.2018.2876171 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Chen J, Wang D. Long short-term memory for speaker generalization in supervised speech separation. The Journal of the Acoustical Society of America. 2017. Jun 23;141(6):4705–14. doi: 10.1121/1.4986931 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Liang R, Kong F, Xie Y, Tang G, Cheng J. Real-time speech enhancement algorithm based on attention LSTM. IEEE Access. 2020. Mar 9;8:48464–76. doi: 10.1109/ACCESS.2020.2979554 [DOI] [Google Scholar]
  • 28.Pandey A, Wang D. Dual-path self-attention RNN for real-time speech enhancement. arXiv preprint arXiv:2010.12713. 2020 Oct 23.
  • 29.Hao X, Shan C, Xu Y, Sun S, Xie L. An attention-based neural network approach for single channel speech enhancement. InICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019 May 12 (pp. 6895–6899). IEEE.
  • 30. Zhang L, Wang M, Zhang Q, Liu M. Environmental attention-guided branchy neural network for speech enhancement. Applied Sciences. 2020. Feb 9;10(3):1167. doi: 10.3390/app10031167 [DOI] [Google Scholar]
  • 31. Nicolson A, Paliwal KK. Masked multi-head self-attention for causal speech enhancement. Speech Communication. 2020. Dec 1;125:80–96. doi: 10.1016/j.specom.2020.10.004 [DOI] [Google Scholar]
  • 32. Xiang X, Zhang X, Chen H. A convolutional network with multi-scale and attention mechanisms for end-to-end single-channel speech enhancement. IEEE Signal Processing Letters. 2021. Jun 30;28:1455–9. doi: 10.1109/LSP.2021.3093859 [DOI] [Google Scholar]
  • 33. Xie Y, Liang R, Liang Z, Huang C, Zou C, Schuller B. Speech emotion classification using attention-based LSTM. IEEE/ACM Transactions on Audio, Speech, and Language Processing. 2019. Jul 1;27(11):1675–85. doi: 10.1109/TASLP.2019.2925934 [DOI] [Google Scholar]
  • 34.Zhang Y, Zhang P, Yan Y. Attention-Based LSTM with Multi-Task Learning for Distant Speech Recognition. InInterspeech 2017 (pp. 3857–3861).
  • 35. Khan TM, Robles-Kelly A. Machine learning: Quantum vs classical. IEEE Access. 2020. Dec 1;8:219275–94. doi: 10.1109/ACCESS.2020.3041719 [DOI] [Google Scholar]
  • 36. Khan TM, Naqvi SS, Meijering E. Leveraging image complexity in macro-level neural network design for medical image segmentation. Scientific Reports. 2022. Dec 24;12(1):22286. doi: 10.1038/s41598-022-26482-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Cali U, Sharma V. Short-term wind power forecasting using long-short term memory based recurrent neural network model and variable selection. Int. J. Smart Grid Clean Energy. 2019;8(2):103–10. doi: 10.12720/sgce.8.2.103-110 [DOI] [Google Scholar]
  • 38. Shan D, Zhang X, Zhang C, Li L. A novel encoder-decoder model via NS-LSTM used for bone-conducted speech enhancement. IEEE Access. 2018. Oct 4;6:62638–44. doi: 10.1109/ACCESS.2018.2873728 [DOI] [Google Scholar]
  • 39.Weninger F, Erdogan H, Watanabe S, Vincent E, Le Roux J, Hershey JR, et al. Speech enhancement with LSTM recurrent neural networks and its application to noise-robust ASR. InLatent Variable Analysis and Signal Separation: 12th International Conference, LVA/ICA 2015, Liberec, Czech Republic, August 25–28, 2015, Proceedings 12 2015 (pp. 91–99). Springer International Publishing.
  • 40. Rothauser EH. IEEE recommended practice for speech quality measurements. IEEE Transactions on Audio and Electroacoustics. 1969;17(3):225–46. [Google Scholar]
  • 41.Hirsch HG, Pearce D. The Aurora experimental framework for the performance evaluation of speech recognition systems under noisy conditions. InASR2000-Automatic speech recognition: challenges for the new Millenium ISCA tutorial and research workshop (ITRW) 2000.
  • 42. Taal CH, Hendriks RC, Heusdens R, Jensen J. An algorithm for intelligibility prediction of time–frequency weighted noisy speech. IEEE Transactions on Audio, Speech, and Language Processing. 2011. Feb 14;19(7):2125–36. doi: 10.1109/TASL.2011.2114881 [DOI] [Google Scholar]
  • 43.Rix AW, Beerends JG, Hollier MP, Hekstra AP. Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs. In2001 IEEE international conference on acoustics, speech, and signal processing. Proceedings (Cat. No. 01CH37221) 2001 May 7 (Vol. 2, pp. 749–752). IEEE.
  • 44. Yu H, Zhu WP, Champagne B. Speech enhancement using a DNN-augmented colored-noise Kalman filter. Speech Communication. 2020. Dec 1;125:142–51. doi: 10.1016/j.specom.2020.10.007 [DOI] [Google Scholar]
  • 45.Roy SK, Nicolson A, Paliwal KK. A Deep Learning-Based Kalman Filter for Speech Enhancement. InINTERSPEECH 2020 Oct (pp. 2692–2696).
  • 46. Saleem N, Gao J, Khattak MI, Rauf HT, Kadry S, Shafi M. Deepresgru: residual gated recurrent neural network-augmented kalman filtering for speech enhancement and recognition. Knowledge-Based Systems. 2022. Feb 28;238:107914. doi: 10.1016/j.knosys.2021.107914 [DOI] [Google Scholar]
  • 47. Nicolson A, Paliwal KK. Deep learning for minimum mean-square error approaches to speech enhancement. Speech Communication. 2019. Aug 1;111:44–55. doi: 10.1016/j.specom.2019.06.002 [DOI] [Google Scholar]
  • 48. Iancu B. Evaluating Google speech-to-text API’s performance for Romanian e-learning resources. Informatica Economica. 2019;23(1):17–25. doi: 10.12948/issn14531305/23.1.2019.02 [DOI] [Google Scholar]
  • 49. Wu M, Wang D. A two-stage algorithm for one-microphone reverberant speech enhancement. IEEE Transactions on Audio, Speech, and Language Processing. 2006. Apr 18;14(3):774–84. doi: 10.1109/TSA.2005.858066 [DOI] [Google Scholar]
  • 50. Al-Utaibi KA, Abdulhussain SH, Mahmmod BM, Naser MA, Alsabah M, Sait SM. Reliable recurrence algorithm for high-order Krawtchouk polynomials. Entropy. 2021. Sep 3;23(9):1162. doi: 10.3390/e23091162 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Mahmmod BM, Abdulhussain SH, Suk T, Hussain A. Fast computation of Hahn polynomials for high order moments. IEEE Access. 2022. Apr 28;10:48719–32. doi: 10.1109/ACCESS.2022.3170893 [DOI] [Google Scholar]
  • 52. Abdulhussain SH, Mahmmod BM. Fast and efficient recursive algorithm of Meixner polynomials. Journal of Real-Time Image Processing. 2021. Dec;18(6):2225–37. doi: 10.1007/s11554-021-01093-z [DOI] [Google Scholar]
  • 53. Xu Y, Liu K, Ni J, Li Q. 3D reconstruction method based on second-order semiglobal stereo matching and fast point positioning Delaunay triangulation. Plos one. 2022. Jan 25;17(1):e0260466. doi: 10.1371/journal.pone.0260466 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Abdullah M Mutawa

26 Dec 2022

PONE-D-22-32395Causal Speech Enhancement using Dynamical-Weighted Loss and Attention Encoder-Decoder Recurrent Neural NetworkPLOS ONE

Dear Dr. Saleem,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Feb 09 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Abdullah M. Mutawa, Ph.D

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1.  Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf  and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. New software must comply with the Open Source Definition.

3. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

4. Please upload a new copy of Figure 1 as the detail is not clear. Please follow the link for more information: " ext-link-type="uri" xlink:type="simple">https://blogs.plos.org/plos/2019/06/looking-good-tips-for-creating-your-plos-figures-graphics/" https://blogs.plos.org/plos/2019/06/looking-good-tips-for-creating-your-plos-figures-graphics/

5. Please include a separate caption for each figure in your manuscript.

Additional Editor Comments:

Dear Author,

Please answer all points mentioned by the reviewers and prepare a point-by-point table with reply to every point showing the location of change in the manuscript.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Causal Speech Enhancement using Dynamical-Weighted Loss and Attention EncoderDecoder Recurrent Neural Network" is a good topic for paper but having some suggestions

1) Compare results with some current developed methods and use those references in your introduction.

2) use one more different input data for analysis and evaluation.

3) use proper references

Reviewer #2: Summary:

In this work, a causal data-driven model is proposed for single-microphone SE operating in real-time systems. The proposed system utilizes attention encoder-decoder long short-term memory (LSTM) to estimate the time-frequency mask from noisy speech.

The manuscript is interesting; however, the following comment should be addressed :

Abstract :

- - - - - - - - - - -

1 – Please include problem statement .

2 - Improvement ratio between the proposed and existing works should be included .

Introduction Section :

- - - - - - - - - - - - - - - - - - - - - -

3 – In the Introduction, the authors need to refer to other speech enhancement algorithms such as : i) doi: 10.1088/1757-899X/1090/1/012102, ii) doi: 10.3390/s21217025, and ii) doi: 10.1109/ACCESS.2019.2929864.

4 – The contribution should be included as a list for better readability.

Proposed Speech Enhancement Section :

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

5 – Please check the numbering of the subsections .

6 – Define the functions used such as “|⋅|” , “||⋅||”, “*” , etc .

Experiments Section :

- - - - - - - - - - - - - - - - - - - - - -

7 – The authors utilize STFT; however, there are different types of transforms which are based on orthogonal polynomials. The authors need to refer to the difference between the Fourier Transform and the following transforms Krawtchouk transform (doi: 10.3390/e23091162), Hahn Transform (doi: 10.1109/ACCESS.2022.3170893), and Meixner transform (doi: 10.1007/s11554-021-01093-z). This will help the researchers to utilize other transforms for SE.

Conclusion Section :

- - - - - - - - - - - - - - - - - - - - - -

8 – This section is fine. No comments .

General Comments:

- - - - - - - - - - - - - - - - -

9 - There are some grammatical error should be checked and corrected .

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

quillbot-extension-portal/quillbot-extension-portal

PLoS One. 2023 May 11;18(5):e0285629. doi: 10.1371/journal.pone.0285629.r002

Author response to Decision Letter 0


23 Jan 2023

Reviewer #1: Causal Speech Enhancement using Dynamical-Weighted Loss and Attention Encoder Decoder Recurrent Neural Network" is a good topic for paper but having some suggestions

1- Compare results with some current developed methods and use those references in your introduction.

Response: Thank you for the important suggestion, the results are compared to the recently developed methods and the reference methods are reflected in the introduction part of the revised manuscript. The Tables are revised to address the suggestion of respected reviewer. Table captions are highlighted with Red to show the changes.

2- Use one more different input data for analysis and evaluation.

Response: Thank you for the important suggestion. This algorithm was intended for the additive noisy backgrounds; however, to address the concern of the respected reviewer, the authors have included noisy reverberation as other input data.

Action: The following text and results are added to the revised paper.

This section examined the dereverberation performance of the proposed SE. To train the SE model, three reverberation times (0.4 sec, 0.6 sec, and 0.8 sec) are considered. A total of 100 anechoic speech utterances from the IEEE dataset [40] are used to create the training dataset. The testing dataset contains 40 reverberant speech utterances. There is no overlapping between the speech utterances used during model training, and testing. The proposed method with reverberant speech utterances are compared and examined for the dereverberation. The results are compared with study of Wu and Wang [49], where estimated inverse filters and spectral subtraction are used to reduce the reverberation. Table 11 shows the results using STOI and PESQ. The proposed method delivered the best STOI and PESQ scores, i.e., STOI≥78.3%, and PESQ≥2.45 at RT≥4 sec. The spectrograms are provided in Fig. 8 where the smearing energy produced by reverberation is considerably reduced, showing that the reverberation performance of the proposed method.

3- Use proper references.

Response: Thank you for the suggestion, the references is arranged in the proper manner to address the concern

Reviewer#2: The manuscript is interesting; however, the following comment should be addressed

1 – Please include problem statement.

Response: Thank you for the important suggestion, the problem statement is included in the revised manuscript.

Action: Speech enhancement (SE) reduces background noise signals in target speech and is applied at the front end in various real-world applications, including robust ASRs and real-time processing in mobile phone communications. SE systems are commonly integrated into mobile phones to increase quality and intelligibility. As a result, a low-latency system is required to operate in real-world applications. On the other hand, these systems need efficient optimization. This research focuses on the single-microphone SE operating in real-time systems with better optimization.

2 - Improvement ratio between the proposed and existing works should be included.

Response: Thank you for the suggestion, the improvement ratios between the proposed and related studies are included in the revised manuscript.

3 – In the Introduction, the authors need to refer to other speech enhancement algorithms such as : i) doi: 10.1088/1757-899X/1090/1/012102, ii) doi: 10.3390/s21217025, and ii) doi: 10.1109/ACCESS.2019.2929864.

Response: The introduction part is modified with the references suggested. Thank you

4 – The contribution should be included as a list for better readability.

Response: The contributions are listed in the revised manuscript to address the concern of the reviewer. Thank you

5 – Please check the numbering of the subsections.

Response: Thank you for the correction, the sections and subsections are corrected in the revised manuscript.

6 – Define the functions used such as “|⋅|” , “||⋅||”, “*” , etc .

Response: Thank you for the important correction, the typos in equations is corrected and these notations are defined in the revised manuscript.

7 – The authors utilize STFT; however, there are different types of transforms which are based on orthogonal polynomials. The authors need to refer to the difference between the Fourier Transform and the following transforms Krawtchouk transform (doi: 10.3390/e23091162), Hahn Transform (doi: 10.1109/ACCESS.2022.3170893), and Meixner transform (doi: 10.1007/s11554-021-01093-z). This will help the researchers to utilize other transforms for SE.

Response: Thank you for the important suggestion, the authors have used STFT transform widely used in the speech signal processing. The other mentioned transforms are associated to the different applications and the authors will conduct a separate study based on these transforms. The mentioned transforms are added with the references to the revised manuscript.

Attachment

Submitted filename: Response to Reviewers.pdf

Decision Letter 1

Abdullah M Mutawa

27 Apr 2023

Causal Speech Enhancement using Dynamical-Weighted Loss and Attention Encoder-Decoder Recurrent Neural Network

PONE-D-22-32395R1

Dear Dr. Saleem,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Abdullah M. Mutawa, Ph.D

Academic Editor

PLOS ONE

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: Summary:

In this work, a causal data-driven model is proposed for single-microphone SE operating in real-time systems. The proposed system utilizes attention encoder-decoder long short-term memory (LSTM) to estimate the time-frequency mask from noisy speech.

The authors have addressed the raised comments. No further comments.

Comments:

Abstract :

- - - - - - - - - - -

1 – The abstract is fine. No further comments.

Introduction Section :

- - - - - - - - - - - - - - - - - - - - - -

2 – This section is fine. No further comments.

Proposed Speech Enhancement Section :

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

3 – This section is fine. No further comments.

Experiments Section :

- - - - - - - - - - - - - - - - - - - - - -

4 – This section is fine. No further comments.

Conclusion Section :

- - - - - - - - - - - - - - - - - - - - - -

5 – This section is fine. No further comments.

- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -

Reviewer #3: I noticed that the revised version corrects all recommendations form the reviewers, I accept the paper in this form.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Reviewer #3: No

**********

Acceptance letter

Abdullah M Mutawa

2 May 2023

PONE-D-22-32395R1

Causal Speech Enhancement using Dynamical-Weighted Loss and Attention Encoder-Decoder Recurrent Neural Network

Dear Dr. Saleem:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Abdullah M. Mutawa

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File

    (ZIP)

    S2 File

    (ZIP)

    S3 File

    (ZIP)

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting information files.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES