Skip to main content
PLOS One logoLink to PLOS One
. 2022 Nov 23;17(11):e0277975. doi: 10.1371/journal.pone.0277975

When less is more powerful: Shapley value attributed ablation with augmented learning for practical time series sensor data classification

Arijit Ukil 1,*, Leandro Marin 2, Antonio J Jara 3
Editor: Anand Nayyar4
PMCID: PMC9683574  PMID: 36417477

Abstract

Time series sensor data classification tasks often suffer from training data scarcity issue due to the expenses associated with the expert-intervened annotation efforts. For example, Electrocardiogram (ECG) data classification for cardio-vascular disease (CVD) detection requires expensive labeling procedures with the help of cardiologists. Current state-of-the-art algorithms like deep learning models have shown outstanding performance under the general requirement of availability of large set of training examples. In this paper, we propose Shapley Attributed Ablation with Augmented Learning: ShapAAL, which demonstrates that deep learning algorithm with suitably selected subset of the seen examples or ablating the unimportant ones from the given limited training dataset can ensure consistently better classification performance under augmented training. In ShapAAL, additive perturbed training augments the input space to compensate the scarcity in training examples using Residual Network (ResNet) architecture through perturbation-induced inputs, while Shapley attribution seeks the subset from the augmented training space for better learnability with the goal of better general predictive performance, thanks to the “efficiency” and “null player” axioms of transferable utility games upon which Shapley value game is formulated. In ShapAAL, the subset of training examples that contribute positively to a supervised learning setup is derived from the notion of coalition games using Shapley values associated with each of the given inputs’ contribution into the model prediction. ShapAAL is a novel push-pull deep architecture where the subset selection through Shapley value attribution pushes the model to lower dimension while augmented training augments the learning capability of the model over unseen data. We perform ablation study to provide the empirical evidence of our claim and we show that proposed ShapAAL method consistently outperforms the current baselines and state-of-the-art algorithms for time series sensor data classification tasks from publicly available UCR time series archive that includes different practical important problems like detection of CVDs from ECG data.

Introduction

With the advent of Internet of Things (IoT) and ever-increasing adoptions of sensors in the physical world, analytics problems with practical relevance are growing in numbers. One of the typical real-world challenges is to solve different classification problems, particularly that deal with time series sensor data to build sensing intelligence as one of the most useful practical implementations of artificial intelligent (AI) technique. We like to acknowledge the capability of the remarkably improved deep learning algorithms powered by the computational strength of high-powered computing infrastructure including different cloud platforms and Graphics Processing Unit (GPU)-based servers and work-stations [1, 2]. The ubiquity of smartphones and smart devices including smart bands, smart watches, smart gears, and the development of advanced sensors are playing an important role to leverage the substantial improvement in the sensing technologies to capture the physical and physiological information. High end GPU-enabled computing, cloud infrastructure, public availability of useful data sources and the emergence of powerful AI techniques like deep learning algorithms pose us the opportunity of developing vast number of worthy applications [35]. Currently, we are witnessing the learning revolution paradigm, where providing examples or training instances are often sufficient for a machine or computer to learn substantially such that it can be comparable to human-level ability. Sensors capture the physical world information from its ambience and provide the required inputs to the intelligent system such that it can sense the given physical space and perform different decision-making processes. Sensors can be considered as the micro-representation of our physical and physiological spaces.

The fundamental focus of this work is to find solutions of such practical yet diverse real world problems. Time series data are omnipresent in large set of practical applications, especially where sensor data are used to build intelligent systems. Sensors like Electrocardiogram (ECG), accelerometer, Infra-red spectroscopy, smart electric meter, etc. generate time series outputs, which motivate us such to build reliable time series classification model. For example, the important problems like cardio-vascular disease conditions like Atrial Fibrillation detection [6, 7] or Myocardial Infarction (commonly known as heart attack condition) detection from ECG data are of immense practical importance [8]. However, real-world problems come with different types of practical challenges. We particularly consider the time series sensor data classification problem, where the task is to build multi-class classification models by training given time series sensor data. We observe that large set of real-world time series sensor datasets ([8]) often suffer from the scarcity in labeled training examples for various reasons like expensive process of experimental setup or limited availability of the experimental setup (for e.g., “SonyAIBORobotSurface1” dataset [8] requires a robot to walk on different kinds of surfaces like cement or carpet) as well as the expenses associate with annotation process (“ECG200” dataset [8] requires cardiologists to annotate data whether the ECG recording is a normal sinus rhythm or Myocardial Infarction condition). “SonyAIBORobotSurface1” dataset contains mere 20 number of training examples, “ECG200” contains 100 number of training examples. Traditionally, deep neural networks require large set of training datasets for reliable and generalized learning. For example, CIFAR-10 dataset consists of 50,000 training images, while classical ImageNet 2012 classification dataset consists 1.28 million training datasets [9, 10]. CIFAR-10 and CIFAR-100 are actually labeled subsets of 80 million image [11]. Such abundance of training dataset availability is infeasible in case of practical time series sensor signal analysis problems. In fact, deep learning algorithms rely on the sufficiency of the training examples with the assumption that the learned embeddings preserve latent structures and the distribution of the given time series data [12].

Typically the solution of training data limitation is tackled by augmented learning through adversarial training [1315], where the input training space is augmented through perturbation. Adversarial examples, which in simple terms are the perturbed forms of the input training data, have potential benefit as a data augmentation method to solve the training data scarcity issue [16, 17]. However, adversarial examples need finer control and it is shown that adversarial training mostly positively helps when the training data is sufficient and hurts the accuracy when training data is small in size [18].

On the other hand, we understand that suitable feature space has immense impact on the model learning. If an apt feature set or in our context, appropriate inputs in terms of training data are provided to a suitable deep learning model, the learned model can have better prediction capability. In this paper, we consider Shapley value [19, 20] to estimate the importance of each of the inputs of the model towards the prediction. Shapley values attempt to fairly commensurate a player’s contribution in a coalition game. In fact, Shapley value estimation has been applied in diverse disciplines [5, 21]. We incorporate Shapley Value attribution to discard the unnecessary or negatively impacting input data. While augmented learning using adversarial training provides a generic augmentation of the given time series data, the augmented-learned model when gets trained with the suitable input subset using associated Shapley values ensures better learnability. In ShapAAL, data augmentation and input ablation jointly provide the impetus towards learning with better data. ShapAAL can be considered as a push-pull architecture, where augmented learning pushes the model towards getting trained by learning newer (adversarial) examples and Shapley value estimated subset selection pulls the model towards a suitable lower dimension for better learnability and prediction. We introduce the concept of Learn → Unlearn → Re-learn, where the model is initially learned through augmented training; next, Shapley value attribution forces the model to unlearn few detrimental features and subsequently, the model re-learns with the selected subset features using augmented learning. With series of empirical study, we demonstrate the efficacy of our proposed model ShapAAL: Shapley Attributed Augmented Learning and establish the performance superiority over relevant state-of-the-art algorithms.

Related works

Sensor data-centric classification tasks are mostly likely to undergo the training data scarcity issue owing to its universally acknowledge problem of high expenses and difficulties associated with the generation, collection and the cost with labeling by human experts [22]. Classically, the emphasis was to analyze the time series (given that the senor data is a time series) and to build strong classifiers to solve time series classification tasks [8]. Nearest neighbor based classification with distance function as dynamic time warping distance (1NN-DTW) has been traditionally considered as the classical baseline algorithm for time series classification [23]. COTE or Collective of Transformation-based Ensembles is an ensemble learning algorithm with collection of 35 classifiers [24]. Random Interval Spectral Ensemble (RISE) algorithm builds decision trees with set of Fourier, auto-correlation and partial auto-correlation features and perform ensembling operation [25]. Recently, Time Series Combination of Heterogeneous and Integrated Embeddings Forest (TS-Chief), an tree-based ensemble learning classifier is proposed [26]. In fact, Time Series Forest (TSF) is one of the pioneering works that combine entropy gain with a distance measure to provide evaluation of the split in tree-based ensembling learning [27]. Similarly, Proximity Forest, which is ensembles of highly randomized proximity trees is another ensemble learning algorithm that has been developed for time series classification tasks [28]. Recently, CAnonical Time-series CHaracteristics (Catch22), a feature-engineered time series classifier is proposed that has shown promising results [29]. With deep learning models showing outstanding performances in computer vision tasks, time series classification also employs strong deep learning architecture like Residual Network (ResNet) [30]. In [31], authors have proposed convolution layer-based residual blocks to develop ResNet-based model, which is considered as a strong baseline for time series classification tasks.

The state-of-the-art techniques as cited above are mostly concerned with the development of a decent time series classification model without the consideration of training data scarcity. It is observed in [22] that time series classification tasks need to emphasis on training data scarcity issue in order to construct a practical analytics system. However, the research direction towards mitigating the learning impairment problem due to training instance insufficiency in time series classification under common machine learning or deep learning framework seems to be an open practical challenge. The typical research attempts are focused towards sophistication of the architecture and detailed extraction of time series representation. Under the constraint of inadequate availability of training set, such attempts may not always be the ideal choice and the diversity of time series applications limit the scalability of such models. In this paper, we propose a novel method ShapAAL, that contains intrinsic capability to demonstrate consistent accurate performance and improves upon the state-of-the art models through the learn, unlearn and re-learn principle of learning with positive impact towards the predictive capability of the model.

In general, machine learning algorithms need to carefully select the supervised features to build a robust model [32]. Optimization method plays an important role in various aspects towards better learned model development under practical constraints [3339]. For instance, evolutionary processes with consistent equilibrium for high-quality performance and optimization that achieves quicker convergence is proposed in [35]. It is well-known that the search for global optimization in deep learning algorithms often suffer through spurious local optimization issues. In [36], fusion-based meta-heuristic optimization methods are proposed to solve global optimization tasks.

Materials and methods

Problem sketch

We focus on time series classification tasks for sensor signal analysis, where typically a time series is represented as an ordered set of real values as: x = [x1, x2, x3, …, xT], xRT and x is of length T and x1, x2, x3, …, xT are the scalar measurements at time intervals 1, 2, 3, …, T from a given sensor. For example, an ECG signal x contains continuous time stamp measurements, where the first time stamp measurement is denoted as x1, the second time stamp measurement is denoted as x2 and so on.

Consider a set of N examples that constitute the training dataset XTrain=[x(1),x(2),,x(N)], where each of x(n), n = 1, 2, &, N is a time series and each of which consists of T number of data samples, i.e. each training instances can be considered as consisting of T number of time stamp measurements from the given sensor. The complete training set consists of corresponding labels: DTrain=[XTrain,YTrain]=[{x(1),y(1)},{x(2),y(2)},,{x(N),y(N)}] y(n)[1,C], ∀n are the labels correspond to one of the C classes. We are particularly concerned to solve the supervised learning tasks for time series classification problems such that a model is constructed from the given input variables or training instances along with its associated labels or ground truths such that model correctly attempts to predict the class that a sensor data belongs to. In supervised learning setting, we find a model or function hθ(.), parameterized by θ that describes the random vector x associated with label or target y with joint distribution pdata(x, y). However, we tacitly assume that [{x(1),y(1)},{x(2),y(2)},,{x(N),y(N)}]i.i.d.pdata, which means that the model learning is imposing independently and identically distribution (i.i.d.) condition, i.e. the given training examples are drawn independently and identically from pdata.

In machine learning, the principal aim is to minimize an objective function that penalizes the model hθ(.) when it makes mistake, which is denoted by loss function as L(hθ(x), y). We consider the expected risk in Eq 1.

R(h)=E(x,y)pdata[L(hθ(x),y)] (1)

However, we do not have complete idea of pdata(x, y), we simply know the training dataset DTrain=(x(n),y(n)). Hence, we focus on empirical risk minimization (ERM), which is defined in Eq 2.

R^emp(h)=1Nn=1NL(hθ(x(n)),y(n))) (2)

Considering negative log-likelihood as the loss function under maximum likelihood estimation (MLE) principle, which is a special case of ERM, we the MLE cost function as follows:

J(θ)=E(x,y)p^data-logpθ(y|x)

Thus, we need to minimize the cost function J(θ) to find the parameter θ from the empirical distribution p^data. The optimization problem is as follows:

θ*=argminθJ(θ)

The practical challenge in time series classification is that N and T are typically much smaller than (or usually as expected by) the corresponding learning methods that solve classical computer vision problems [40]. For example, N can be as low as 50 and often less than 200 and T can be less than 300 for time series classification tasks and on the contrary, classical ImageNet 2012 classification dataset consists of 1.28 million training datasets [9, 10]. However, when N is small (which is profound in time series sensor signal classification tasks), it is not practical to assume the closeness of p^data and pdata and consequently, the learned model tends to get over-fitted to the given training dataset DTrain. Hence, the estimation from J(θ) is poor when the model is directly constructed from the available training dataset DTrain.

Let ηR be the learning rate and the gradient descent function updates the deep neural network model parameter θ as follows: θθηJ(θ). In training data insufficiency problem, the estimation from J(θ) is incomplete, which in high probability leads the model parameter θ gets directed towards incomplete or wrongly learned direction from usual gradient descent method. Therefore, we can safely assume that learning degradation is a common problem and approaches that minimize the learning degradation due to training data scarcity needs to be developed in order to achieve better performance from time series classification tasks. Thus, our objective is to find a “good” learned model for diverse set of time series classification tasks (especially where the data is sourced from sensors) in order to minimize the adverse effect of limited training data availability. In a nutshell, the above problem formulation is a generic one that motivates us to build robust time series sensor data classification models, which often suffer due to insufficiency in the training instances.

Proposed methodology

We propose three stage approach of deep learning model construction, where the first stage learns through augmented learning with additive perturbation of the input samples. Next, the unlearning part identifies the subset of features or the training samples that do not have positive contribution to the model predictive capability through Shapley value computation for each of the input samples. Finally, a new model is re-learned with the subset of samples or with the identified important samples of the training set.

Model training for augmented learning

We consider Residual Network or ResNet [30] architecture with controlled perturbation of input space that compensates the lack of training data for time series classification tasks. We consider adversarial perturbation as the set of invariants such that a robust model can be constructed under the practical constraints of training sample scarcity that attempts to minimize the worst-case classification error due to the data perturbation by the adversary [15]. The adversary, in turn augments the training space as an automated (machine generated) labeler, replacing the human labeler. Hence, we not only gain in the enrichment of the training process, but also avoid the expensive process of collection and labeling of time series examples. The adversarial perturbation is to force the classifier to learn hidden representations of unseen neighbor feature in order to estimate the true distribution pdata. Let Jadv(θadv, x, y) be the cost (associated with the adversarial loss Ladv for training the network (in our case, we primarily consider the neural network as ResNet [30]) to derive the model parameter θadv. ResNet has shown tremendous success in different classification tasks. It aims to tackle the learnability issue of deep neural networks by minimizing the exploring and vanishing gradient problems through norm preservation of error gradient [41]. ResNet transforms the traditional representation learning to learn H(x)=F(x)+x at each layer [30] as depicted in Fig 2, where one typical Residual Block (RB) is shown. The main advantage of H(x) is to ensure that the information in x flows throughout the network [41]. In ResNet, the original mapping is recast into F(x)+x [30] and it is hypothesized that optimization of the residual mapping becomes easier [30]. In fact, the identity or shortcut connection does the desirable effect of norm preservation as error gradient [41] as shown in Fig 1. In ShapAAL, we transform xx + δ for perturbed identity and it learns through H(x+δ)=F(x)+x+δ. Therefore, the transformation of xx + δ into the identity connection augments the learnability through perturbation-induced shortcut connection. The identity connection becomes perturbation connection as shown in Fig 2.

Fig 2. A residual block (RB) in ShapAAL with perturbed input.

Fig 2

Fig 1. A residual block (RB) in ResNet ([30]).

Fig 1

We design the ShapAAL model with ResNet architecture using restrained learning principle [42]. It consists of variable number of residual blocks (RB) between 10 and 2 and the residual block depth (i.e. the number of total RBs in the model) depends on the training data. We estimate the network depth (measured in term of number of RBs) using restrained learning principle that analyzes the training dataset distribution to adjust the network depth. We depict the typical RB of ShapAAL in Fig 3.

Fig 3. Single residual block in ShapAAL architecture.

Fig 3

We convert the 1D data to 2D through reshape operation such that the features of 2D convolutions can be utilized. The batch size is variable, depending upon the number of training instances. The batch size is calculated as: min(ceil(numberoftraininginstances10),16). When the training examples are small in number (≤10), batch size of 2 is considered. We consider fixed learning rate 10−3, which is default in Keras. We z-normalized the training data as: XTrain-mean(XTrain)standarddeviation(XTrain) as well as the test data as: Xtest-mean(XTrain)standarddeviation(XTrain). We first calculate x + δ and after that z-normalization is performed. Please note that the statistical estimation of z-normalization operation for test data is made from the provided train data as the statistics of test data is unknown. After the final residual block, Global Average Pooling is used. We use softmax function in the output layer for the classification task and cross-entropy as the loss function. The input data for identity connection is x + δ. Depending upon the estimation of residual block depth from restrained learning algorithm [42], the number of RBs are constructed. Let, the number of RBs be χ, where 10 ≤ χ ≤ 2. For different XTrain, the value of χ might be different owing to the differences of the underlying training distribution and accordingly, the model with χ number of RBs is constructed. We illustrate ShapAAL architecture in Fig 4.

Fig 4. ShapAAL architecture.

Fig 4

The expected risk of ShapAAL under augmented learning is defined in Eq 3.

Raug(h)=E(x,y)p^data[maxδΔL(hθ(x+δ),y)] (3)

where Δ represents the set of adversarial perturbations in δ to induce mis-classification. The input is perturbed with noise δ such that the network gets the opportunity to learn training examples outside the given training set while the unperturbed shortcut connection makes the gradient to avoid being trapped into the spurious local optimum [43]. Hence, we hypothesize that the network learns well with the identity propagation of a ResNet model through shortcut connection that guides the algorithm to move easily towards global optimum [43] and the perturbed input space forces the model to learn unseen examples through augmented learning gain. The perturbation needs to be controlled and introduction of controlled additive perturbation compels the learning to be more generic and it learns examples beyond the given training data. We perturb the input data by adding small amount of Gaussian noise δ and the parameters (mean and standard deviation) of the Gaussian noise δ are derived from XTrain, i.e. δ is sampled from N(μ,σ2), where, μ being the mean of XTrain and σ2 being the variance of XTrain. In order to maintain a reasonably high signal-to-noise ratio between the perturbed data and original data, a scaling factor α is introduced and δ is sampled from α×N(μ,σ2). In fact, we change the view of a training set to a learner system such that the leaner’s robustness is examined and eventually we expect a stronger model with higher generalization gain and lower over-fitting error. We have considered α = 0.020 throughout the experimental process. The controlled additive perturbation enforces the ResNet model to learn the less confident solutions to lower the generalization loss. Hence, the learnability of the model improves when it faces newer types of challenges (as expected in the field when unseen test data are encountered).

We attempt to generate the augmented learning model to minimize the generalization loss by introducing perturbation into the learning space. Further, we incorporate restrained learning for adjusting the depth of the network (more precisely, the number of residual blocks), which is training data distribution-aware [42]. For a given training data DTrain=[XTrain,YTrain], we estimate the network depth through restrained learning approach of elastic depth estimation [42]. Elastic depth minimizes the negative impact of data perturbation. When the perturbed training data results in redundancy, the network depth shrinks and vice-versa. The restrained learning which dynamically configures the network depth acts as a regularizer to restrict the learning when the data redundancy due perturbation process is high. Let us denote the adversarially trained augmented model with adversarial risk Raug(h) minimization as Maug.

Subset selection from input samples for model re-learning

Augmented training has the advantage of better learning due to perturbation in the learning process, but such learning may not always do the good for model predictability. Using adversarial training for data augmentation requires to know the worst-case δ that augments the training data “most beneficial” way with “highest confusion creation” to the training data [44]. However, such search is computationally (extremely) expensive. We propose that an apt process of important feature selection or sampling the input data ensures the required better learning for the model. Hence, we identify DTrainsub=[XTrainsub,YTrainsub] which is a subset of DTrain, i.e. DTrainsubDTrain such that the samples with positive impact on the model predictability are chosen.

Our objective is to estimate the importance of a feature or training sample x(n)XTrain such that the worth of x(n) is significant to consider it as an important and positively contributing sample. We use Shapley value [19, 20], a fundamental concept in transferable utility cooperative game theory [45] to quantify the attribution of x(n) in the prediction capability of the constructed model. Let N be a finite set of training samples (in cooperative game theory context, we call the training samples as players) or player [46, 47].

  1. (Definition I) (Transferable utility game). We define a game that maps v:2NR such that v(∅) = 0. We interpret v(ψ) where ψ in 2N, as the estimated value of coalition ψ. The value function v(ψ) intends to identify the collective payoff a player’s or a set of players’ gain when they cooperate and the model M is trained with nth sample on all possible subset ψ ⊆ 2N.

  2. (Definition II) (Marginal contribution). We define the marginal contribution Δv(n, ψ) of player n with respect to the coalition ψ as: Δv(n, ψ) = v(ψn) − v(ψ).

With Λ being denoted as the integer permutations up to N and λ ∈ Λ and we represent the predecessor set of players preceding nth player in λ as: ψn = {m: λ(m)<λ(n)}. With this definition, Shapley value φn of nth player is formulated as the weighted average of the marginal contribution of it to all other possible subset of players in the game. Accordingly, Shapley value φv(n) of nth player with the function v is:

ϕv(n)=1N!λΛΔv(n,ψn,λ).

From the permutation logic, we can compute the Shapley value φv(n) of nth training sample as:

ϕv(n)=1N!ψ{1,2,3..,N}|ψ|!(N-|ψ|-1)!Δv(n,ψ).

The above equation needs to be solved to get the estimation of Shapley value for each of the training examples in N, but that process is computationally expensive. In this paper, we consider the high-speed approximation of φv(n) using DeepLIFT algorithm [48] with DeepExplainer implementation (https://github.com/slundberg/shap).

From the computed Shapley values φv(n), ∀nN for each of the training samples in DTrain=[XTrain,YTrain], we discard the negative valued ones, i.e. the training samples which contain negative magnitude in their Shapley value are removed and new training set with NeffectiveN number of training examples are formed and the expected risk of ShapAAL is depicted in Eq 4.

RShapAAL(h)=E(x,y)p^data[maxδΔL(hθ(xs+δ),ys)] (4)

where, {xs, ys} belong to the Shapley value attributed dataset. A set of axioms namely “efficiency” and “null player” are the prime motivations to claim that the context of Shapley value for finding out the right subset [46, 47].

  1. (Axiom I) (Fairness). The worth of a complete model v(N) in a transferable utility game is a lossless distribution among the given features: ∑nN φ(n) = v(N).

  2. (Axiom II) (Null player). If a feature n contributes nothing in a transferable utility game v, its Shapley value is zero. [(∀ψ)v(n ∪ {n}) = v({n})] ⇒ ϕ(n) = 0.

Axiom I and Axiom II help us to develop the subset selection algorithm. Let us denote the newly formed training set with Neffective number of training samples as DTrain_effectiveDTrain. The unlearning part gets completed with newly formed training set DTrain_effective. The model previously learned with DTrain with adversarial risk Raug(h) (Eq 3) minimization as Maug, subsequently, re-learns DTrain_effective by minimizing textitRShapAAL(h) (Eq 4) to construct ShapAAL model MaugShapAAL through unlearning the negatively impacting dataset. When training dataset is large, the negative contribution of few data may not have some impact, but in case of smaller number of training datasets, the negatively contributing ones can have higher impact on the learning of the model. Classically, the model learning flow is: training data → model training → classification by the trained model. With data augmentation training the flow is: training data → augmentation → augmented model training → classification by the augmented trained model. With Shapley value-vbased feature attribution the training flow is: Training data → Subset selection from the knowledge of Shapley values of each of the input data → Shapley-attributed model training → Classification by the Shapley-attributed model. We propose the model training algorithm ShapAAL that takes advantage of augmented training for training space augmentation as well as subset selection through Shapley value-attribution as defined in Eq 4. The proposed model training flow is: training data → augmentation → augmented model training → Subset selection from the augmented set wit the knowledge of Shapley values → Shapley-attributed augmented model training → classification by the Shapley-attributed augmented model. We depict the ShapAAL Algorithm 1 below.

  1. Construct the model Maug with adversarial risk of augmented learning Raug(h) from Eq 3 through risk minimization from the given training dataset DTrain=[XTrain,YTrain], with N number of training examples according to the ShapAAL architecture in Figs 24, where input is perturbed with noise δ to provide the network with the capability to learn outside the given training example set.

    Learning part

  2. With the model Maug as reference, for each of the training instances nN, φ(n) Shapley values are computed from DeepLIFT algorithm [48].

  3. We find those n where φ(n)≤0, (Axiom II) which create the set of N′ number of examples, N′ ≤ N.

  4. Discard those N′ number of training samples and rest Neffective training samples create new training dataset DTrain_effective, where, Neffective = N-N′.

    Unlearning part

  5. ShapAAL model MaugShapAAL is generated by training with DTrain_effective containing {xs, ys} according to the ShapAAL architecture in Figs 24, where Shapley attributed inputs xs are additively perturbed with noise δ to construct ShapAAL model MaugShapAAL by minimizing RShapAAL(h) from Eq 4.

    Re-learning part

In summary, we define a transferable utility game for the selecting useful training data, which are made inputs to the learning algorithm. The attribution of each of the inputs into the model predictability is estimated through Shapley value computation. The non-contributing inputs defined according to Axiom II are discarded as the not worthy inputs and the remaining inputs are used for re-learning the model. The subset finding operation is performed over the perturbed set with the assumption that the perturbed input space provides augmentation when the training data is insufficient.

Results

We conduct series of empirical studies to investigate the performance efficacy of ShapAAL in time series sensor data classification tasks particularly when the training data sample size is small.

Data description

Currently, UCR [8] is one of the most recognized time series classification benchmark archives [49]. We find number of time series sensor datasets along with three important ECG datasets which fulfill our criteria of being limited in number of training instances (≤200). The datasets are sourced from sensing devices. These datasets are diverse in different characteristics like sensor types, number of training examples, length of the data etc. as depicted in Table 1. Each of the time series datasets in UCR has fixed and exclusive training and testing splits. The test data is completely hidden. In this work, we have generated the learning model using the training datasets and the trained model is tested on the provided testing datasets and the ‘test accuracy’ (as per the convention in the UCR time series archive benchmark [8, 49]) is considered as the classification inference performance measure.

Table 1. Experimental dataset description.

Dataset Training size Data length Sensor type Application type
ChinaTown 20 24 IR Sensor Pedestrian counting
Coffee 28 286 Food Spectrometer Detection of two different coffee types for food safety and quality assurance
ECG200 100 96 Electrocardiogram (ECG) Myocardial infarction detection
ECGFiveDays 23 136 Electrocardiogram (ECG) Change detection
FreezerRegularTrain 150 301 Smart energy meter Engery Efficiency in domestic appliances
FreezerSmallTrain 28 301 Smart energy meter Engery Efficiency of freezers
ItalyPowerDemand 67 24 Smart energy meter Power demand identification
MoteStrain 20 84 Humidity sensor Weather feature
PowerCons 180 144 Smart energy meter Household energy consumption pattern identification
SonyAIBO1 20 70 Accelerometer Surface type identification
SonyAIBO2 27 65 Accelerometer Surface type identification
TwoLeadECG 23 82 Electrocardiogram (ECG) Change detection

Development environment

ShapAAL is implemented in Keras 2.1.2 on Python 3.5.4 on Tensorflow 1.4.0 library. The hardware environment for training the model consists of 64-bit x86 architecture 16 cores Intel Xeon CPU E5–2623 v4 with 2.60GHz clock speed with two Nvidia GeForce GTX 1080 GPUs, which are powered by Pascal architecture and each of the GPUs has 10 GB memory. We have used DeepExplainer implementation of DeepLIFT algorithm [48]. In DeepExplainer (https://github.com/slundberg/shap), a distribution of background samples is used instead of a single reference point in DeepLIFT. In order to minimize the impact of non-reproducibility (https://glaringlee.github.io/notes/randomness.html) with run-to-run variability due to nondeterminism in neural networks [50, 51], we have considered at least 50 different random seeds for each of the experimental datasets and the reported empirical results are the highest occurring (mode) of the obtained test accuracies.

Empirical investigation

We perform number of empirical investigations including ablation study, comparative study with relevant baselines and state-of-the-art algorithms to illustrate the practical utility of ShapAAL when performing diverse set of real-world sensor time series classification tasks including that of critical prediction task of Myocardial Infarction condition detection from ECG sensor. We show in Figs 59 that the Shapley value responses of the training instances in different datasets. It is clearly observed that few of the training instances are in fact negatively impacting towards the model prediction and it is evident in practice that each training samples are not essentially contributing positively towards model’s prediction.

Fig 5. Estimation of input attribution for “ECGFivedays” dataset.

Fig 5

Fig 9. Estimation of input attribution for “SonyAIBOSurface1” dataset.

Fig 9

Fig 6. Estimation of input attribution for “FreezerSmallTrain” dataset.

Fig 6

Fig 7. Estimation of input attribution for “ItalyPowerDemand” dataset.

Fig 7

Fig 8. Estimation of input attribution for “MoteStrain” dataset.

Fig 8

Subsequently, in Fig 10, we depict the distribution of subset selection from Shapley attribution in ShapAAL, where in some cases, more than 30% of training samples are rendered unimportant and subsequently discarded in the process of learning.

Fig 10. Selection of subset by ShapAAL algorithm for different datasets.

Fig 10

Next, we conduct ablation study to understand the efficacy of the proposed model. An ablation study in general, investigates the performance of a machine learning system by removing few components in order to evaluate the impact of those components in the complete system. Similarly, ShapAAL model construction consists of four components that include the base model (ResNet), Shapley value attribution over the base model, data augmented training on the base model and data augmented training with Shapley attributed feature selection on the base model. We denote M as the base model that is trained with each of the training data, MShapley as the model that is trained with the training data after discarding the negatively contributing Shapley valued features, Maug is the model that is adversarially trained over over entire augmented training data. MShapAAL or MaugShapley is the adversarially trained with the augmented training data with discarding the negatively contributing Shapley valued ones following the deep architecture in Fig 4 as depicted in Figs 1114. In Table 2, we depict the “test accuracy” performances of M, Maug, MShapley and MaugShapAAL over the experimental datasets. The ablation study unambiguously indicates that our proposed model MShapAAL is the superior one. In fact, the trend is also clear that both augmented training and Shapley attributed re-learning have significant positive impact on the learnability of the model, which reflects in the consistent superlative performance of MShapAAL w.r.t the others. Conceptually, the ShapAAL model is evolved from the base model M, which learns from R(h)=E(x,y)pdata[L(hθ(x),y)]. Maug model, derived from M directly helps the base model M to get trained over unseen training examples due to additive perturbation with benefit of addressing training data scarcity problem by learning from Raug(h)=E(x,y)p^data[maxδΔL(hθ(x+δ),y)]. On the other hand, model MShapley gets trained over a subset of the seen training examples according to Shapley-value attribution that discards the non-important input features. Our proposed model considers the strengths of both Maug and MShapley to construct an unique deep learning algorithm that renders data augmentation as well as input feature reduction (i.e. getting advantages from adversarial training and apt feature selection) to allow the ResNet base model M to appropriately learn over augmented yet selected input set. Hence, we establish with the empirical support that less number of input features (Refer Fig 10) when properly selected can provide better test accuracy. Under training data size constraint scenario, the push-pull architecture of ShapAAL as a coalition game with Shapley attributed push towards lower dimension and concurrently pulling or augmenting the learning capability of the model over unseen data indeed demonstrates significantly improved performance.

Fig 11. Constructing the base model M.

Fig 11

Fig 14. Constructing the proposed model MShapAAL.

Fig 14

Table 2. Ablation study through test accuracies of ShapAAL model (MShapAAL) with M, MShapley, Maug.

Algorithm M M Shapley M aug M ShapAAL
ChinaTown 0.890 0.901 0.9211 0.9722
Coffee 0.976 1.00 0.998 1.00
ECG200 0.83 0.86 0.87 0.92
ECGFiveDays 0.989 1.00 1.00 1.00
FreezerRegularTrain 0.9865 0.9901 0.9933 0.9984
FreezerSmallTrain 0.8640 0.8640 0.8613 0.9309
ItalyPowerDemand 0.8910 0.8901 0.9356 0.9704
MoteStrain 0.8101 0.8233 0.9087 0.9084
PowerCons 0.8576 0.8571 0.9083 0.9633
SonyAIBO1 0.8121 0.8439 0.8907 0.9682
SonyAIBO2 0.9355 0.9451 0.9406 0.9461
TwoLeadECG 0.8860 0.9006 0.9304 0.9994

Fig 12. Constructing the Shapley-value ablated model MShapley.

Fig 12

Fig 13. Constructing the augmented learning model Maug.

Fig 13

Given that generic time series classification is well-studied [8], we do an exhaustive comparative study with the baseline algorithms like 1NN-DTW-based model [23] as well as state-of-the-art methods including RISE [25], COTE [24], TS-Chief [26], Time Series Forest (TSF) [27], Proximity Forest (PF) [28], Catch22 [29], and time series ResNet [31]. In Table 3, the comparative study of test accuracies of relevant state-of-the-art algorithms are shown and we observe that ShapAAL consistently outperforms the state-of-the-art algorithms.

Table 3. Comparative study of test accuracies of ShapAAL model with baseline and state-of-the-art algorithms INN-DTW ([23]), COTE ([24]), TS-Chief ([26]), ResNet ([31]), PF ([28]), RISE ([25]), TSF ([27]), Catch22 ([29]).

Algorithm INN-DTW([23]) COTE([24]) TS-Chief ([26]) ResNet([31]) PF([28]) RISE([25]) TSF([27]) Catch22([29]) ShapAAL
ChinaTown 0.9618 0.9701 0.94801 0.8885 0.9529 0.9344 0.9722
Coffee 0.821 1.00 0.9904 0.9964 0.9916 0.9845 0.9869 0.9797 1.00
ECG200 0.88 0.88 0.855 0.8836 0.873 0.851 0.86 0.7886 .92
ECGFiveDays 0.7967 1.00 0.9988 0.9510 0.8828 0.97286 0.9519 0.8158 1.00
FreezerRegularTrain 0.9984 0.9967 0.9423 0.9522 0.9970 0.9981 0.9984
FreezerSmallTrain 0.9954 0.9494 0.8233 0.8787 0.9614 0.9597 0.9309
ItalyPowerDemand 0.9553 0.9611 0.9624 0.9571 0.9560 0.9445 0.9594 0.8774 0.9704
MoteStrain 0.8658 0.9369 0.9301 0.9031 0.9149 0.8780 0.8554 0.8484 0.9084
PowerCons 0.9794 0.8861 0.9874 0.9579 0.9931 0.8862 0.9633
SonyAIBO1 0.6955 0.8453 0.8897 0.9603 0.9201 0.8669 0.8637 0.8833 0.9682
SonyAIBO2 0.8594 0.9517 0.9010 0.9688 0.8990 0.9124 0.8743 0.9023 0.9461
TwoLeadECG 0.86 0.993 0.9900 0.9994 0.9817 0.9107 0.8706 0.8539 0.9994

marked results are not available.

Another classical performance merit is the “outperforming” the benchmark. In recent years, number of time series classification algorithms have been proposed in literature, which might not have been updated in the UCR archive repository. However, we can consider the available benchmark or the best results in the UCR repository of the respective datasets as the “reported benchmark”. In Fig 15, we depict the differential test accuracy gain of the algorithms (which has reported results available in public domain) including ShapAAL model w.r.t the reported best results and it is computed as testaccuracyofthealgorithm-reportedbenchmarktestaccuracyreportedbenchmarktestaccuracy with the aim of being the test accuracy result to be positive, indicating that the concerned algorithm has outperformed the currently reported benchmark result. We observe that proposed ShapAAL steadily outperforms the reported benchmark results in comparison with the relevant benchmark algorithms.

Fig 15. Differential test accuracy gain of different algorithms and proposed ShapAAL from the current reported (best) benchmark results.

Fig 15

Mean Per-Class Error (MPCE) ([31])is another useful metric to evaluate the classification performance of the model as: the expected error rate for a single class across each of the test data. For Υ number of test data with class cυ and corresponding error rate errυ, we compute MPCE as: 1ϒerrυcυ.

MPCE seems to a robust as an evaluator of model performance for different datasets of the classes [31]). Below in Table 4, we demonstrate the MPCE results for the ablation study. In MPCE, our aim is to have a lower value, approaching zero.

Table 4. Ablation study through MPCE of ShapAAL model (MShapAAL) with M, MShapley, Maug.

Algorithm M M Shapley M aug M ShapAAL
ChinaTown 0.1115 0.1002 0.0751 0.0281
Coffee 0.0409 0.0 0.011 0.0
ECG200 0.1785 0.1454 0.1417 0.0846
ECGFiveDays 0.0174 0.00 0.0 0.00
FreezerRegularTrain 0.0119 0.0085 0.0079 0.0021
FreezerSmallTrain 0.1587 0.1587 0.1532 0.0879
ItalyPowerDemand 0.1089 0.1082 0.0649 0.0305
MoteStrain 0.1981 0.1772 0.1045 0.1049
PowerCons 0.1459 0.1457 0.1006 0.0457
SonyAIBO1 0.1870 0.1640 0.0995 0.0401
SonyAIBO2 0.0615 0.0604 0.0588 0.0583
TwoLeadECG 0.1239 0.0902 0.0689 0.0012

Another unique feature of the current work is its response to higher number of test instances when it gets trained with smaller number of training examples. We can quantify the learning gain of ShapAAL at the time of testing as: testaccurcayShapAAL-testaccurcayBasetestaccurcayBase and also define training insufficiency factor as: NumberoftrainingexamplesNumberoftestinginstances. In Fig 16, we demonstrate the comparative study of learning gain of ShapAAL on testing data over base model and the insufficiency in the training. We observe that the learning gain of ShapAAL is mostly ≥1, while training insufficiency factor ≤1. Hence, we further establish our claim that ShapAAL model is the apt choice under practical constraint of training data limitation in solving the time series classification tasks.

Fig 16. Empirical support of consistency in learning gain over test data of ShapAAL (MShapley) over base model (M) under typical practical constraint of training insufficiency factor ≤1.

Fig 16

The significance of ShapAAL as a time series sensor data classification model is well-established both from ablation study (Table 2) and comparative study with current state-of-the-art algorithms (Table 3, Fig 15). ShapAAL not only improves upon through joint augmented training and Shapley value based feature attribution, but also it creates new benchmark in time series sensor signal classification tasks. With the support of the above empirical study, we claim that ShapAAL is the apt choice for time series classification tasks under the practical constraint of training data insufficiency. The proposed model attempts to maximize the worst-case classification accuracy owing to the presence of data perturbation, which in philosophy, expands the training space to act as machine generated annotator that creates the possibility of human annotator replacement. Hence, another substantial gain we incur other than better learnability with enriched training process is the avoidance of expensive data labeling processes.

Discussion

It is well-established in literature with empirical evidences in support of neural scaling law, which hypothesizes that the test error generally decreases as a power law with the number of training data, i.e. more training data is often beneficial for the learnability of a deep learning model and motivated by this neural scaling law, significant investments have been made in data collection [52]. In this work, we have presented our novel ShapAAL algorithm that can potentially overcome the limitation of practical scenarios of insufficiency of training data while performing time series classification tasks including practically important application of cardio-vascular disease detection from ECG recordings. ShapAAL augments the learning method such that unseen training examples are made part of the model learning process along with selection of important training instances through Shapley value computation such that only positively impacting data are included while constructing the computational model. The conventional Shapley value-based feature subset identification relies upon choosing k highest ranking ones [46]. However, aprior knowledge of k is practically infeasible. For instance, the “best” result may be k = 90% or may be k = 100% or k = 60%. Hence, the classical approach is not the appropriate choice. Our proposed algorithm is intuitively appealing and principled upon the “Efficiency” and “Null player” Shapley value axioms [46, 47], which is theoretically sound, tractable and practically feasible and supported with empirical investigation as depicted in Tables 2 and 3. Firstly, we have proposed and validated the unique idea augmentation and ablation of the input features to generate a better learned model. Controlled augmentation of the seen examples to learn better on the unseen examples through introduction of perturbed or virtual data points helps the model to combat the insufficiency in training examples and Shapley-attributed input feature selection refines the input space such that the model gets the opportunity of training more (through augmentation) yet better (Shapley-value based feature ablation). While the augmentation and feature attribution separately improve the test accuracy of the model over different tasks, the combined effect is significant, and it is evident from Tables 2 and 4. The study in Tables 2 and 4 clearly indicates that data augmentation through adversarial learning and subsequent feature space identification for re-learning with appropriate features provide significant impetus to the learning process to learn that compensates the limitation in seen examples and learn appropriately. Secondly, we have provided state-of-the-art comparison of the proposed method and the ShapAAL model with both data augmentation and input attribution features has demonstrated consistently outstanding classification performances over different time series classification tasks, conveniently outperforming the current benchmark and state-of-the-art algorithms as depicted in Tables 3 and 4 and Fig 15.

From a purely pragmatic standpoint, ShapAAL has demonstrated capability of accurately performing diverse set of time series sensor signal classification tasks including identification of time-critical conditions like Myocardial Infarction or heart attack using ECG signal and consistently outperform the state-of-the-art algorithms. Smartphone-based ECG applications are indeed one of the important practical utility of IoT and AI technology [4]. It is known that cardio-vascular diseases are leading cause of human deaths globally [53]. We envisage that the automated ECG analysis is capable of ensuring on-demand, remote monitoring of heart health and can issue accurate alerts when the disease condition is detected with notifying the user and other stakeholders to take relevant clinical actions.

Internet has reached remotest corner of the globe, medical facility is not. We can enable early warning and on-demand automated cardiac care provisioning by leveraging wide-scale deployment of Internet of Things applications acts for developing wireless health monitoring using smartphone and smart ECG sensors like MAX3003 (https://www.maximintegrated.com/en/products/analog/data-converters/analog-front-end-ics/MAX30003.html). It is well-known that early detection and timely intervention can lead to significant life-saving outcomes with substantial reduction of clinical burden. For instance, Myocardial Infarction is to be diagnosed and treated in an urgent manner and an appropriate treatment within first hour can lead to considerable avoidance of deaths and reversal of heart condition. Automated digital screening of cardio-vascular diseases through Internet infrastructure can potentially lead to early detection and in-time screening even at home or at a remote place without real-time access to doctors or cardiologists. Remote screening and monitoring are especially imperative for cardio-vascular disease management. We understand that ShapAAL performs significantly better than the state-of-the-art in cardio-vascular disease detection using ECG signals (for e.g., “ECG200”, “ECGFiveDays”, “TwoLeadECG” datasets results in Table 3). ShapAAL outperforms the current benchmark in Myocardial Infarction detection with test accuracy of 0.92. ShapAAL as part of the analytic engine for automated detection of Myocardial Infraction condition. The primary objective is to build an early warning and on-demand automated cardiac care provisioning that does not get hindered by the immediate absence of a specialist or the user being in a remote place. As a generic setup, the components of the eco-system can be modularized as applications for user end, medical caregiver end and analytics engine end (where the ECG classification model is hosted. In presence of powerful local machine, smartphone, ECG analysis can be done at edge or locally). Users or the patients install the user end application in his/her smartphone (or it can be installed in a laptop) to proactively interact for receiving the cardiac care from the smart healthcare systems with digital therapeutics as part of a typical m-health eco-system. The analytics engine does the job of ECG data interpretation to predict the cardio-vascular disease class. For instance, analytics engine predicts whether the user suffers from Myocardial Infarction condition and sends alerts to the medical caregivers for urgent clinical attention and intervention. The model is trained off-line, and the trained model is deployed on the cloud or at the local workstation as a clinical analytics engine. The on-field ECG data is given as input to the trained model MShapAAL and the output as one of the disease classes (considering binary or multi-class classification) is considered as the screening outcome. We illustrate the system, which can be potentially developed as an early warning platform for basic CVD screening in Fig 17. Further, we like to mention that clinical screening scenario of the conventional CVD screening and diagnosis need to be changed from a reactive mode to proactive mode. In current conventional setup, users will react when the symptoms flareup. In the most likely scenario, the milder symptoms will be ignored when the clinical facility is far-off. Even the routine check-up, which is necessary for CVD patients may be skipped by the remote patients. Another serious consideration is the missing response of subclinical or non-symptomatic condition of CVDs, where the patient might suddenly develop life-threatening conditions. With the proposed automated CVD screening method that can be conveniently performed at home, we expect that the CVD screening will be proactive with early warning of sub-clinical or non-symptomatic CVDs. We are hopeful that the paradigm shift towards automated basic cardio-vascular disease screening can enable us to achieve the goal of 25% relative reduction in premature mortality due to cardio vascular diseases before 2025 [54].

Fig 17. Early warning, emergency, and on-demand cardiac care provision through automated clinical analytics engine with ShapAAL.

Fig 17

We like to mention that the ECG-based automated cardio-vascular disease detection as early warning system is illustrated as an example use case scenario. The proposed method is a generic one and would be an ideal choice for different analytics tasks involving the requirement of time series sensor data classification. Another interesting practical application is in food safety and quality assurance (“Coffee” dataset) to identify the type of coffee beans through food spectrographs.

Conclusion

Our aim of this study is to develop solution for solving the important practical problem of training data scarcity in time series sensor data classification tasks when deploying diverse type of real-world applications including smart cardio-vascular disease detection using ECG data to build effective early-warning, on-demand heart health monitoring eco-system. Our proposed augmented learning with input subset selection approach through Shapley value-based attribution has demonstrated significantly accurate performance over diverse time series sensor data analysis tasks. We have proposed a novel learning mechanism that learns with augmented training to compensate the inadequacy of the training data; unlearns the non-important samples by identifying their contributions to the model predictability through Shapley value computation from coalition game setup with transferable utility; and re-learns with those subset samples. Our novel three-stage time series classification model with learning through augmentation, unlearning the non-contributing input features with Shapley value attribution and finally, relearning through augmentation of selected input features has demonstrated classification efficacy not only through ablation study but also through comparative state-of-the-art investigation. In fact, the intentional introduction of perturbations in the training process of the deep neural network (ResNet) model compels it to learn generalization with crafted and controlled perturbations to create important, unseen input space. The main objective for constructing the learned model when training data is less is to find a way towards minimize the generalization loss over unseen or test or on-filed data. The unique feature of ShapAAL algorithm is the augmentation for learning the unseen data as well as removing the negatively-contributing seen examples in the learning process, which in tandem constitutes superior and effective input space to learn better under training data scarcity problem. Given that Shapley values provide quantitative understanding of fairly attributing the contribution of the input features, the unlearning of detrimental input features has theoretical benefits and we have demonstrated that ablation of such input features has positive impact towards the learnability of the model.

We sincerely hope that the proposed model has the capability to demonstrate practical significance in the development cycle of real-world sensor data classification-based applications including automated prediction of cardio-vascular diseases from physiological marker of heart health like Electrocardiogram to build remote, on-demand smart cardio-vascular health monitoring and early warning system. The proposed method is a generic one for solving time series classification tasks. We envisage that automated analysis with algorithmic screening for cardio-vascular disease identification purpose has the right potential to step towards the long-cherished quest for the availability of a cardio-vascular health management system to intervene for the initial disease screening without expert-in loop.

Our future scope of study includes more exploration towards game theoretic understanding in the construction of a deep learning model with an intuitive rationality perspective of model’s dilemma for prediction over unseen data. The general step for Shapley value computation is using sampling method to estimate the expectation over a distribution of marginals and interpretable machine learning fits to such type of quantified notion of an input feature’s contribution. We intend to explore the model interpretability and algorithmic transparency as a future research initiative with model-agnostic interpretability indicating marginal contributions for individual input features. Another interesting idea is to investigate virtual adversarial regularization such that we can consider the perspective of model robustness. While a sophisticated model provides outstanding performance on given dataset, the model may be over-sensitive towards a little adversarial attack. Data augmentation is in fact capable of improving the stability of the model where the model does not have a high confidence at the prediction, but those augmented examples are close to the given seen examples. From practical utility perspective, we shall further focus on introducing prescriptive analytics such that the initial treatment directive can be urgently delivered as a basic critical care, which can be lifesaving as well as provides the emergency caregivers the information to immediately start the basic yet immensely important initial basic clinical procedures. For example, after heart attack, each passing minutes cause more heart tissues to get damaged. When the analytics engine detects heart attack, immediate commencement of medications like aspirins, thrombolytics before a cardiologist’s intervention is of immense clinical importance. We intend to bring out a robust remote cardio-vascular management system with automation in the basic screening methods that utilizes the Internet backbone to enable healthcare services to the remotest part of the globe for on-demand screening and basic treatment with both screening and prescriptive functions.

Supporting information

S1 Data. Data source.

Experimental datasets are publicly available at https://figshare.com/articles/dataset/Data_zip/21532440/1.

(TXT)

S1 Table. Hyperparameters.

The hyperparameters used in ShapAAL model construction.

(PDF)

S1 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in ChinaTown dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S2 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in Coffee dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S3 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in ECG200 dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S4 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in ECGFiveDays dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S5 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in FreezerRegularTrain dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S6 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in FreezerSmallTrain dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S7 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in ItalyPowerDemandn dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S8 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in MoteStrain dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S9 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in PowerCons dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S10 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in SonyAIBO1 dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S11 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in SonyAIBO2 dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S12 Fig. Study on the training augmentation control.

We depict the trend of the test accuracy data augmentation control parameter α in TwoLeadECG dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

(TIF)

S13 Fig. ShapAAL model plot.

We present the complete model description for reproducibility, where the input is “ECG200” training dataset.

(TIF)

Acknowledgments

Leandro Marin acknowledges the support of PID2020-112675RB-C44 by MCIN/AEI/10.13039/5011000011033 for this research work execution.

Antonio J. Jara, Libelium acknowledges the cooperation for data identification and experimentation in QUAFAIR experiment for the Smart and Healthy Ageing through People Engaging in Supportive Systems: SHAPES—H2020 project (857159) and Comunidad Autonoma de la Region de Murcia (CARM) in HORECOV-21—RIS3MUR FEDER Strengthen research, technological development and innovation.

Data Availability

All relevant data is available from the figshare repository (https://doi.org/10.6084/m9.figshare.21532440.v2).

Funding Statement

The work is partially funded by Grant PID2020-112675RB-C44 funded by MCIN (Ministry for Science and Innovation)/AEI (Agencia Estatal de Investigación - State Research Agency)/10.13039/5011000011033. Tata Consultancy Services is funding the work with generous support and entire APC funding.

References

  • 1. LeCun Y. The Power and Limits of Deep Learning: In his IRI Medal address, Yann LeCun maps the development of machine learning techniques and suggests what the future may hold. Research-Technology Management. 2018;61(6):22–27. doi: 10.1080/08956308.2018.1516928 [DOI] [Google Scholar]
  • 2. Bengio Y, Lecun Y, Hinton G. Deep learning for AI. Communications of the ACM. 2021;64(7):58–65. doi: 10.1145/3448250 [DOI] [Google Scholar]
  • 3.Ukil A, Marin L, Jara A, Farserotu J. Knowledge-driven analytics and systems impacting human quality of life. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management; 2019. p. 2989–2990.
  • 4. Gropler MR, Dalal AS, Van Hare GF, Silva JNA. Can smartphone wireless ECGs be used to accurately assess ECG intervals in pediatrics? A comparison of mobile health monitoring to standard 12-lead ECG. PLoS One. 2018;13(9):e0204403. doi: 10.1371/journal.pone.0204403 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Jafari H, Shohaimi S, Salari N, Kiaei AA, Najafi F, Khazaei S, et al. A full pipeline of diagnosis and prognosis the risk of chronic diseases using deep learning and Shapley values: The Ravansar county anthropometric cohort study. PloS one. 2022;17(1):e0262701. doi: 10.1371/journal.pone.0262701 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Clifford GD, Liu C, Moody B, Li-wei HL, Silva I, Li Q, et al. Af classification from a short single lead ecg recording: the physionet/computing in cardiology challenge 2017. In 2017 Computing in Cardiology (CinC); 2017. [DOI] [PMC free article] [PubMed]
  • 7. Ukil A, Marin L, Mukhopadhyay SC, Jara AJ. AFSense-ECG: Atrial Fibrillation Condition Sensing from Single Lead Electrocardiogram (ECG) Signals. IEEE Sensors Journal. 2022;. doi: 10.1109/JSEN.2022.3162691 [DOI] [Google Scholar]
  • 8. Bagnall A, Lines J, Bostrom A, Large J, Keogh E. The Great Time Series Classification Bake Off: a Review and Experimental Evaluation of Recent Algorithmic Advances. Data Mining and Knowledge Discovery. 2017;31:606–660. doi: 10.1007/s10618-016-0483-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Krizhevsky A, Sutskever I, Hinton GE. ImageNet Classification with Deep Convolutional Neural Networks. In: Pereira F, Burges CJC, Bottou L, Weinberger KQ, editors. Advances in Neural Information Processing Systems 25. Curran Associates, Inc.; 2012. p. 1097–1105.
  • 10.Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L. ImageNet: A Large-Scale Hierarchical Image Database. In: CVPR09; 2009.
  • 11.Krizhvesky A. The Need for Biases in Learning Generalizations. University of Toronto; 2009.
  • 12.Huang C, Wu X, Zhang X, Lin S, Chawla NV. Deep prototypical networks for imbalanced time series classification under data scarcity. In: Proceedings of the 28th ACM International Conference on Information and Knowledge Management; 2019. p. 2141–2144.
  • 13.Mounsaveng S, Vazquez D, Ayed IB, Pedersoli M. Adversarial learning of general transformations for data augmentation. arXiv preprint arXiv:190909801. 2019;.
  • 14.Worzyk N, Yu S. Broad Adversarial Training with Data Augmentation in the Output Space. In: The AAAI-22 Workshop on Adversarial Machine Learning and Beyond; 2021.
  • 15.Goodfellow IJ, Shlens J, Szegedy C. Explaining and harnessing adversarial examples. arXiv preprint arXiv:14126572. 2014;.
  • 16. Ilyas A, Santurkar S, Tsipras D, Engstrom L, Tran B, Madry A. Adversarial examples are not bugs, they are features. Advances in neural information processing systems. 2019;32. [Google Scholar]
  • 17. Iwana BK, Uchida S. An empirical survey of data augmentation for time series classification with neural networks. Plos one. 2021;16(7):e0254841. doi: 10.1371/journal.pone.0254841 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Clarysse J, Hörmann J, Yang F. Why adversarial training can hurt robust accuracy. arXiv preprint arXiv:220302006. 2022;.
  • 19.Roth AE. The Shapley value: essays in honor of Lloyd S. Shapley. Cambridge University Press; 1988.
  • 20.Shapley LS. Notes on the n-Person Game—II: The Value of an n-Person Game. (1951). Lloyd S Shapley. 1951;.
  • 21. Maleki S, Rahwan T, Ghosh S, Malibari A, Alghazzawi D, Rogers A, et al. The Shapley value for a fair division of group discounts for coordinating cooling loads. PloS one. 2020;15(1):e0227049. doi: 10.1371/journal.pone.0227049 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Wang Q, Farahat A, Gupta C, Zheng S. Deep time series models for scarce data. Neurocomputing. 2021;456:504–518. doi: 10.1016/j.neucom.2020.12.132 [DOI] [Google Scholar]
  • 23. Lines J, Bagnall A. Time series classification with ensembles of elastic distance measures. Data Mining and Knowledge Discovery. 2015;29(3):565–592. doi: 10.1007/s10618-014-0361-2 [DOI] [Google Scholar]
  • 24. Bagnall A, Lines J, Hills J, Bostrom A. Time-series classification with COTE: the collective of transformation-based ensembles. IEEE Transactions on Knowledge and Data Engineering. 2015;27(9):2522–2535. doi: 10.1109/TKDE.2015.2416723 [DOI] [Google Scholar]
  • 25.Flynn M, Large J, Bagnall T. The contract random interval spectral ensemble (c-RISE): the effect of contracting a classifier on accuracy. In: International Conference on Hybrid Artificial Intelligence Systems. Springer; 2019. p. 381–392.
  • 26. Shifaz A, Pelletier C, Petitjean F, Webb GI. TS-CHIEF: a scalable and accurate forest algorithm for time series classification. Data Mining and Knowledge Discovery. 2020;34(3):742–775. doi: 10.1007/s10618-020-00679-8 [DOI] [Google Scholar]
  • 27. Deng H, Runger G, Tuv E, Vladimir M. A time series forest for classification and feature extraction. Information Sciences. 2013;239:142–153. doi: 10.1016/j.ins.2013.02.030 [DOI] [Google Scholar]
  • 28. Lucas B, Shifaz A, Pelletier C, O’Neill L, Zaidi N, Goethals B, et al. Proximity forest: an effective and scalable distance-based classifier for time series. Data Mining and Knowledge Discovery. 2019;33(3):607–635. doi: 10.1007/s10618-019-00617-3 [DOI] [Google Scholar]
  • 29. Lubba CH, Sethi SS, Knaute P, Schultz SR, Fulcher BD, Jones NS. catch22: Canonical time-series characteristics. Data Mining and Knowledge Discovery. 2019;33(6):1821–1852. doi: 10.1007/s10618-019-00647-x [DOI] [Google Scholar]
  • 30.He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2016. p. 770–778.
  • 31.Wang Z, Yan W, Oates T. Time series classification from scratch with deep neural networks: A strong baseline. In: 2017 International joint conference on neural networks (IJCNN). IEEE; 2017. p. 1578–1585.
  • 32. Mahajan S, Pandit AK. Hybrid method to supervise feature selection using signal processing and complex algebra techniques. Multimedia Tools and Applications. 2021; p. 1–22. [Google Scholar]
  • 33. Mahajan S, Abualigah L, Pandit AK, Altalhi M. Hybrid Aquila optimizer with arithmetic optimization algorithm for global optimization tasks. Soft Computing. 2022;26(10):4863–4881. doi: 10.1007/s00500-022-06873-8 [DOI] [Google Scholar]
  • 34. Mahajan S, Pandit AK. Image segmentation and optimization techniques: a short overview. Medicon Eng Themes. 2022;2(2):47–49. [Google Scholar]
  • 35. Mahajan S, Abualigah L, Pandit AK. Hybrid arithmetic optimization algorithm with hunger games search for global optimization. Multimedia Tools and Applications. 2022; p. 1–24. [Google Scholar]
  • 36. Mahajan S, Abualigah L, Pandit AK, Nasar A, Rustom M, Alkhazaleh HA, et al. Fusion of modern meta-heuristic optimization methods using arithmetic optimization algorithm for global optimization tasks. Soft Computing. 2022; p. 1–15. [Google Scholar]
  • 37. Lakshmi YV, Singh P, Abouhawwash M, Mahajan S, Pandit AK, Ahmed AB. Improved Chan Algorithm Based Optimum UWB Sensor Node Localization Using Hybrid Particle Swarm Optimization. IEEE Access. 2022;10:32546–32565. doi: 10.1109/ACCESS.2022.3157719 [DOI] [Google Scholar]
  • 38. Salgotra R, Abouhawwash M, Singh U, Saha S, Mittal N, Mahajan S, et al. Multi-population and dynamic-iterative cuckoo search algorithm for linear antenna array synthesis. Applied Soft Computing. 2021;113:108004. doi: 10.1016/j.asoc.2021.108004 [DOI] [Google Scholar]
  • 39. Singh H, Abouhawwash M, Mittal N, Salgotra R, Mahajan S, Pandit AK. Performance evaluation of Non-Uniform circular antenna array using integrated harmony search with Differential Evolution based Naked Mole Rat algorithm. Expert Systems with Applications. 2022;189:116146. doi: 10.1016/j.eswa.2021.116146 [DOI] [Google Scholar]
  • 40.Bansal MA, Sharma DR, Kathuria DM. A Systematic Review on Data Scarcity Problem in Deep Learning: Solution and Applications. ACM Computing Surveys (CSUR). 2020;.
  • 41. Zaeemzadeh A, Rahnavard N, Shah M. Norm-preservation: Why residual networks can become extremely deep? IEEE transactions on pattern analysis and machine intelligence. 2020;43(11):3980–3990. doi: 10.1109/TPAMI.2020.2990339 [DOI] [PubMed] [Google Scholar]
  • 42.Ukil A, Jara AJ, Marin L. Blend-Res 2 net: Blended Representation Space by Transformation of Residual Mapping with Restrained Learning for Time Series Classification. In: ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE; 2021. p. 3555–3559.
  • 43. Liu T, Chen M, Zhou M, Du SS, Zhou E, Zhao T. Towards understanding the importance of shortcut connections in residual networks. Advances in neural information processing systems. 2019;32. [Google Scholar]
  • 44.Tsipras D, Santurkar S, Engstrom L, Turner A, Madry A. Robustness may be at odds with accuracy. arXiv preprint arXiv:180512152. 2018;.
  • 45.Rozemberczki B, Watson L, Bayer P, Yang HT, Kiss O, Nilsson S, et al. The Shapley Value in Machine Learning. arXiv preprint arXiv:220205594. 2022;.
  • 46. Fryer D, Strümke I, Nguyen H. Shapley values for feature selection: the good, the bad, and the axioms. IEEE Access. 2021;9:144352–144360. doi: 10.1109/ACCESS.2021.3119110 [DOI] [Google Scholar]
  • 47.Roth AE. The shapley value; 2005.
  • 48.Shrikumar A, Greenside P, Kundaje A. Learning important features through propagating activation differences. In: International conference on machine learning. PMLR; 2017. p. 3145–3153.
  • 49.Gay D, Lemaire V. Should we Reload Time Series Classification Performance Evaluation?(a position paper). arXiv preprint arXiv:190303300. 2019;.
  • 50.Summers C, Dinneen MJ. Nondeterminism and instability in neural network optimization. In: International Conference on Machine Learning. PMLR; 2021. p. 9913–9922.
  • 51. Zhuang D, Zhang X, Song S, Hooker S. Randomness in neural network training: Characterizing the impact of tooling. Proceedings of Machine Learning and Systems. 2022;4:316–336. [Google Scholar]
  • 52.Sorscher B, Geirhos R, Shekhar S, Ganguli S, Morcos AS. Beyond neural scaling laws: beating power law scaling via data pruning. arXiv preprint arXiv:220614486. 2022;.
  • 53. Roth GA, Mensah GA, Johnson CO, Addolorato G, Ammirati E, Baddour LM, et al. Global burden of cardiovascular diseases and risk factors, 1990–2019: update from the GBD 2019 study. Journal of the American College of Cardiology. 2020;76(25):2982–3021. doi: 10.1016/j.jacc.2020.11.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Ordunez P, Prieto-Lara E, Pinheiro Gawryszewski V, Hennis AJ, Cooper RS. Premature Mortality from Cardiovascular Disease in the Americas–Will the Goal of a Decline of “25% by 2025” be Met? PloS one. 2015;10(10):e0141685. doi: 10.1371/journal.pone.0141685 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Anand Nayyar

14 Oct 2022

PONE-D-22-26084When less is more powerful: Shapley value attributed ablation with augmented learning for practical time series sensor data classificationPLOS ONE

Dear Dr. Ukil,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Nov 28 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Anand Nayyar, Ph.D.

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1.  Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf  and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse. New software must comply with the Open Source Definition.

3. Thank you for stating the following in the Acknowledgments Section of your manuscript:

“Leandro Marin is partially funded by Grant PID2020-112675RB-C44 funded by MCIN 542

(Ministry for Science and Innovation)/AEI (Agencia Estatal de Investigaci´on - State 543

Research Agency)/10.13039/5011000011033.”

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

 “The work is partially funded by Grant PID2020-112675RB-C44 funded by MCIN (Ministry for Science and Innovation)/AEI (Agencia Estatal de Investigación - State Research Agency)/10.13039/5011000011033. Tata Consultancy Services is funding the work with generous support and entire APC funding.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

Additional Editor Comments :

The Paper needs revisions and is subject for re-review.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors take up a very important topic, which is time series sensor data classification. Autrozy solidly designed the manuscript. It is very interesting to describe the abstract and explain what the novelty is at work. The authors explain the topic in an interesting way. The strength of the work is a solid literature review, also based on the latest studies and methodology with interesting diagrams and drawings. The weak point is the lack of underlining that the hatch fills the work with. I recommend that you complete this point.

Reviewer #2: Paper Title: When less is more powerful: Shapley value attributed ablation with augmented learning for practical time series sensor data classification

Discusses: Time series sensor data classification tasks often suffer from training data scarcity issue due to the expenses associated with the expert-intervened annotation efforts. For example, Electrocardiogram (ECG) data classification for cardio-vascular disease detection requires expensive labeling procedures with the help of cardiologists. The current state-of-the-art algorithms like deep learning models have shown outstanding performance under the general requirement of availability of large set of training examples. In this paper, we propose Shapley Attributed Ablation with Augmented Learning: ShapAAL, which demonstrates that deep learning algorithm with suitably selected subset of the seen examples or ablating the unimportant ones from the given limited training dataset can ensure consistently better classification performance under augmented training. In ShapAAL, additive perturbed training augments the input space to compensate the scarcity in training examples and Shapley attribution seeks the subset from the augmented training space for better learnability with the goal of better general predictive performance, thanks to the ”efficiency” and ”null player” axioms of transferable utility games upon which Shapley value game is formulated. In ShapAAL, the subset of training examples that contribute positively in a supervised learning setup is derived from the notion of coalition games using Shapley values associated with each of the given examples’ contribution into the model prediction. ShapAAL is a novel push-pull deep architecture where the subset selection through Shapley value attribution pushes the model to lower dimension while augmented training augments the learning capability of the model over unseen data. We perform ablation study to provide the empirical evidence of our claim and we show that proposed ShapAAL method outperforms the current baselines and state-of-the-art results for time series sensor data classification tasks including the practical important ones that detect cardio-vascular diseases from ECG data.

1.Abstract and Conclusion should be concise yet. But should give complete overview of the work and study.

2.Authors can use latest related works from reputed journals like IEEE/ACM Transactions, MDPI, Elsevier, Inderscience, Springer, Taylor & Francis etc and write the references in proper format, from year 2021-2022. Like https://link.springer.com/article/10.1007/s11042-021-11474-y, https://link.springer.com/article/10.1007/s00500-022-06873-8, https://themedicon.com/pdf/engineeringthemes/MCET-02-016.pdf, https://link.springer.com/article/10.1007/s00500-022-07079-8, https://link.springer.com/article/10.1007/s11042-022-12922-z,

https://ieeexplore.ieee.org/abstract/document/9729866/, https://www.sciencedirect.com/science/article/abs/pii/S095741742101472X, https://www.sciencedirect.com/science/article/abs/pii/S1568494621009261

3.The authors seem to disregard or neglect some important finding in results that have been achieved in paper. So, elaborate and explain the results in more details.

4.Improve the results and discussion section in paragraph.

5.Mention the future scope of your present works.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: SHUBHAM MAHAJAN

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Nov 23;17(11):e0277975. doi: 10.1371/journal.pone.0277975.r002

Author response to Decision Letter 0


26 Oct 2022

Review comments and Response:

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors take up a very important topic, which is time series sensor data classification. Autrozy solidly designed the manuscript. It is very interesting to describe the abstract and explain what the novelty is at work. The authors explain the topic in an interesting way. The strength of the work is a solid literature review, also based on the latest studies and methodology with interesting diagrams and drawings. The weak point is the lack of underlining that the hatch fills the work with. I recommend that you complete this point.

Thanks for the review comment. We have modified the manuscript with careful investigation. The Abstract and Conclusion Sections are modified considerably.

Abstract:

Time series sensor data classification tasks often suffer from training data scarcity issue due to the expenses associated with the expert-intervened annotation efforts. For example, Electrocardiogram (ECG) data classification for cardio-vascular disease (CVD) detection requires expensive labeling procedures with the help of cardiologists. Current state-of-the-art algorithms like deep learning models have shown outstanding performance under the general requirement of availability of large set of training examples. In this paper, we propose Shapley Attributed Ablation with Augmented Learning: ShapAAL, which demonstrates that deep learning algorithm with suitably selected subset of the seen examples or ablating the unimportant ones from the given limited training dataset can ensure consistently better classification performance under augmented training. In ShapAAL, additive perturbed training augments the input space to compensate the scarcity in training examples using Residual Network (ResNet) architecture through perturbation-induced inputs, while Shapley attribution seeks the subset from the augmented training space for better learnability with the goal of better general predictive performance, thanks to the "efficiency" and "null player" axioms of transferable utility games upon which Shapley value game is formulated. In ShapAAL, the subset of training examples that contribute positively to a supervised learning setup is derived from the notion of coalition games using Shapley values associated with each of the given inputs' contribution into the model prediction. ShapAAL is a novel push-pull deep architecture where the subset selection through Shapley value attribution pushes the model to lower dimension while augmented training augments the learning capability of the model over unseen data. We perform ablation study to provide the empirical evidence of our claim and we show that proposed ShapAAL method consistently outperforms the current baselines and state-of-the-art algorithms for time series sensor data classification tasks from publicly available UCR time series archive that includes different practical important problems like detection of CVDs from ECG data.

Conclusion:

Our aim of this study is to develop solution for solving the important practical problem of training data scarcity in time series sensor data classification tasks when deploying diverse type of real-world applications including smart cardio-vascular disease detection using ECG data to build effective early-warning, on-demand heart health monitoring eco-system. Our proposed augmented learning with input subset selection approach through Shapley value-based attribution has demonstrated significantly accurate performance over diverse time series sensor data analysis tasks. We have proposed a novel learning mechanism that learns with augmented training to compensate the inadequacy of the training data; unlearns the non-important samples by identifying their contributions to the model predictability through Shapley value computation from coalition game setup with transferable utility; and re-learns with those subset samples. Our novel three-stage time series classification model with learning through augmentation, unlearning the non-contributing input features with Shapley value attribution and finally, relearning through augmentation of selected input features has demonstrated classification efficacy not only through ablation study but also through comparative state-of-the-art investigation. In fact, the intentional introduction of perturbations in the training process of the deep neural network (ResNet) model compels it to learn generalization with crafted and controlled perturbations to create important, unseen input space. The main objective for constructing the learned model when training data is less is to find a way towards minimize the generalization loss over unseen or test or on-filed data. The unique feature of ShapAAL algorithm is the augmentation for learning the unseen data as well as removing the negatively-contributing seen examples in the learning process, which in tandem constitutes superior and effective input space to learn better under training data scarcity problem. Given that Shapley values provide quantitative understanding of fairly attributing the contribution of the input features, the unlearning of detrimental input features has theoretical benefits and we have demonstrated that ablation of such input features has positive impact towards the learnability of the model.

We sincerely hope that the proposed model has the capability to demonstrate practical significance in the development cycle of real-world sensor data classification-based applications including automated prediction of cardio-vascular diseases from physiological marker of heart health like Electrocardiogram to build remote, on-demand smart cardio-vascular health monitoring and early warning system. The proposed method is a generic one for solving time series classification tasks. We envisage that automated analysis with algorithmic screening for cardio-vascular disease identification purpose has the right potential to step towards the long-cherished quest for the availability of a cardio-vascular health management system to intervene for the initial disease screening without expert-in loop.

Our future scope of study includes more exploration towards game theoretic understanding in the construction of a deep learning model with an intuitive rationality perspective of model's dilemma for prediction over unseen data. The general step for Shapley value computation is using sampling method to estimate the expectation over a distribution of marginals and interpretable machine learning fits to such type of quantified notion of an input feature's contribution. We intend to explore the model interpretability and algorithmic transparency as a future research initiative with model-agnostic interpretability indicating marginal contributions for individual input features. Another interesting idea is to investigate virtual adversarial regularization such that we can consider the perspective of model robustness. While a sophisticated model provides outstanding performance on given dataset, the model may be over-sensitive towards a little adversarial attack. Data augmentation is in fact capable of improving the stability of the model where the model does not have a high confidence at the prediction, but those augmented examples are close to the given seen examples. From practical utility perspective, we shall further focus on introducing prescriptive analytics such that the initial treatment directive can be urgently delivered as a basic critical care, which can be lifesaving as well as provides the emergency caregivers the information to immediately start the basic yet immensely important initial basic clinical procedures. For example, after heart attack, each passing minutes cause more heart tissues to get damaged. When the analytics engine detects heart attack, immediate commencement of medications like aspirins, thrombolytics before a cardiologist’s intervention is of immense clinical importance. We intend to bring out a robust remote cardio-vascular management system with automation in the basic screening methods that utilizes the Internet backbone to enable healthcare services to the remotest part of the globe for on-demand screening and basic treatment with both screening and prescriptive functions.

We have added some additional results to further substantiate our claim.

Next, we conduct ablation study to understand the efficacy of the proposed model. An ablation study in general, investigates the performance of a machine learning system by removing few components in order to evaluate the impact of those components in the complete system. Similarly, ShapAAL model construction consists of four components that include the base model (ResNet), Shapley value attribution over the base model, data augmented training on the base model and data augmented training with Shapley attributed feature selection on the base model. We denote $M$ as the base model that is trained with each of the training data, $M^{Shapley}$ as the model that is trained with the training data after discarding the negatively contributing Shapley valued features, $M_{aug}$ is the model that is adversarially trained over over entire augmented training data. $M^{ShapAAL}$ or $M_{aug}^{Shapley}$ is the adversarially trained with the augmented training data with discarding the negatively contributing Shapley valued ones following the deep architecture in Figure ~\\ref{fig862} as depicted in Figure ~\\ref{fig1112}. In Table ~\\ref{table_4}, we depict the "test accuracy" performances of $M$, $M_{aug}$, $M^{Shapley}$ and $M_{aug}^{ShapAAL}$ over the experimental datasets. The ablation study unambiguously indicates that our proposed model $M^{ShapAAL}$ is the superior one. In fact, the trend is also clear that both augmented training and Shapley attributed re-learning have significant positive impact on the learnability of the model, which reflects in the consistent superlative performance of $M^{ShapAAL}$ w.r.t the others. Hence, we establish with the empirical support that less number of input features (Refer Figure ~\\ref{fig2345}) when properly selected can provide better test accuracy. Under training data size constraint scenario, the push-pull architecture of ShapAAL as a coalition game with Shapley attributed push towards lower dimension and concurrently pulling or augmenting the learning capability of the model over unseen data indeed demonstrates significantly improved performance.

Another classical performance merit is the "outperforming" the benchmark. In recent years, number of time series classification algorithms have been proposed in literature, which might not have been updated in the UCR archive repository

\\footnote {available at each of the dataset description URL, for e.g., \\url{https://www.timeseriesclassification.com/description.php?Dataset=ECG200}}. However, we can consider the available benchmark or the best results in the UCR repository of the respective datasets \\footnote{\\url{https://www.timeseriesclassification.com/dataset.php}} as the "reported benchmark". In Figure ~\\ref{fig1d}, we depict the differential test accuracy gain of the algorithms (which has reported results available in public domain) including ShapAAL model w.r.t the reported best results and it is computed as $\\frac{test\\; accuracy\\; of\\; the \\;algorithm\\;-\\; reported\\; benchmark \\;test \\;accuracy}{reported\\; benchmark\\; test\\; accuracy}$ with the aim of being the test accuracy result to be positive, indicating that the concerned algorithm has outperformed the currently reported benchmark result. We observe that proposed ShapAAL steadily outperforms the reported benchmark results in comparison with the relevant benchmark algorithms.

In this manuscript, it is felt that MPCE-score-based classification performance evaluation needs to be part of the main content, which was earlier in the Appendix.

Mean Per-Class Error (MPCE) (~\\cite{wang2017time}) is another useful metric to evaluate the classification performance of the model as: the expected error rate for a single class across each of the test data. For $\\Upsilon$ number of test data with class $\\textit{c}_{\\upsilon}$ and corresponding error rate $err_{\\upsilon}$, we compute MPCE as: $\\frac{1}{\\Upsilon}\\sum \\frac{err_{\\upsilon}}{\\textit{c}_{\\upsilon}}$.\\\\

MPCE seems to a more robust as an evaluator of model performance for different datasets of the classes ~\\cite{wang2017time}). Below in Table ~\\ref{tab_76540}, we demonstrate the MPCE results for the ablation study. In MPCE, our aim is to have a lower value, approaching zero.

Another unique feature of the current work is its response to higher number of test instances when it gets trained with smaller number of training examples. We can quantify the learning gain of ShapAAL at the time of testing as: $\\frac{test\\: accurcay_{ShapAAL} \\; - test\\: accurcay_{Base}}{test\\: accurcay_{Base}}$ and also define training insufficiency factor as: $\\frac{Number\\: of\\: training\\: examples}{Number\\: of \\:testing\\: instances}$. In Fig ~\\ref{fig561}, we demonstrate the comparative study of learning gain of ShapAAL on testing data over base model and the insufficiency in the training. We observe that the learning gain of ShapAAL is mostly $\\ge 1$, while training insufficiency factor $\\le 1$. Hence, we further establish our claim that ShapAAL model is the apt choice under practical constraint of training data limitation in solving the time series classification tasks.

We have provided more details in the Discussion Section.

Firstly, we have proposed and validated the unique idea augmentation and ablation of the input features to generate a better learned model. Controlled augmentation of the seen examples to learn better on the unseen examples through introduction of perturbed or virtual data points helps the model to combat the insufficiency in training examples and Shapley-attributed input feature selection refines the input space such that the model gets the opportunity of training more (through augmentation) yet better (Shapley-value based feature ablation). While the augmentation and feature attribution separately improve the test accuracy of the model over different tasks, the combined effect is significant, and it is evident from Table 2 and 4. The study in Table 2 4 clearly indicates that data augmentation through adversarial learning and subsequent feature space identification for re-learning with appropriate features provide significant impetus to the learning process to learn that compensates the limitation in seen examples and learn appropriately. Secondly, we have provided state-of-the-art comparison of the proposed method and the ShapAAL model with both data augmentation and input attribution features has demonstrated consistently outstanding classification performances over different time series classification tasks, conveniently outperforming the current benchmark and state-of-the-art algorithms as depicted in Table 3, Figure 12, and Table 4.

The model is trained off-line, and the trained model is deployed on the cloud or at the local workstation as a clinical analytics engine. The on-field ECG data is given as input to the trained model ShapAAL and the output as one of the disease classes (considering binary or multi-class classification) is considered as the screening outcome. We illustrate the system, which can be potentially developed as an early warning platform for basic CVD screening in Figure 14. Further, we like to mention that clinical screening scenario of the conventional CVD screening and diagnosis need to be changed from a reactive mode to proactive mode. In current conventional setup, users will react when the symptoms flareup. In the most likely scenario, the milder symptoms will be ignored when the clinical facility is far-off. Even the routine check-up, which is necessary for CVD patients may be skipped by the remote patients. Another serious consideration is the missing response of subclinical or non-symptomatic condition of CVDs, where the patient might suddenly develop life-threatening conditions. With the proposed automated CVD screening method that can be conveniently performed at home, we expect that the CVD screening will be proactive with early warning of subclinical or non-symptomatic CVDs.

Reviewer #2: Paper Title: When less is more powerful: Shapley value attributed ablation with augmented learning for practical time series sensor data classification

Discusses: Time series sensor data classification tasks often suffer from training data scarcity issue due to the expenses associated with the expert-intervened annotation efforts. For example, Electrocardiogram (ECG) data classification for cardio-vascular disease detection requires expensive labeling procedures with the help of cardiologists. The current state-of-the-art algorithms like deep learning models have shown outstanding performance under the general requirement of availability of large set of training examples. In this paper, we propose Shapley Attributed Ablation with Augmented Learning: ShapAAL, which demonstrates that deep learning algorithm with suitably selected subset of the seen examples or ablating the unimportant ones from the given limited training dataset can ensure consistently better classification performance under augmented training. In ShapAAL, additive perturbed training augments the input space to compensate the scarcity in training examples and Shapley attribution seeks the subset from the augmented training space for better learnability with the goal of better general predictive performance, thanks to the ”efficiency” and ”null player” axioms of transferable utility games upon which Shapley value game is formulated. In ShapAAL, the subset of training examples that contribute positively in a supervised learning setup is derived from the notion of coalition games using Shapley values associated with each of the given examples’ contribution into the model prediction. ShapAAL is a novel push-pull deep architecture where the subset selection through Shapley value attribution pushes the model to lower dimension while augmented training augments the learning capability of the model over unseen data. We perform ablation study to provide the empirical evidence of our claim and we show that proposed ShapAAL method outperforms the current baselines and state-of-the-art results for time series sensor data classification tasks including the practical important ones that detect cardio-vascular diseases from ECG data.

1.Abstract and Conclusion should be concise yet. But should give complete overview of the work and study.

We have modified the Abstract and Conclusion to provide the overview of the work and study.

Abstract:

Time series sensor data classification tasks often suffer from training data scarcity issue due to the expenses associated with the expert-intervened annotation efforts. For example, Electrocardiogram (ECG) data classification for cardio-vascular disease (CVD) detection requires expensive labeling procedures with the help of cardiologists. Current state-of-the-art algorithms like deep learning models have shown outstanding performance under the general requirement of availability of large set of training examples. In this paper, we propose Shapley Attributed Ablation with Augmented Learning: ShapAAL, which demonstrates that deep learning algorithm with suitably selected subset of the seen examples or ablating the unimportant ones from the given limited training dataset can ensure consistently better classification performance under augmented training. In ShapAAL, additive perturbed training augments the input space to compensate the scarcity in training examples using Residual Network (ResNet) architecture through perturbation-induced inputs, while Shapley attribution seeks the subset from the augmented training space for better learnability with the goal of better general predictive performance, thanks to the "efficiency" and "null player" axioms of transferable utility games upon which Shapley value game is formulated. In ShapAAL, the subset of training examples that contribute positively to a supervised learning setup is derived from the notion of coalition games using Shapley values associated with each of the given inputs' contribution into the model prediction. ShapAAL is a novel push-pull deep architecture where the subset selection through Shapley value attribution pushes the model to lower dimension while augmented training augments the learning capability of the model over unseen data. We perform ablation study to provide the empirical evidence of our claim and we show that proposed ShapAAL method consistently outperforms the current baselines and state-of-the-art algorithms for time series sensor data classification tasks from publicly available UCR time series archive that includes different practical important problems like detection of CVDs from ECG data.

Conclusion:

Our aim of this study is to develop solution for solving the important practical problem of training data scarcity in time series sensor data classification tasks when deploying diverse type of real-world applications including smart cardio-vascular disease detection using ECG data to build effective early-warning, on-demand heart health monitoring eco-system. Our proposed augmented learning with input subset selection approach through Shapley value-based attribution has demonstrated significantly accurate performance over diverse time series sensor data analysis tasks. We have proposed a novel learning mechanism that learns with augmented training to compensate the inadequacy of the training data; unlearns the non-important samples by identifying their contributions to the model predictability through Shapley value computation from coalition game setup with transferable utility; and re-learns with those subset samples. Our novel three-stage time series classification model with learning through augmentation, unlearning the non-contributing input features with Shapley value attribution and finally, relearning through augmentation of selected input features has demonstrated classification efficacy not only through ablation study but also through comparative state-of-the-art investigation. In fact, the intentional introduction of perturbations in the training process of the deep neural network (ResNet) model compels it to learn generalization with crafted and controlled perturbations to create important, unseen input space. The main objective for constructing the learned model when training data is less is to find a way towards minimize the generalization loss over unseen or test or on-filed data. The unique feature of ShapAAL algorithm is the augmentation for learning the unseen data as well as removing the negatively-contributing seen examples in the learning process, which in tandem constitutes superior and effective input space to learn better under training data scarcity problem. Given that Shapley values provide quantitative understanding of fairly attributing the contribution of the input features, the unlearning of detrimental input features has theoretical benefits and we have demonstrated that ablation of such input features has positive impact towards the learnability of the model.

We sincerely hope that the proposed model has the capability to demonstrate practical significance in the development cycle of real-world sensor data classification-based applications including automated prediction of cardio-vascular diseases from physiological marker of heart health like Electrocardiogram to build remote, on-demand smart cardio-vascular health monitoring and early warning system. The proposed method is a generic one for solving time series classification tasks. We envisage that automated analysis with algorithmic screening for cardio-vascular disease identification purpose has the right potential to step towards the long-cherished quest for the availability of a cardio-vascular health management system to intervene for the initial disease screening without expert-in loop.

Our future scope of study includes more exploration towards game theoretic understanding in the construction of a deep learning model with an intuitive rationality perspective of model's dilemma for prediction over unseen data. The general step for Shapley value computation is using sampling method to estimate the expectation over a distribution of marginals and interpretable machine learning fits to such type of quantified notion of an input feature's contribution. We intend to explore the model interpretability and algorithmic transparency as a future research initiative with model-agnostic interpretability indicating marginal contributions for individual input features. Another interesting idea is to investigate virtual adversarial regularization such that we can consider the perspective of model robustness. While a sophisticated model provides outstanding performance on given dataset, the model may be over-sensitive towards a little adversarial attack. Data augmentation is in fact capable of improving the stability of the model where the model does not have a high confidence at the prediction, but those augmented examples are close to the given seen examples. From practical utility perspective, we shall further focus on introducing prescriptive analytics such that the initial treatment directive can be urgently delivered as a basic critical care, which can be lifesaving as well as provides the emergency caregivers the information to immediately start the basic yet immensely important initial basic clinical procedures. For example, after heart attack, each passing minutes cause more heart tissues to get damaged. When the analytics engine detects heart attack, immediate commencement of medications like aspirins, thrombolytics before a cardiologist’s intervention is of immense clinical importance. We intend to bring out a robust remote cardio-vascular management system with automation in the basic screening methods that utilizes the Internet backbone to enable healthcare services to the remotest part of the globe for on-demand screening and basic treatment with both screening and prescriptive functions.

2.Authors can use latest related works from reputed journals like IEEE/ACM Transactions, MDPI, Elsevier, Inderscience, Springer, Taylor & Francis etc and write the references in proper format, from year 2021-2022. Like https://link.springer.com/article/10.1007/s11042-021-11474-y, https://link.springer.com/article/10.1007/s00500-022-06873-8, https://themedicon.com/pdf/engineeringthemes/MCET-02-016.pdf, https://link.springer.com/article/10.1007/s00500-022-07079-8, https://link.springer.com/article/10.1007/s11042-022-12922-z,

https://ieeexplore.ieee.org/abstract/document/9729866/, https://www.sciencedirect.com/science/article/abs/pii/S095741742101472X, https://www.sciencedirect.com/science/article/abs/pii/S1568494621009261

Thanks for the advice. We have incorporated the latest related works.

In general, machine learning algorithms need to carefully select the supervised 125 features to build a robust model [32]. Optimization method plays an important role in 126 various aspects towards better learned model development under practical constraints 127 [33], [34], [35], [36], [37], [38], [39]. For instance, evolutionary processes with 128 consistent equilibrium for high-quality performance and optimization that achieves 129 quicker convergence is proposed in [35]. It is well-known that the search for global 130 optimization in deep learning algorithms often suffer through spurious local 131 optimization issues. In [36], fusion-based meta-heuristic optimization methods are 132 proposed to solve global optimization tasks.

Added References:

32. Mahajan S, Pandit AK. Hybrid method to supervise feature selection using signal processing and complex algebra techniques. Multimedia Tools and Applications. 2021; p. 1–22.

33. Mahajan S, Abualigah L, Pandit AK, Altalhi M. Hybrid Aquila optimizer with arithmetic optimization algorithm for global optimization tasks. Soft Computing. 2022;26(10):4863–4881.

34. Mahajan S, Pandit AK. Image segmentation and optimization techniques: a short overview. Medicon Eng Themes. 2022;2(2):47–49.

35. Mahajan S, Abualigah L, Pandit AK. Hybrid arithmetic optimization algorithm with hunger games search for global optimization. Multimedia Tools and Applications. 2022; p. 1–24.

36. Mahajan S, Abualigah L, Pandit AK, Nasar A, Rustom M, Alkhazaleh HA, et al. Fusion of modern meta-heuristic optimization methods using arithmetic optimization algorithm for global optimization tasks. Soft Computing. 2022; p. 1–15.

37. Lakshmi YV, Singh P, Abouhawwash M, Mahajan S, Pandit AK, Ahmed AB. Improved Chan Algorithm Based Optimum UWB Sensor Node Localization Using Hybrid Particle Swarm Optimization. IEEE Access. 2022;10:32546–32565.

38. Salgotra R, Abouhawwash M, Singh U, Saha S, Mittal N, Mahajan S, et al. Multi-population and dynamic-iterative cuckoo search algorithm for linear antenna array synthesis. Applied Soft Computing. 2021;113:108004.

39. Singh H, Abouhawwash M, Mittal N, Salgotra R, Mahajan S, Pandit AK. Performance evaluation of Non-Uniform circular antenna array using integrated harmony search with Differential Evolution based Naked Mole Rat algorithm. Expert Systems with Applications. 2022;189:116146.

3.The authors seem to disregard or neglect some important finding in results that have been achieved in paper. So, elaborate and explain the results in more details.

We have elaborated the results to detail out and establish the efficacy of the proposed method. We have made number of additions and modifications in the revised manuscript as depicted below.

{Modification/addition in the Result Section}

Next, we conduct ablation study to understand the efficacy of the proposed model. An ablation study in general, investigates the performance of a machine learning system by removing few components in order to evaluate the impact of those components in the complete system. Similarly, ShapAAL model construction consists of four components that include the base model (ResNet), Shapley value attribution over the base model, data augmented training on the base model and data augmented training with Shapley attributed feature selection on the base model. We denote $M$ as the base model that is trained with each of the training data, $M^{Shapley}$ as the model that is trained with the training data after discarding the negatively contributing Shapley valued features, $M_{aug}$ is the model that is adversarially trained over over entire augmented training data. $M^{ShapAAL}$ or $M_{aug}^{Shapley}$ is the adversarially trained with the augmented training data with discarding the negatively contributing Shapley valued ones following the deep architecture in Figure ~\\ref{fig862} as depicted in Figure ~\\ref{fig1112}. In Table ~\\ref{table_4}, we depict the "test accuracy" performances of $M$, $M_{aug}$, $M^{Shapley}$ and $M_{aug}^{ShapAAL}$ over the experimental datasets. The ablation study unambiguously indicates that our proposed model $M^{ShapAAL}$ is the superior one. In fact, the trend is also clear that both augmented training and Shapley attributed re-learning have significant positive impact on the learnability of the model, which reflects in the consistent superlative performance of $M^{ShapAAL}$ w.r.t the others. Hence, we establish with the empirical support that less number of input features (Refer Figure ~\\ref{fig2345}) when properly selected can provide better test accuracy. Under training data size constraint scenario, the push-pull architecture of ShapAAL as a coalition game with Shapley attributed push towards lower dimension and concurrently pulling or augmenting the learning capability of the model over unseen data indeed demonstrates significantly improved performance.

Another classical performance merit is the "outperforming" the benchmark. In recent years, number of time series classification algorithms have been proposed in literature, which might not have been updated in the UCR archive repository

\\footnote {available at each of the dataset description URL, for e.g., \\url{https://www.timeseriesclassification.com/description.php?Dataset=ECG200}}. However, we can consider the available benchmark or the best results in the UCR repository of the respective datasets \\footnote{\\url{https://www.timeseriesclassification.com/dataset.php}} as the "reported benchmark". In Figure ~\\ref{fig1d}, we depict the differential test accuracy gain of the algorithms (which has reported results available in public domain) including ShapAAL model w.r.t the reported best results and it is computed as $\\frac{test\\; accuracy\\; of\\; the \\;algorithm\\;-\\; reported\\; benchmark \\;test \\;accuracy}{reported\\; benchmark\\; test\\; accuracy}$ with the aim of being the test accuracy result to be positive, indicating that the concerned algorithm has outperformed the currently reported benchmark result. We observe that proposed ShapAAL steadily outperforms the reported benchmark results in comparison with the relevant benchmark algorithms.

In this manuscript, it is felt that MPCE-score-based classification performance evaluation needs to be part of the main content, which was earlier in the Appendix.

Mean Per-Class Error (MPCE) (~\\cite{wang2017time}) is another useful metric to evaluate the classification performance of the model as: the expected error rate for a single class across each of the test data. For $\\Upsilon$ number of test data with class $\\textit{c}_{\\upsilon}$ and corresponding error rate $err_{\\upsilon}$, we compute MPCE as: $\\frac{1}{\\Upsilon}\\sum \\frac{err_{\\upsilon}}{\\textit{c}_{\\upsilon}}$.\\\\

MPCE seems to a more robust as an evaluator of model performance for different datasets of the classes ~\\cite{wang2017time}). Below in Table ~\\ref{tab_76540}, we demonstrate the MPCE results for the ablation study. In MPCE, our aim is to have a lower value, approaching zero.

Another unique feature of the current work is its response to higher number of test instances when it gets trained with smaller number of training examples. We can quantify the learning gain of ShapAAL at the time of testing as: $\\frac{test\\: accurcay_{ShapAAL} \\; - test\\: accurcay_{Base}}{test\\: accurcay_{Base}}$ and also define training insufficiency factor as: $\\frac{Number\\: of\\: training\\: examples}{Number\\: of \\:testing\\: instances}$. In Fig ~\\ref{fig561}, we demonstrate the comparative study of learning gain of ShapAAL on testing data over base model and the insufficiency in the training. We observe that the learning gain of ShapAAL is mostly $\\ge 1$, while training insufficiency factor $\\le 1$. Hence, we further establish our claim that ShapAAL model is the apt choice under practical constraint of training data limitation in solving the time series classification tasks.

4.Improve the results and discussion section in paragraph.

We have improved the Results Section with detailed discussion and more results to consolidate our claim as mentioned above. The Discussion Section is also improved with additional details.

{Modification/addition in the Discussion Section}

Firstly, we have proposed and validated the unique idea augmentation and ablation of the input features to generate a better learned model. Controlled augmentation of the seen examples to learn better on the unseen examples through introduction of perturbed or virtual data points helps the model to combat the insufficiency in training examples and Shapley-attributed input feature selection refines the input space such that the model gets the opportunity of training more (through augmentation) yet better (Shapley-value based feature ablation). While the augmentation and feature attribution separately improve the test accuracy of the model over different tasks, the combined effect is significant, and it is evident from Table 2 and 4. The study in Table 2 4 clearly indicates that data augmentation through adversarial learning and subsequent feature space identification for re-learning with appropriate features provide significant impetus to the learning process to learn that compensates the limitation in seen examples and learn appropriately. Secondly, we have provided state-of-the-art comparison of the proposed method and the ShapAAL model with both data augmentation and input attribution features has demonstrated consistently outstanding classification performances over different time series classification tasks, conveniently outperforming the current benchmark and state-of-the-art algorithms as depicted in Table 3, Figure 12, and Table 4.

The model is trained off-line, and the trained model is deployed on the cloud or at the local workstation as a clinical analytics engine. The on-field ECG data is given as input to the trained model ShapAAL and the output as one of the disease classes (considering binary or multi-class classification) is considered as the screening outcome. We illustrate the system, which can be potentially developed as an early warning platform for basic CVD screening in Figure 14. Further, we like to mention that clinical screening scenario of the conventional CVD screening and diagnosis need to be changed from a reactive mode to proactive mode. In current conventional setup, users will react when the symptoms flareup. In the most likely scenario, the milder symptoms will be ignored when the clinical facility is far-off. Even the routine check-up, which is necessary for CVD patients may be skipped by the remote patients. Another serious consideration is the missing response of subclinical or non-symptomatic condition of CVDs, where the patient might suddenly develop life-threatening conditions. With the proposed automated CVD screening method that can be conveniently performed at home, we expect that the CVD screening will be proactive with early warning of subclinical or non-symptomatic CVDs.

5.Mention the future scope of your present works.

The future scope of work is elaborated in the revised manuscript.

Our future scope of study includes more exploration towards game theoretic understanding in the construction of a deep learning model with an intuitive rationality perspective of model's dilemma for prediction over unseen data. The general step for Shapley value computation is using sampling method to estimate the expectation over a distribution of marginals and interpretable machine learning fits to such type of quantified notion of an input feature's contribution. We intend to explore the model interpretability and algorithmic transparency as a future research initiative with model-agnostic interpretability indicating marginal contributions for individual input features. Another interesting idea is to investigate virtual adversarial regularization such that we can consider the perspective of model robustness. While a sophisticated model provides outstanding performance on given dataset, the model may be over-sensitive towards a little adversarial attack. Data augmentation is in fact capable of improving the stability of the model where the model does not have a high confidence at the prediction, but those augmented examples are close to the given seen examples. From practical utility perspective, we shall further focus on introducing prescriptive analytics such that the initial treatment directive can be urgently delivered as a basic critical care, which can be lifesaving as well as provides the emergency caregivers the information to immediately start the basic yet immensely important initial basic clinical procedures. For example, after heart attack, each passing minutes cause more heart tissues to get damaged. When the analytics engine detects heart attack, immediate commencement of medications like aspirins, thrombolytics before a cardiologist’s intervention is of immense clinical importance. We intend to bring out a robust remote cardio-vascular management system with automation in the basic screening methods that utilizes the Internet backbone to enable healthcare services to the remotest part of the globe for on-demand screening and basic treatment with both screening and prescriptive functions.________________________________________

Attachment

Submitted filename: Response to Reviewers.pdf

Decision Letter 1

Anand Nayyar

8 Nov 2022

When less is more powerful: Shapley value attributed ablation with augmented learning for practical time series sensor data classification

PONE-D-22-26084R1

Dear Dr. Ukil,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Anand Nayyar, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

The Revised Paper stands Accepted with no further revisions.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuscript has been adapted to the comments of the reviewer. It is legible, well planned, and substantive.

Reviewer #2: all the comments raised are addressed well the authors

the paper is accepted in present form

i wish the authors best of luck

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

Acceptance letter

Anand Nayyar

14 Nov 2022

PONE-D-22-26084R1

When less is more powerful: Shapley value attributed ablation with augmented learning for practical time series sensor data classification

Dear Dr. Ukil:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Anand Nayyar

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Data. Data source.

    Experimental datasets are publicly available at https://figshare.com/articles/dataset/Data_zip/21532440/1.

    (TXT)

    S1 Table. Hyperparameters.

    The hyperparameters used in ShapAAL model construction.

    (PDF)

    S1 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in ChinaTown dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S2 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in Coffee dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S3 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in ECG200 dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S4 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in ECGFiveDays dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S5 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in FreezerRegularTrain dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S6 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in FreezerSmallTrain dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S7 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in ItalyPowerDemandn dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S8 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in MoteStrain dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S9 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in PowerCons dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S10 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in SonyAIBO1 dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S11 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in SonyAIBO2 dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S12 Fig. Study on the training augmentation control.

    We depict the trend of the test accuracy data augmentation control parameter α in TwoLeadECG dataset by varying α from 0.00 ≤ α ≤ 0.07 to understand the response of the model under different strengths of perturbations.

    (TIF)

    S13 Fig. ShapAAL model plot.

    We present the complete model description for reproducibility, where the input is “ECG200” training dataset.

    (TIF)

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Data Availability Statement

    All relevant data is available from the figshare repository (https://doi.org/10.6084/m9.figshare.21532440.v2).


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES