Abstract
Snoring, which may be decisive for many diseases, is an important indicator especially for sleep disorders. In recent years, many studies have been performed on the snore related sounds (SRSs) due to producing useful results for detection of sleep apnea/hypopnea syndrome (SAHS). The first important step of these studies is the detection of snore from SRSs by using different time and frequency domain features. The SRSs have a complex nature that is originated from several physiological and physical conditions. The nonlinear characteristics of SRSs can be examined with chaos theory methods which are widely used to evaluate the biomedical signals and systems, recently. The aim of this study is to classify the SRSs as snore/breathing/silence by using the largest Lyapunov exponent (LLE) and entropy with multiclass support vector machines (SVMs) and adaptive network fuzzy inference system (ANFIS). Two different experiments were performed for different training and test data sets. Experimental results show that the multiclass SVMs can produce the better classification results than ANFIS with used nonlinear quantities. Additionally, these nonlinear features are carrying meaningful information for classifying SRSs and are able to be used for diagnosis of sleep disorders such as SAHS.
1. Introduction
Sleep apnea/hypopnea syndrome (SAHS) is a sleep disorder which is defined by instances of low-level breathing or pauses in breathing during sleep. SAHS is diagnosed by polysomnography (PSG). However, PSG is highly troublesome test owing to the patients sleeping for an overnight in the hospital. Researchers indicate that there is a serious requirement to simplified systems instead of PSG for diagnosing SAHS [1]. For this purpose, some approaches have been developed by using several physiological signals such as electrocardiography (ECG), electroencephalography (EEG), and snore related sounds (SRSs). Recently, there are many studies available for diagnosing SAHS working with the SRSs [2–9].
The snoring is one of the SAHS's symptoms, and it may be interpreted as an indicator for many different diseases such as cardiovascular and sleep disorders [10]. It is stated that snoring is associated with physical and functional defects of upper airway responses and generally occurring in patients having narrower upper airway than the normal ones [3, 4, 6, 11]. Hence, the knowledge about physiology and functions of upper airway can be obtained from SRSs [4]. Because SAHS is concerned with partial or full collapse of upper airways, many studies using different linear techniques have been performed for SRSs analysis [5, 7, 12–16]. The results of these studies indicate that SRSs carry significant information on SAHS.
Recently, some studies have been shown to classify SRSs that is a first task for detecting snoring and SAHS [17–20]. Duckitt et al. [17] presented a method which provides the automatic monitoring of snoring characteristic depending on intensity and frequency of data. They used mel-frequency cepstral coefficients (MFCCs) with hidden Markov model (HMM). Their system obtained 82–89% accuracy in specifying snores. Cavusoglu et al. [18] estimated energy and zero-crossing rates of SRSs and used linear regression method for classifying SRSs' segments as snore/nonsnore. They obtained 86.8% accuracy for obstructive sleep apnea (OSA) patients. Karunajeewa et al. [19] calculated zero-crossing rates, normalized autocorrelation coefficients, energy, and first predictor coefficient of linear predictive coding (LPC) analysis as features for classification of SRSs into snoring, breathing, and silence. They also used three noise reduction techniques and compared their performance. While they had an overall classification accuracy of 90.74%, their accuracy is up to 96.78% with noise reduction techniques and a proper choice of features. Yadollahi and Moussavi [20] tried to classify breath and snore sounds by using Fisher linear discriminant (FLD) based on zero-crossing rates, energy, and first formant of the SRSs that are recorded by two microphones (tracheal and ambient). They investigated the body and neck positions' effects on the classification results. The overall accuracies were found as 95.7% and 93.2% for tracheal and ambient recordings regardless of the neck position, respectively [20].
It is seen that these studies have focused on examining snoring sounds by linear features analysis such as Fourier transform, LPC, but the nonlinear characteristics of SRSs for classifications have not yet been evaluated so far. SRSs are known to have nonstationary and complex behaviours [21, 22]. Some parts of SRSs have nonlinear acoustic vibrations depending on out- and in-factors such as the respiratory airflow strength, vibrations on the soft palate, the shape and differences of upper airway, and the airway obstruction due to tongue subsidence [22, 23]. These make it difficult to solve the identification of SRSs depending on traditional linear frequency and time domain based methods. In these cases, the nonlinear analysis can be performed to reveal the undefined characteristics of the signal. It is possible to use the linear and nonlinear analyses together since these approximations can produce the meaningful results which support each other. For example, a wideband spectrum may be indicator of chaos, but it cannot only identify definition of chaos. However, spectral analysis may give some information for examining the source of variations in the signals.
From the biomedical perspective, nonlinear analyses methods coming from chaos theory have been used to evaluate the states of biological systems and signals from systemic level to the cellular scale [24–28]. The behaviours of dynamical biological systems are interpreted by evaluating the chaotic indicators such as Lyapunov exponents, correlation dimension, Poincare section, and entropy. While some systems show a chaotic behaviour under the normal operating conditions, at the pathological conditions a degree of systems' chaotic behaviour may be reduced or increased, or system behaviour can turn to a regular state [23–28]. Hence, some contributions related to commentating of system functions can be provided by evaluating the effects of pathological or normal conditions on the system behaviour. There are a few studies which investigated the nonlinear properties of SRSs in the literature [29–33]. Sakakura [29] used Poincare section and the largest Lyapunov exponent (LLE) for indicating the chaotic characteristics of snoring sounds. Mikami [30] performed surrogate analysis by using correlation integral and showed that snoring sounds have nonlinear properties. Yılmaz and Ankışhan [31] calculated the LLE values for snore segments obtained from apnea/hypopnea patients and simple snorers. Their results showed that the LLE of simple snorer group was significantly higher than apnea/hypopnea group [31]. Ankışhan and Arı [32] showed from their experimental results that SRSs have chaotic behaviour and carry meaningful information, so the LLE can be used for classification of SRSs. Moreover, they also showed that entropy can be used in classification of SRSs with time domain features [33]. These studies have shown that nonlinear measures can bring to light significant outcomes for SRSs.
The LLE quantifies the dynamic stability of a system and gives information about the predictability of a system [34, 35]. The positive value of the LLE is accepted as a key indicator of chaotic behaviour. Entropy which has also multiple definitions may be interpreted as the average amount of new information obtained by making the measurement in information theory sense [35]. In previous studies, it was reported that these features are the quantitative measures of chaos and frequently used in the nonlinear analysis of the experimental time series [23–33]. So, the aims of this work are to extract meaningful information from SRSs related to its nonlinear characteristics, to investigate the variability of these chaotic measures for different SRSs segments, and to test the availability of these quantities for classifying SRSs. For this purpose, SRSs were analysed by using entropy and the LLE and classified as snore/breathing/silence with support vector machines (SVMs) and adaptive network fuzzy inference system (ANFIS).
There have been several methods which are used for classification of biological data. Two of them are ANFIS and SVMs which are effective methods for classification and machine learning systems such as face detection, speech recognition, pattern recognition and lots of biomedical applications [36–41]. Recently, multiclass SVMs have been very popular and successful for different classification issues [42, 43]. In previous studies, for classifying SRSs, some classification methods such as the linear regression, LPC, HMM, and FLD have been used [17–20]. Unlike the previous ones, in this study, it has been taken into consideration the successfulness of SVMs and ANFIS in biomedical applications; the multiclass SVMs and ANFIS were chosen as classifiers.
2. Material and Methods
2.1. Data
In this study, Cavusoglu et al.'s [18] data set, obtained from Gülhane Military Medical Science Sleep Laboratory, was used. While recording data, a Sennheiser ME 64 condenser microphone with a 40 Hz–20 kHz ± 2.5 dB frequency response was used and placed 15 cm over the patient's head. So, a BNC cable, UA-1000 model multichannel data acquisition system, and a personal computer were used for obtaining the data. The computer was placed out of the room for avoiding its noise. Signal records were made at 16 kHz sampling frequency and 16-bit resolution and stored with patient information. The details about this database are referred to in Cavusoglu et al.'s study [18]. 12 patients' data were used in the experiments of this study. The information of related patients is given in Table 1.
Table 1.
Subject no. | AHI | BMI | Age | Gender |
---|---|---|---|---|
1 | 1.27 | 26.03 | 39 | F |
2 | 9.2 | 27.28 | 51 | M |
3 | 14.3 | 24.38 | 40 | M |
4 | 14.7 | 27.36 | 42 | M |
5 | 17.7 | 27.46 | 44 | M |
6 | 19 | 32.47 | 55 | F |
7 | 20.3 | 28.41 | 36 | M |
8 | 22 | 26.25 | 55 | M |
9 | 24 | 26.78 | 25 | M |
10 | 24.6 | 33.31 | 59 | F |
11 | 24.7 | 26.1 | 67 | M |
12 | 29.86 | 29.67 | 48 | M |
2.2. Entropy
Entropy is defined in terms of a random variable's probability distribution and generally interpreted as a measure of uncertainty. The concept of entropy is used in very large area, from physic to information theory [35]. The amount of entropy is often thought of as the amount of information used to define the system's state [35]. Shannon gives the definition of entropy which states the content of expected information or uncertainty of a probability distribution [44]. In this work, with the considering successfulness about the short signal entropy estimation, Shannon entropy algorithm was used to evaluate the average amount of new information obtained from segments of SRSs [44]. Shannon proposed a quantity called self-information which is a logarithmic function on the interval (0,1] and related to the probability of event X = x i, (i = 1,2,…, n). Self-information h(p i) is given by
(1) |
Average self-information or entropy is derived by weighing the n numbers of self-information values h(p i) by their individual probabilities [44]:
(2) |
2.3. Phase Space Reconstruction
The concepts coming from nonlinear dynamic theory are applied to biomedical time series by using some methods proposed for calculating the chaotic measures (Lyapunov exponents, correlation dimension, etc.). For estimation of these measures, the first step is the reconstruction of system's attractor (phase space reconstruction) from the scalar time series data [34, 35]. Even though the dynamical system might have many variables, the obtaining of one variable from the system is sufficient for reconstructing the phase space of the system [35]. The time-delay technique is easy, systematic, and the most popular approach to reconstruct the attractor [34, 35]. In this study, Takens' time-delay embedding method [45] was used for the reconstruction of the attractor from SRSs segment parts in the phase space. According to embedding theorem of Takens [45], m-dimensional time-delay vectors are reconstructed from SRS segment parts (x(n) = x(nΔt), n = 1,2,…, N), as follows:
(3) |
where x(i) is the ith value of the sound segment part, X i is the ith point on the attractor (i = N − (m − 1)τ), m is the dimension of the phase space (embedding dimension), and τ is the lag or time delay [45]. In this theorem, selection criterion is m ≥ 2D + 1 for embedding dimension; herein, D is the attractor dimension. According to Abarbanel [34], m > D would be sufficient. The chosen embedding dimension and time-delay parameters are important and determined from time series data, since the optimal values of them are not known in advance. There are many proposed methods for choosing their values [35]. In this paper, the false nearest neighbours (FNN) method was applied to select the embedding dimension [46]. The suitable embedding dimension is determined as the phase space dimension for the zero percentage value of FNN where the number of FNN is below a threshold value. The average mutual information (AMI), that is, a nonlinear autocorrelation function, was used for finding the time delay [34, 47]. In this study, the proper value of time delay is selected as the first minimum value of AMI.
2.4. The Largest Lyapunov Exponent
The Lyapunov exponents or characteristic exponents specify the sensitivity to initial conditions of a dynamical system and measure the average rate of convergence or divergence of nearby trajectories on the attractor [34, 35, 48]. For detection of the chaotic behaviour, the calculation of the LLE is sufficient since a positive LLE can be accepted as good evidence for a chaotic system [34, 35, 48]. In this study, the LLE measure was used in the determination of the complexity of the SRS segments.
Several methods estimating the LLE or Lyapunov exponent spectrum have been proposed [48–51]. A simple method which calculates only the LLE is Rosenstein et al.'s algorithm [51], and this algorithm is suitable especially for small noisy datasets. In this paper, Rosenstein et al.'s algorithm [51] was chosen to calculate the LLE of snore episodes. It was stated that this algorithm is a suitable method for embedding dimension is smaller than Takens' selection criterion and for a wide range of time delays [51].
According to Rosenstein et al.'s algorithm [51], after reconstructing the time-delay vectors, the nearest neighbours of each state on the phase space trajectory are searched. It is opined that the jth nearest neighbours' pair diverges nearly at a rate dedicated by the LLE (λ 1) [51]:
(4) |
where C j is the initial separation, Δt is the sampling period, and d j(i) is the distance between the jth pair of nearest neighbours after (i × Δt) seconds. Equation (5) is obtained by taking the logarithm of both sides of (4):
(5) |
This equation gives a set of approximately parallel lines (for j = 1,2,…, M), each with a slope roughly proportional to λ 1. The LLE values of sound segment parts are computed by using a least squares fit to the averaging line defined by
(6) |
where 〈⋯〉 represents the average of all j values [51].
2.5. Support Vector Machines
Support vector machines (SVMs) is a tool used in machine learning. It was proposed by Cortes and Vapnik in 1995 [52]. SVMs that motivate minimizing Vapnik-Chervonenkis (VC) dimension have proved to be very successful in classification learning [52–54]. In this algorithm, it is easy to formulate the decision functions in terms of a symmetric, positive definite and square integrable function k(·, ·) referred to as a kernel. The class of the decision functions also known as kernel classifiers [55] is then given by
(7) |
where training data x i ∈ R d and labels y i ∈ {±1}.
2.5.1. Multiclass Support Vector Machines
SVM has been used in numerous studies of interest in different areas. When SVMs are performed for multiclass problems, the generally used technique is to solve the problem as a collection of some two-class classifications that can be solved by binary SVMs [54].
According to Vapnik [54], some information is stated as follows: suppose that we have a set of training data ((x i, y i), for i = 1,…, n and x i ∈ R d and y i ∈ {±1}) and a nonlinear transformation to a higher dimensional space, the feature space , the SVMs solves
(8) |
(9) |
where (w, b) defines the linear classifier in the feature space. The SVMs try to enforce positive samples to present an output greater than +1, and the negative samples present an output less than −1 for two classes. Those samples are not fulfilling this condition and need a nonzero ζ i in (9), so they will introduce a penalty in the objective function (8). The inclusion of the norm of w in (8) checks whether the solution is maximum margin [54].
The purpose of C∑i ζ i is to control the number of misclassified samples. The usage of a larger value of parameter C, which is chosen by the user, corresponds to assigning a higher penalty to errors. When application of this optimization for solving problems, generally suffering the balance of the samples between the classes and the unequal density of the clusters in the feature space [56]. In multiclass support vector machines, the binary SVMs, that is, using a vector of different weight and bias for each class (w j and b j for j ∈ {1,2,…, k}), are applied. The classifier function of SVMs is given as
(10) |
Accordingly, we can impose (9) for each class it does not belong to, as suggested in [54, 57], leading to
(11) |
where x ij is the ith sample in class j and n j is the number of training samples in that class. Herein, this optimization problem searching the jth output for x ij is larger than any other. In conclusion, the penalization for any incorrectly classified sample will be based on the incorrectly assigned class and on the number of outputs larger than the output of true assigned class [57]. We have focused our study on that this simple model can be extended to solve nonlinear separable problem. So, there are some studies available in the literature, which have proposed to use kernel-based methods. These methods used mapping functions on the input features for carrying them into a very high-dimensional space. So, the methods can construct a hyperplane in that feature space properties rely on kernel functions of SVMs [54, 58].
2.6. Adaptive Network Fuzzy Inference System (ANFIS)
Adaptive network fuzzy inference system (ANFIS) is a tool for using different purposes. Its structure has a fuzzy set “if-then rules.” For generating the suitable input-output pairs, it has a correct membership function [59]. ANFIS has a learning ability having a neuro-fuzzy-type structure, and this network have nodes, where each combination of these nodes placed in the different layers for completed specific functions.
ANFIS has some hybrid learning algorithm, that is, Sugeno-type fuzzy inference systems for identifying parameters. Sugeno-type fuzzy system uses least-squares method which is combined for training membership function parameters of fuzzy system. This system emulates a given training data set [59].
Sugeno-type ANFIS model has five layers for generating inference system [59]. There are several nodes in each layer. These layers are working as follows: the input signals in the present layer are the output signals obtained from nodes in the previous layers [59]. Jang [59] has shown that first-order Sugeno-type model is given as
(12) |
where x 1,2 are inputs, A i and B i are fuzzy sets, f i outputs within the fuzzy region identified by the rules, and the design parameters (p i, q i, and r i) are determined during the training process [59]. ANFIS architecture is given in Figure 1.
In Figure 1, square nodes which are called adaptive nodes are accepted to represent the parameter sets in these nodes which are adjustable [59]. However, circle nodes are fixed nodes that are accepted to represent the fixed parameter sets in the system [59]. The details of ANFIS model can be seen in [59].
3. Results
3.1. Data Analysis and Feature Extraction
In this study, snore, breathing, and silent segments were firstly extracted from the SRSs. The samples of the recorded ambient sounds in time and time-frequency domains are shown in Figure 2. Herein, snore, breathing and silent segments were manually marked in both domains for examining the signals' characteristics. Each segment was automatically subdivided into parts having 3200 samples by developed MATLAB program with thin black lines. These segmentation and subdivision procedure were also shown in Figure 2. In Figure 2, Br.1, Br.2,…, and Br.4 represent breathing segment parts, Sn.1,…, and Sn.3 represent snore segment parts, and Sl.1, Sl.2,…, and Sl.5 represent silent segment parts. We can see from Figure 2, time duration of each segment (in case of inspiration and expiration interval) is different from each other. The frequency and amplitude spectrums of SRSs segments have different characteristics (in Figure 2). It is seen that snore segment parts have a few components for different frequency range as high and low amplitude peaks in Figure 2. However, breathing and silent segments have many components which are similar peak values in a wide frequency band. The separation of breathing and silent segments is difficult since their characteristics are similar to each other, and the noise originated from ambient recordings substantially affects breathing and silent segments. Therefore, auditory and visual examinations were used for marking of the segments.
In case of experiments, the segments were normalized to be in the range of [−0.5,0.5]. This method is used because of each subject having different characteristics (i.e., all snore, breathing, and silent segments' amplitudes of subjects are different owing to their sounds intensity). Therefore, changing range of each sound segment could be different without normalization. It would affect the calculated feature values.
In order to investigate the chaotic behaviour of SRSs, LLE and entropy of each segment parts are calculated. For the LLE calculation, all segment parts are reconstructed in the phase space according to Takens' theorem. Therefore, embedding dimension (m) and time delay (τ) values must be determined, firstly. In order to make a decision about these values, for all SRS segments, embedding dimension (m) and time delay (τ) values are estimated by using the false nearest neighbour (FNN) and average mutual information (AMI) methods, respectively. Looking at these results, it was seen that m and τ values of snore, breathing and silent segment parts are close to each other. Hence, if we chose m to be 5 and τ to be 8 for all segment parts, the estimation of LLE is not significantly affected. So, it is most suitable to set m = 5 and τ = 8 for all segment parts. Herein, some reconstructed attractors depending on SRSs chosen segments projected onto three-dimensional phase space are shown in Figure 3. Certain type of geometrical shape of snore, breathing and silent segments can be seen from the figure. Moreover, sizes of attractors give information to us about how they are differentiated from each other depending on the amplitude values of SRS segments.
Figure 4 gives information about LLE and entropy features of SRS segments. The graph is manually divided and marked with vertical lines to show the segments. The LLE and entropy values of segment parts (square and triangle markers, resp.) are automatically estimated by developed MATLAB program.
It can be seen from Figure 4 that the LLE and entropy values can be differentiated into snore, breathing, and silent segments. For snore segments, LLE values changing between (0.1 × 2) and (1.2 × 2) are less than breathing and silent ones, and entropy values changing between (0.1 × 5) and (1 × 5) are greater than breathing and silent ones. However, at the breathing segments, entropy values are usually greater and LLE values are usually less than silent segments. When looking at silent segments, they generally have bigger values of LLE and smaller values of entropy. Figure 5 gives us any information about all snore/breathing/silent parts distribution in all data sets.
In Figure 5, snore, breathing, and silent segment parts of SRSs are denoted by square, circle, and triangle symbols, respectively. It is seen that snoring segment parts are usually clustered between 2.0 and 4.5 entropy levels and 0.03–0.15 LLE levels; however, some parts of snoring segments are clustered between 0.0 and 1.0 entropy levels and 0.1–0.15 LLE levels. Although snoring segment parts are clustered into two areas, the results show they can be discriminated from breathing and silent segment parts. If looking at the breathing segment parts, which are usually clustered between 0.0 and 3.3 entropy levels and 0.03–0.25 LLE levels. The results show that breathing segment parts cannot easily be classified depending on the distribution area. Because feature distribution of breathing parts is not significantly clustered. However, silent segment parts are generally clustered into two areas: first, between 0.0 and 0.3 entropy levels and 0.15–0.25 LLE levels; second, between 0.3 and 2.5 entropy levels and 0.4–1.0 LLE levels. It can be said that silent segment parts can be discriminated and classified into their own classes.
3.2. Experiments
In this study, the performance of methods in classifying snore, breathing, and silent segments' parts were evaluated in two experimental ways.
Experiment I. The training and test datasets were obtained from the recordings of same patients. The first half interval of the recordings was training data set and the rest was test data set. These datasets obtained from 12 patients include 672 and 668 segments for training and test, respectively. After the subdivision procedure of these segments, training and test databases have 4037 and 5234 segment parts, respectively.
Experiment II. The training and test data sets were obtained from all recordings of different patients. Training data set contains 553 segments from 6 patients, and the rest 6 patients recordings were used for test data set with 787 segments. After the subdivision procedure of these segments, training and test databases have 3322 and 4110 segment parts, respectively.
The details of the training and test datasets in Experiments I and II are summarized in Tables 2 and 3, respectively.
Table 2.
Subject no. | Training | Test | ||||
---|---|---|---|---|---|---|
12 subjects (first half interval) |
12 subjects (rest interval) |
|||||
Snore | Breathing | Silent | Snore | Breathing | Silent | |
1 | 25 | 23 | 16 | 25 | 23 | 15 |
2 | 15 | 15 | 15 | 15 | 15 | 15 |
3 | 15 | 15 | 15 | 15 | 15 | 15 |
4 | 15 | 15 | 15 | 15 | 15 | 15 |
5 | 25 | 25 | 8 | 25 | 25 | 8 |
6 | 39 | 25 | 11 | 39 | 25 | 10 |
7 | 25 | 15 | 25 | 25 | 15 | 25 |
8 | 25 | 25 | 25 | 25 | 25 | 25 |
9 | 25 | 15 | 25 | 25 | 15 | 25 |
10 | 25 | 25 | 10 | 25 | 25 | 10 |
11 | 0 | 10 | 7 | 0 | 10 | 6 |
12 | 25 | 25 | 8 | 25 | 25 | 7 |
| ||||||
Total | 259 | 233 | 180 | 259 | 233 | 176 |
| ||||||
Total | 672 | 668 |
Table 3.
Training | Test | ||||||
---|---|---|---|---|---|---|---|
Subject no. | Snore | Breathing | Silent | Subject no. | Snore | Breathing | Silent |
2 | 30 | 30 | 30 | 1 | 50 | 46 | 31 |
3 | 30 | 30 | 30 | 5 | 50 | 50 | 16 |
4 | 30 | 30 | 30 | 6 | 78 | 50 | 21 |
9 | 50 | 30 | 50 | 7 | 50 | 30 | 50 |
10 | 50 | 50 | 20 | 8 | 50 | 50 | 50 |
11 | 0 | 20 | 13 | 12 | 50 | 50 | 15 |
Total | 190 | 190 | 173 | Total | 328 | 276 | 183 |
| |||||||
Total | 553 | Total | 787 |
3.3. Results of Classifiers
In this study, we presented experimental results obtained from the M-SVMs and ANFIS algorithms, which are three class nonlinearly separable problems. In case of experiments, optimal C value was chosen as 107 and regularization parameter (Lambda) was 10−7 for M-SVMs. However, for ANFIS predictions, maximum number of epochs was chosen as 10. It is seen that this is enough to allow the system converge to a final value. The learning rate was chosen as 0.0, and examined membership functions for ANFIS predictions are Gaussian combination, Gaussian curve, generalized bell-shaped, pi-shaped, sigmoidally shaped, trapezoidal-shaped, and triangular-shaped membership functions. It is observed that the best prediction results in the same conditions obtained by generalized bell-shaped membership function.
For each experiment, the classification performance was evaluated in terms of sensitivity defined as 100 × TP/(TP + FN), positive predictive value (PPV) defined as 100 × TP/(TP + FP), and accuracy 100 × (TP + TN)/(TP + TN + FP + FN), where TP, TN, FP, and FN are the numbers of true positive, true negative, false positive, and false negative classified segment parts, respectively.
Tables 4 and 5 show the classifiers performances of Experiments I and II for snore, breathing, and silent segment parts. Herein, all segments have different number of segment parts (each part having 3200 samples); owing to the segment time interval, the total numbers of segment parts in training and test data sets are different from each other.
Table 4.
Number of segment parts | Snore | Breathing | Silence | Sensitivity (%) | PPV (%) | Total accuracy (%) | |
---|---|---|---|---|---|---|---|
M-SVMs training | |||||||
Snore | 1260 | 1190 | 10 | 60 | 94.44 | 100 | 92.79 |
Breathing | 746 | 0 | 546 | 200 | 73.19 | 94.63 | |
Silence | 2031 | 0 | 21 | 2010 | 98.97 | 88.55 | |
| |||||||
M-SVMs test | |||||||
Snore | 1175 | 1075 | 42 | 58 | 91.49 | 97.64 | 91.61 |
Breathing | 746 | 10 | 457 | 279 | 61.26 | 85.74 | |
Silence | 3313 | 16 | 34 | 3263 | 98.49 | 90.64 |
Number of segment parts | Snore | Breathing | Silence | Sensitivity (%) | PPV (%) | Total accuracy (%) | |
---|---|---|---|---|---|---|---|
ANFIS training | |||||||
Snore | 1260 | 1150 | 91 | 19 | 91.26 | 96.88 | 86.97 |
Breathing | 746 | 32 | 597 | 117 | 80.02 | 62.84 | |
Silence | 2031 | 5 | 262 | 1764 | 86.86 | 92.84 | |
| |||||||
ANFIS test | |||||||
Snore | 1175 | 932 | 202 | 41 | 79.31 | 94.81 | 86.75 |
Breathing | 746 | 35 | 426 | 285 | 57.10 | 57.41 | |
Silence | 3313 | 16 | 114 | 3183 | 96.07 | 90.70 |
Table 5.
Number of segment parts | Snore | Breathing | Silence | Sensitivity (%) | PPV (%) | Total accuracy (%) | |
---|---|---|---|---|---|---|---|
M-SVMs training | |||||||
Snore | 1156 | 1097 | 42 | 17 | 94.90 | 96.65 | 90.07 |
Breathing | 721 | 32 | 525 | 164 | 72.82 | 82.55 | |
Silence | 1445 | 6 | 69 | 1370 | 94.81 | 88.33 | |
| |||||||
M-SVMs test | |||||||
Snore | 1280 | 1121 | 135 | 24 | 87.58 | 85.57 | 84.65 |
Breathing | 772 | 175 | 523 | 74 | 67.75 | 60.32 | |
Silence | 2058 | 14 | 209 | 1835 | 89.16 | 94.93 |
Number of segment parts | Snore | Breathing | Silence | Sensitivity (%) | PPV (%) | Total accuracy (%) | |
---|---|---|---|---|---|---|---|
ANFIS training | |||||||
Snore | 1156 | 971 | 162 | 23 | 83.99 | 96.14 | 84.61 |
Breathing | 721 | 39 | 423 | 259 | 58.67 | 69.00 | |
Silence | 1445 | 0 | 28 | 1417 | 98.06 | 83.40 | |
| |||||||
ANFIS test | |||||||
Snore | 1280 | 938 | 303 | 39 | 73.28 | 87.91 | 80.55 |
Breathing | 772 | 128 | 414 | 230 | 53.62 | 50.79 | |
Silence | 2058 | 1 | 98 | 1959 | 95.19 | 87.92 |
Tables 4 and 5 represent information about percentages of accuracy results of training and test datasets. It can be seen that M-SVMs give a good performance concerning both training and test procedures in Experiments I and II. The best detection performance was achieved in Experiment I where both training and test datasets have different time intervals of same subjects' recordings (Table 4). In Experiment II (Table 5), the SVMs test accuracy was dropped by 6.96% (from 91.61% to 84.65%), while in ANFIS test accuracy was dropped by 6.2% (from 86.75% to 80.55%). Herein, training and test datasets were obtained from different individuals. Additionally, looking at the testing sensitivities of snore and silent segment parts for SVMs was close to training performance, especially in Experiment I. However, breathing segment parts' sensitivity results are worse than others because breathing segment beginning and ending parts are close to snore or silent parts. That is, the entropy and LLE values are similar to others on segment beginning and ending parts. So, classifier is having difficulty in classifying breathing segment parts. The similar results were again obtained by ANFIS, but the accuracy was not feasiblly good as SVMs. Figure 6 gives information about ROC analyses of methods.
4. Discussion and Conclusion
It can be seen from the previously performed studies that some different features, obtained from the analysis of sound signals in spectral and temporal domains, were used for examining the SRSs. In majority of these works, the energy of sounds, zero-crossing rates, MFCCs, and formant frequencies, and so forth, were estimated, and linear frequency analyses such as LPC and Fourier transform were examined [17–20]. SRSs cannot be adequately examined by linearly based methods since they have complex nature [21, 22]. The nonlinear analysis can bring out the SRSs' characteristics related to irregular behaviour.
In this study, the significant information to understanding of chaotic behaviour of SRSs and useful features to classifying the SRSs was expected from the nonlinear measures. For this goal, the LLE and entropy measures were used for classifying SRSs into three classes. The LLE and entropy values of SRSs segments in Experiment II were compared with each other by using unpaired Student's t-test with level of significance P < 0.001 (with Bonferroni correction), statistically. The LLE calculated for all SRSs segments have a positive value, and this result can be accepted as an evidence of chaotic behaviour. It can be recognized that from experiments, that the LLE values of breathing and silent groups were found higher, and entropy values of these two groups were found lower than snore group, significantly. However, for breathing group, the LLE value is lower and the entropy value is higher than silent group, significantly. So, the results have shown that while the LLE values tend to decrease, the entropy values tend to increase from silent to snore segments. This decrement for LLE indicates that dynamics of snore segment have low-level complex behaviour with respect to breathing and silent segments. The reason for this can be interpreted depending on frequency characteristics of SRSs [21]. SRSs segments have different frequency and amplitude spectrums. Snore segments generally consist of simple and complex waveforms and have irregular appearance [21]. They have some high and low amplitude peaks for different frequency range in domain [21]. However, breathing and silent segments show noise-like characteristics which have many similar peak values in a wide frequency band [17, 20]. These segments generally appear as intertwined with each other, with some background noises especially originated from ambient recordings [17, 20]. Because of giving a measure of predictability and complexity of sounds, the increments of LLE show that the predictability of system decreases from snore to silent segment.
The entropy values may be interpreted as the average amount of new information obtained by making the measurement in information theory sense [35]. According to entropy values, it can be said that the information expected from snore segment is greater than breathing and silent. This result can be evaluated from a different perspective; according to Williams [35], high entropy is interpreted as high accuracy in the data, and it shows that the state of system can be determined with great accuracy. The LLE and entropy measurements produced significant values for differentiating these segments if these results were taken into account.
The structure of SRSs shows variations such as silent, breathing, and snore segments, so the degree of chaotic behaviour of SRSs can vary based on these segments. According to LLE values, the complexity of snore segment has lower level than others, and it can be said that, considering entropy values, the state of system can be defined as more accurate than breathing and silent segments. It is found that the complexity increases and amount of new information gained from measurement decreases from snore to silent segments. Since the silent segment has not any information because of noisy characteristics, SRSs are actually produced during breathing process, and therefore the dynamics of these signals depend on the breathing cycle. Cervantes and Femat [60] have studied the phenomenon of breathing in a class of piecewise continuous systems by using a model of linear-driven switched system. It is seen that chaotic and regular phases of breathing evolve irregularly in their system; however, their average behaviour is surprisingly regular depending on bifurcation parameter. Therefore, they have shown that breathing phenomenon occurs the similar structural characteristics with intermittency. If this result is taken into account, it can be said that SRSs show intermittency behaviour since they have different forms of chaotic dynamics.
Different from previous studies [17–20], although there are many classifier methods, multiclass SVMs and ANFIS were selected in this study. First, data was trained by one-against-all SVMs and ANFIS which provide a feasible way to categorize and classify SRSs in experiments. While experiments were realized, different kernel functions such as polynomial, gauss, RBF gauss. were used for obtaining best accuracy results. It has been shown that when gauss-based kernel functions are used for both classifiers, the system accuracy was the best.
The previous studies have shown that classification accuracies of them have varied values depending on their experimental procedures [17–20]. Duckitt et al. [17] obtained snoring detection sensitivity as 89% with training and test data from same patients and 82.2% with training and test data sets from different patients. In Cavusoglu et al.'s study [18], their sensitivities for snore detection were 97.3% with only simple snorers and 90.2% with both simple snorers and OSA patients. For OSA patients, their sensitivity was 86.8%. These results were found when training and test data sets were formed of different individuals [18]. Karunejeewa et al.'s method [19] provided 90.74% total sensitivity for snoring, breathing, and silence classification. They also used three different noise reduction techniques and obtained total maximum sensitivity of 96.78% with proper choices of noise reduction technique and features. Their test set contained different subjects' recordings than those used in training set [19]. Yadollahi and Moussavi [20] used both tracheal and ambient sounds gathered from different body and neck positions to classify them into breath and snore groups. Their results were obtained by using different subjects' samples as training and test data sets [20], and their total accuracies were found as 95.7% and 93.2% for tracheal and ambient recordings, respectively. Table 6 gives information about previous studies and our study accuracy results.
Table 6.
Study | Duckitt et al. [17] | Cavusoglu et al. [18] | Karunajeewa et al. [19] | Yadollahi and Moussavi [20] | Our method |
---|---|---|---|---|---|
Sound types | Ambient sound | Ambient sound | Ambient sound | Ambient and tracheal sound | Ambient sound |
Classes | Snoring and other sounds (silence, breathing, and other types of sounds); snore detection | Snore/nonsnore | Snore, breathing, and silence | Snore and breathing | Snore, breathing, and silence |
Features | 39-dimensional feature vector of energy and MFCC | Spectral energy distributions | Zero-crossings and signal's energy | Zero-crossings signal's energy, and first formant | The largest Lyapunov exponent (LLE) and entropy |
Classifier | HMM | Linear regression | Minimum-probability-of-error decision rule | FLD | M-SVMs and ANFIS |
Accuracy | 82–89% snore sensitivity | 86.8% snore sensitivity | 90.74% total sensitivity | 93.2% for ambient sound total accuracy | In Exp. I: 91.61% (SVMs), 86.75% (ANFIS) total accuracies; and 91.49% (SVMs), 79.31% (ANFIS) snore sensitivities |
In this paper, it is seen that SRSs have nonlinear characteristics, which carry any information about snore, breathing, and silent segments of SRSs. With the help of this information, these segments can be differentiated into different groups. Two experimental procedures were implemented so that training and test data sets were formed in two ways: same (for Experiment I) or different (for Experiment II) subjects' data. The total classifiers performances of Experiment II (84.65% for M-SVMs and 80.55% for ANFIS) were found lower than Experiment I (91.61% for M-SVMs and 86.75% for ANFIS). In previous studies, they generally focused on estimating snore detection sensitivities (Table 6). However, snore detection sensitivities were found as 91.49% for M-SVMs, 79.31% for ANFIS in Experiment I and 87.58% for M-SVMs, 73.28% for ANFIS in Experiment II. The results show that proposed methods especially M-SVMs classification accuracy is as good as previous studies.
Table 6 shows that previous studies have generally focused on linear analysis of SRSs segments and have tried to detect snore segments. In fact, time domain linear features are important for extracting any information from physiological signals. It can be seen from these studies that their results were good enough to classify and differentiate the SRSs. However, keeping in mind the chaotic behaviour of these sounds is also considerable. Experimental results have shown that the LLE and entropy can be important features for clinical diagnosis systems.
Acknowledgments
The authors would like to thank Dr. Tolga Çiloğlu and Dr. Osman Eroğul for the data.
References
- 1.Flemons WW, Littner MR, Rowley JA, et al. Home diagnosis of sleep apnea: a systematic review of the literature. Chest. 2003;124(4):1543–1579. doi: 10.1378/chest.124.4.1543. [DOI] [PubMed] [Google Scholar]
- 2.Perez-Padilla JR, Slawinski E, Difrancesco LM, Feige RR, Remmers JE, Whitelaw WA. Characteristics of the snoring noise in patients with and without occlusive sleep apnea. American Review of Respiratory Disease. 1993;147(3):635–644. doi: 10.1164/ajrccm/147.3.635. [DOI] [PubMed] [Google Scholar]
- 3.Beck R, Odeh M, Oliven A, Gavriely N. The acoustic properties of snores. European Respiratory Journal. 1995;8(12):2120–2128. doi: 10.1183/09031936.95.08122120. [DOI] [PubMed] [Google Scholar]
- 4.Dalmasso F, Prota R. Snoring: analysis, measurement, clinical implications and applications. European Respiratory Journal. 1996;9(1):146–159. doi: 10.1183/09031936.96.09010146. [DOI] [PubMed] [Google Scholar]
- 5.Fiz JA, Abad J, Jané R, et al. Acoustic analysis of snoring sound in patients with simple snoring and obstructive sleep apnoea. European Respiratory Journal. 1996;9(11):2365–2370. doi: 10.1183/09031936.96.09112365. [DOI] [PubMed] [Google Scholar]
- 6.Agrawal S, Stone P, Mcguinness K, Morris J, Camilleri AE. Sound frequency analysis and the site of snoring in natural and induced sleep. Clinical Otolaryngology and Allied Sciences. 2002;27(3):162–166. doi: 10.1046/j.1365-2273.2002.00554.x. [DOI] [PubMed] [Google Scholar]
- 7.Cavusoglu M, Ciloglu T, Serinagaoglu Y, Kamasak M, Erogul O, Akcam T. Investigation of sequential properties of snoring episodes for obstructive sleep apnoea identification. Physiological Measurement. 2008;29(8):879–898. doi: 10.1088/0967-3334/29/8/003. [DOI] [PubMed] [Google Scholar]
- 8.Solà-Soler J, Jané R, Fiz JA, Morera J. Variability of snore parameters in time and frequency domains in snoring subjects with and without Obstructive Sleep Apnea. Proceedings of the 27th Annual International Conference of the Engineering in Medicine and Biology Society (IEEE-EMBS '05); September 2005; New York, NY, USA. pp. 2583–2586. [DOI] [PubMed] [Google Scholar]
- 9.Solà-Soler J, Fiz JA, Morera J, Jané R. Multiclass classification of subjects with sleep apnoea-hypopnoea syndrome through snoring analysis. Medical Engineering and Physics. 2012;34:1213–1220. doi: 10.1016/j.medengphy.2011.12.008. [DOI] [PubMed] [Google Scholar]
- 10.Wilson K, Stoohs RA, Mulrooney TF, Johnson LJ, Guilleminault C, Huang Z. The snoring spectrum: acoustic assessment of snoring sound intensity in 1,139 individuals undergoing polysomnography. Chest. 1999;115(3):762–770. doi: 10.1378/chest.115.3.762. [DOI] [PubMed] [Google Scholar]
- 11.Fujita S, Conway W, Zorick F, Roth T. Surgical correction of anatomic abnormalities in obstructive sleep apnea syndrome: uvulopalatopharyngoplasty. Otolaryngology. 1981;89(6):923–934. doi: 10.1177/019459988108900609. [DOI] [PubMed] [Google Scholar]
- 12.Issa FG, Morrison D, Hadjuk E, et al. Digital monitoring of obstructive sleep apnea using snoring sound and arterial oxygen saturation. Sleep. 1993;16(8):p. S132. doi: 10.1093/sleep/16.suppl_8.s132. [DOI] [PubMed] [Google Scholar]
- 13.Lim PVH, Curry AR. A new method for evaluating and reporting the severity of snoring. Journal of Laryngology and Otology. 1999;113(4):336–340. doi: 10.1017/s0022215100143919. [DOI] [PubMed] [Google Scholar]
- 14.Lee TH, Abeyratne UR. Analysis of snoring sounds for the detection of obstructive sleep apnea. Medical & Biological Engineering & Computing. 1999;37:538–539. [Google Scholar]
- 15.Lee TH, Abeyratne UR, Puvanendran K, Goh KL. Formant-structure and phase-coupling analysis of human snoring sounds for detection of obstructive sleep apnea. Computer Methods in Biomechanics and Biomedical Engineering. 2001;3:243–248. [Google Scholar]
- 16.Abeyratne UR, Patabandi CKK, Puvanendran K. Pitch-jitter analysis of snoring sounds for the diagnosis of sleep apnea. Proceedings of the 23rd Annual International Conference of the IEEE Engineering in Medicine and Biology Society; October 2001; pp. 2072–2075. [Google Scholar]
- 17.Duckitt WD, Tuomi SK, Niesler TR. Automatic detection, segmentation and assessment of snoring from ambient acoustic data. Physiological Measurement. 2006;27(10):1047–1056. doi: 10.1088/0967-3334/27/10/010. [DOI] [PubMed] [Google Scholar]
- 18.Cavusoglu M, Kamasak M, Erogul O, Ciloglu T, Serinagaoglu Y, Akcam T. An efficient method for snore/nonsnore classification of sleep sounds. Physiological Measurement. 2007;28(8):841–853. doi: 10.1088/0967-3334/28/8/007. [DOI] [PubMed] [Google Scholar]
- 19.Karunajeewa AS, Abeyratne UR, Hukins C. Silence-breathing-snore classification from snore-related sounds. Physiological Measurement. 2008;29(2):227–243. doi: 10.1088/0967-3334/29/2/006. [DOI] [PubMed] [Google Scholar]
- 20.Yadollahi A, Moussavi Z. Automatic breath and snore sounds classification from tracheal and ambient sounds recordings. Medical Engineering and Physics. 2010;32(9):985–990. doi: 10.1016/j.medengphy.2010.06.013. [DOI] [PubMed] [Google Scholar]
- 21.Beck R, Odeh M, Oliven A, Gavriely N. The acoustic properties of snores. European Respiratory Journal. 1995;8(12):2120–2128. doi: 10.1183/09031936.95.08122120. [DOI] [PubMed] [Google Scholar]
- 22.Inoue Y, Yamashiro Y. Sleep Disordered Breathing: Update. Tokyo, Japan: Nippon Hyoronsha; 2006. [Google Scholar]
- 23.Jiang JJ, Zhang Y, Stern J. Modeling of chaotic vibrations in symmetric vocal folds. Journal of the Acoustical Society of America. 2001;110(4):2120–2128. doi: 10.1121/1.1395596. [DOI] [PubMed] [Google Scholar]
- 24.Femat R, Alvarez-Ramirez J, Zarazua M. Chaotic behavior from a human biological signal. Physics Letters A. 1996;214(3-4):175–179. [Google Scholar]
- 25.Goldberger AL, Rigney DR, West BJ. Chaos and fractals in human physiology. Scientific American. 1990;262(2):42–49. doi: 10.1038/scientificamerican0290-42. [DOI] [PubMed] [Google Scholar]
- 26.Glass L, Kaplan D. Time series analysis of complex dynamics in physiology and medicine. Medical Progress through Technology. 1993;19(3):115–128. [PubMed] [Google Scholar]
- 27.Fukunaga MS, Arita S, Ishino S, Nakai Y. Quantiative analysisof gastric electric stress response with chaos theory. Biomedical Soft Computing and Human Sciences. 2000;5:59–64. [Google Scholar]
- 28.Quiroz G, Bonifas I, Barajas-Ramirez JG, Femat R. Chaos evidence in catecholamine secretion at chromaffin cells. Chaos, Solitons and Fractals. 2012;45(7):988–997. [Google Scholar]
- 29.Sakakura A. Acoustic analysis of snoring sounds with chaos theory. Proceedings of the International Congress Series; December 2003; pp. 227–230. [Google Scholar]
- 30.Mikami T. Detecting nonlinear properties of snoring sounds for sleep apnea diagnosis. Proceedings of the 2nd International Conference on Bioinformatics and Biomedical Engineering (iCBBE '08); May 2006; Shanghai, China. pp. 1173–1176. [Google Scholar]
- 31.Yılmaz D, Ankışhan H. Analysis of snore sounds by using the largest Lyapunov exponent. Journal of Concrete and Applicable Mathematics. 2011;9:146–153. [Google Scholar]
- 32.Ankışhan H, Arı F. Chaotic analysis of snore related sounds. Proceedings of the IEEE Signal Processing and Communications Applications Conference; April 2012; Fethiye-Muğla, Turkey. IEEE; pp. 1–3. [Google Scholar]
- 33.Ankishan H, Ari F. Snore-related sound classification based on time-domain features by using ANFIS model. Proceedings of the International Symposium on Innovations in Intelligent Systems and Applications (INISTA '11); June 2011; Istanbul, Turkey. IEEE; pp. 441–444. [Google Scholar]
- 34.Abarbanel HDI, Brown R, Sidorowich JJ, Tsimring LS. The analysis of observed chaotic data in physical systems. Reviews of Modern Physics. 1993;65(4):1331–1392. [Google Scholar]
- 35.Williams GP. Chaos Theory Tamed. Washington, DC, USA: Joseph Henry Press; 1997. [Google Scholar]
- 36.Brown MPS, Grundy WN, Lin D, et al. Knowledge-based analysis of microarray gene expression data by using support vector machines. Proceedings of the National Academy of Sciences of the United States of America. 2000;97(1):262–267. doi: 10.1073/pnas.97.1.262. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Furey TS, Cristianini N, Duffy N, Bednarski DW, Schummer M, Haussler D. Support vector machine classification and validation of cancer tissue samples using microarray expression data. Bioinformatics. 2000;16(10):906–914. doi: 10.1093/bioinformatics/16.10.906. [DOI] [PubMed] [Google Scholar]
- 38.Cao L, Gu Q. Dynamic support vector machines for non-stationary time series forecasting. Intelligent Data Analysis. 2002;6:67–83. [Google Scholar]
- 39.Polat K, Yosunkaya Ş, Güneş S. Pairwise ANFIS approach to determining the disorder degree of obstructive sleep apnea syndrome. Journal of Medical Systems. 2008;32(5):379–387. doi: 10.1007/s10916-008-9143-y. [DOI] [PubMed] [Google Scholar]
- 40.Güneş T, Polat E. Feature selection in facial expression analysis and its effect on multi-svm classifiers. Journal of the Faculty of Engineering and Architecture of Gazi University. 2009;24(1):7–14. [Google Scholar]
- 41.Issac Niwas S, Palanisamy P, Chibbar R, Zhang WJ. An expert support system for breast cancer diagnosis using color wavelet features. Journal of Medical Systems. 2011;36:3091–3102. doi: 10.1007/s10916-011-9788-9. [DOI] [PubMed] [Google Scholar]
- 42.Abe S. Analysis of multiclass support vector machines. Proceedings of the International Conference on Computational Intelligence for Modelling and Automation (CIMCA '03); 2003; Vienna, Austria. pp. 385–396. [Google Scholar]
- 43.Angulo C, Parra X, Català A. K-SVCR. A support vector machine for multi-class classification. Neurocomputing. 2003;55(1-2):57–77. [Google Scholar]
- 44.Shannon CE. A mathematical theory of communication. Bell System Technical Journal. 1948;27:623–656. [Google Scholar]
- 45.Takens F. Detecting strange attractors in turbulence. Lecture Notes in Mathematics. 1981;898:366–381. [Google Scholar]
- 46.Kennel MB, Brown R, Abarbanel HDI. Determining embedding dimension for phase-space reconstruction using a geometrical construction. Physical Review A. 1992;45(6):3403–3411. doi: 10.1103/physreva.45.3403. [DOI] [PubMed] [Google Scholar]
- 47.Fraser AM, Swinney HL. Independent coordinates for strange attractors from mutual information. Physical Review A. 1986;33(2):1134–1140. doi: 10.1103/physreva.33.1134. [DOI] [PubMed] [Google Scholar]
- 48.Wolf A, Swift JB, Swinney HL, Vastano JA. Determining Lyapunov exponents from a time series. Physica D. 1985;16(3):285–317. [Google Scholar]
- 49.Sano M, Sawada Y. Measurement of the lyapunov spectrum from a chaotic time series. Physical Review Letters. 1985;55(10):1082–1085. doi: 10.1103/PhysRevLett.55.1082. [DOI] [PubMed] [Google Scholar]
- 50.Briggs K. An improved method for estimating Liapunov exponents of chaotic time series. Physics Letters A. 1990;151(1-2):27–32. [Google Scholar]
- 51.Rosenstein MT, Collins JJ, De Luca CJ. A practical method for calculating largest Lyapunov exponents from small data sets. Physica D. 1993;65(1-2):117–134. [Google Scholar]
- 52.Cortes C, Vapnik V. Support-vector networks. Machine Learning. 1995;20(3):273–297. [Google Scholar]
- 53.Burges CJC, Schölkopf B. Improving the accuracy and speed of support vector machines. In: Mozer M, Jordan M, Petsche T, editors. Neurol Information Processing Systems. Cambridge, Mass, USA: MIT press; 1997. pp. 375–381. [Google Scholar]
- 54.Vapnik V. Statistical Learning Theory. New York, NY, USA: John Wiley & Sons; 1998. [Google Scholar]
- 55.Jaakola T, Diekhans M, Haussler D. Using the fisher kernel method to detect remote protein homologies. Proceedings of the Intelligent Systems for Molecular Biology Conference (ISCB '99); August 1999; Heidelberg, Germany. AAAI press; pp. 149–158. [PubMed] [Google Scholar]
- 56.Arenas-García J, Pérez-Cruz F. Multi-class support vector machines: a new approach. Proceedings of the IEEE International Conference on Accoustics, Speech, and Signal Processing; April 2003; pp. 781–784. [Google Scholar]
- 57.Weston J, Watkins C. Support vector machines for multi-class pattern recognition. Proceedings of the European Symposium on Artificial Neural Networks (ESANN '99); April 1999; Bruges, Belgium. ESANN; pp. 219–224. [Google Scholar]
- 58.Joachims T. Making large-scale SVM learning practical. In: Schِlkopf B, Burges CJC, Smola AJ, editors. Advances in Kernel Methods-Support Vector Learning. Cambridge, Mass, USA: MIT Press; 1999. pp. 169–184. [Google Scholar]
- 59.Jang J-SR. ANFIS: adaptive-network-based fuzzy inference system. IEEE Transactions on Systems, Man and Cybernetics. 1993;23(3):665–685. [Google Scholar]
- 60.Cervantes I, Femat R. Intermittent operation of linear driven switched systems. International Journal of Bifurcation and Chaos. 2008;18(2):495–508. [Google Scholar]