Abstract
In modern manufacturing systems and industries, more and more research efforts have been made in developing effective machine health monitoring systems. Among various machine health monitoring approaches, data-driven methods are gaining in popularity due to the development of advanced sensing and data analytic techniques. However, considering the noise, varying length and irregular sampling behind sensory data, this kind of sequential data cannot be fed into classification and regression models directly. Therefore, previous work focuses on feature extraction/fusion methods requiring expensive human labor and high quality expert knowledge. With the development of deep learning methods in the last few years, which redefine representation learning from raw data, a deep neural network structure named Convolutional Bi-directional Long Short-Term Memory networks (CBLSTM) has been designed here to address raw sensory data. CBLSTM firstly uses CNN to extract local features that are robust and informative from the sequential input. Then, bi-directional LSTM is introduced to encode temporal information. Long Short-Term Memory networks (LSTMs) are able to capture long-term dependencies and model sequential data, and the bi-directional structure enables the capture of past and future contexts. Stacked, fully-connected layers and the linear regression layer are built on top of bi-directional LSTMs to predict the target value. Here, a real-life tool wear test is introduced, and our proposed CBLSTM is able to predict the actual tool wear based on raw sensory data. The experimental results have shown that our model is able to outperform several state-of-the-art baseline methods.
Keywords: machine health monitoring, tool wear prediction, convolutional neural network, recurrent neural network, bi-directional long-short term memory network
1. Introduction
During recent years, machine monitoring systems, including diagnosis and prognosis approaches, have been actively researched [1,2,3,4]. Diagnosis and prognosis systems focus on the detection of faults after the occurrence of certain faults and predictions of the future working conditions and the Remaining Useful Life (RUL), respectively [5,6,7,8,9,10,11]. The methodologies behind these existing machine monitoring systems can be divided into two major categories: physics-based and data-driven models [12,13]. In physics-based models, domain knowledge of physical models and laws with measured data are incorporated into a model constructed via mathematical equations. Previous proposed models, including the Paris crack growth model [14], the Forman crack growth model [15], and so on, all belong to physics-based ones. However, the physics-based models have been criticized for these following points. Firstly, the performance of physics-based models is heavily dependent on the quality and accuracy of domain knowledge about the practical mechanical systems. In real life, due to complexity and noisy working conditions, such a kind of high-quality domain knowledge is often unavailable, which hinders the robustness of these physics-based models. Secondly, most of physics-based models are unable to be updated with on-line measured data, which limits the effectiveness and flexibility of the applications of physics-based models. On the other hand, data-driven models focus on modeling based on historical measured data and try to make a decision from the online data collected from sensors on working machines. Additionally, these model parameters can be updated in real time when the working status of the machines changes. In addition, the development of advanced sensors and computing systems make the research topic of data-driven machine monitoring systems more and more attractive. Therefore, in this paper, our work focuses on this data-driven framework.
As shown in Figure 1, the basic pipeline framework behind data-driven models consists of four major parts: data acquisition, feature extraction/selection, model training and model testing. These systems take various sensor data as inputs and perform feature selections and extractions to derive representation of machine conditions. Then, representations are fed into various algorithms, which normally consist of two parts: one is model training based on historical data, and the other is model prediction based on current sampled data. The core step in data-driven models is representation learning of these sensor data. Sensor data are in nature time series data, which are sampled by sensors and expressed in a sequential form. Some previous works focus on multi-domain feature extractions, including statistical (variance, skewness, kurtosis), frequency (spectral skewness) and time frequency (wavelet coefficients) features. However, these methods do not belong to sequence models, which cannot model the intrinsic sequential characteristic behind sensory data. Additionally, how to select these features is another big challenge for these methods. These models require intensive expert knowledge or feature engineering. Except these methods based on hand-engineered features, some sequence models, including Markov models, Kalman filters and conditional random fields, are powerful for addressing sequential data, which only access raw time series [16,17,18]. However, they have been criticized for the inability to capture long-range dependencies. In sensory data for machine monitoring, two informative and discriminative signals may be separated by many indiscriminative or even noisy signals representing a long time period. Therefore, the long delays that separate some important samples in the time scale may lead to failures of these above sequences models. During recent years, Recurrent Neural Networks (RNNs), especially Long Short-Term Memory (LSTM), that were proposed to relieve the problem of gradient exploding or vanishing in RNN, have emerged as one popular architecture to handle sequential data with various applications, including image captioning, speech recognition, genomic analysis and natural language processing [19,20,21]. LSTM is able to address sequences of varying length data and capture long-term dependencies. As one kind of neural network, LSTM incorporates representation learning and model training together, which require no additional domain knowledge. Additionally, this structure can enable us to discover some unseen structure to improve the generalization capability of the model. Except the necessity of temporal information, raw sensory data usually contain noise. The LSTM models built on top of raw sensory data may not be robust. Therefore, Convolutional Neural Networks (CNN) are introduced to extract local features. The core idea of CNN lies in that abstract features can be extracted by convolutional kernels and the pooling operation. In CNN, the convolutional layers (convolutional kernels) convolve multiple local filters with raw sequential data and generate invariant local features, and the subsequent pooling layers extract the most significant features within fixed length sliding windows. Here, we adopt CNN to extract a sequence of local features from the raw signal firstly.
In this paper, we combine CNN with bi-directional LSTM to propose a novel machine health monitoring system named Convolutional Bi-directional LSTM networks (CBLSTMs). In our proposed CBLSTMs, CNN can extract local robust features, and bi-directional LSTMs, which are built on CNN, are able to encode the temporal information and learn representations. Different from conventional LSTMs that process the input sequence in a feedforward manner, bi-directional LSTMs model the input sequence forward and backward [22]. The core idea behind bi-directional LSTM is that each sequence is presented forwards and backwards to two separate LSTMs, and bi-directional LSTMs can access complete, sequential information about all context information before and after each time step in a given sequence. Here, we adopt one open source dataset: dynamometer, accelerometer and acoustic emission data sampled from high-speed Computer Numerical Control (CNC) milling machine cutters (the dataset has been kindly provided at https://www.phmsociety.org/competition/phm/10). The corresponding task is defined as the estimation of tool wear conditions based on sensory signals, i.e., tool wear depth [23,24]. In our setting, this problem has been transformed into a regression problem with sequential data, in which each sequential datum, i.e., sensory data, represents one certain tool wear condition that corresponds to the actual tool wear width. Several state-of-the-art models are compared with our proposed CBLSTMs model.
This paper is organized as follows. In Section 2, some related work including CNN, LSTM and their various applications, are reviewed. Then, our proposed CBLSTMs are presented in Section 3. Then, in the following Section 4, experimental results of the prediction of tool wear condition are illustrated. Finally, concluding remarks are provided in Section 5.
2. Related Work
2.1. Convolutional Neural Network
CNNs were firstly proposed by LeCun [25] for image processing, which is featured by two key properties: spatially-shared weights and spatial pooling. CNNs have shown their success in various computer vision applications [25,26,27,28] for which the input data are usually 2D data. CNN has also been introduced to address sequential data, including natural language processing and speech recognition [29,30,31]. Generally, to address sequences, the 1D convolutional layer in CNN firstly adopts multiple local filters over the whole sequential input. Each feature map corresponding to each local filter can be generated by sliding the filter over the sequence. Then, the following pooling layer is applied to extract the most vital and fixed-length features from each feature map. Additionally, both convolution and pooling layers can be performed in a stacked way.
In our proposed CBLSTM network, CNN is firstly adopted to process time series data. Then, the outputs of CNN model are then fed into the following bi-directional LSTMs. It is believed that CNN is able to encode more critical information compared to the raw sequential input, considering that the single time step information may not be discriminative enough. In addition, CNN is able to compress the length of the sequence, which increases the capability of the subsequent recurrent models to capture temporal information.
2.2. From RNN to LSTM
Recurrent Neural Networks (RNNs) were proposed for sequence learning. RNNs build connections between units from a directed cycle. Different from the basic neural network, multi-layer perceptron, which can only map from input data to target vectors, RNN is able to map target vectors from the entire history of previous inputs in principal. RNN allows a memory of previous inputs to be kept in the network’s internal state. RNNs can be trained via backpropagation through time for supervised tasks. However, the vanishing gradient problem during backpropagation for model training hinders the performance of RNN. This means that traditional RNN may not capture long-term dependencies. Therefore, LSTMs were firstly presented to prevent backpropagated errors from vanishing or exploding. Forget gates were introduced in LSTMs to avoid the long-term dependency problem. These adopted forget gates are able to control the utilization of information in the cell states. To capture nonlinear dynamics in time series sensory data and learn effective representation of machine conditions, LSTMs should be superior compared to traditional RNNs due to their capability to capture long-term dependencies. Considering that LSTMs are able to capture long-range dependencies and nonlinear dynamics in time series data, LSTMs have been successfully applied in various applications, including speech recognition, image captioning, handwriting recognition, genomic analysis and natural language processing.
Our proposed CBLSTM utilizes bi-directional LSTM to model temporal information. Bi-directional LSTM processes input sequences in forward and backward directions and is able to summarize temporal information from past and future contexts. Adopting the bi-directional structure, the past and future dependency information are both exploited to capture the temporal information.
2.3. Neural Network for Machine Health Monitoring
Due to the strong representation capability of multi-layer neural networks, neural networks have been widely applied to machine health monitoring problems [32,33,34,35,36,37,38,39,40]. Most of the previous works do not consider the sequential nature behind sensory data. Before feeding raw data into the neural network, feature extraction and selection are performed firstly [32,33,34,35,41,42]. These works do not consider the order of the signal and require some feature engineering. After wavelet techniques, CNN has been applied on the time frequency map of time series data for fault diagnosis [38]. Compared to the previous CNN model, CNN in our proposed CBLSTM only takes raw sensory data as input. Additionally, CBLSTM utilizes CNN to extract local features instead of the final representation of the whole sequence. In addition, some papers have applied RNNs including LSTMs to machine health monitoring problems [36,37,39]. In this paper, we further proposed bi-directional LSTMs combined with CNN to address machine health monitoring problems.
3. Models
Before the presentation of our proposed CBLSTMs, some adopted notations in this paper are introduced firstly. In this paper, machine health monitoring problems are cast into a specific one: the tool wear prediction problem. The in-process sensory data as time series observations are used as input data. The task is defined as designing a model to infer the tool wear conditions from these in-process multi-sensory signals. Let a series of observations denote the acquired data for the i-th machine condition sample. Additionally, represents the multi-sensory data sampled at time step t, which is a vector, and d is the dimensionality of sensory data. l is the length of the sensory signal. For each sequential datum , the corresponding actual tool wear condition (flank wear width) is measured and recorded as . The tasks are defined to predict based on sequential sensory data .
Our proposed CBLSTMs consists of two major parts: one is the local feature extractor, CNN, and the other one is the temporal encoder, bi-directional LSTMs. After applying one-layer CNN on the raw input sequence to extract local and discriminative features, two-layer bi-directional LSTMs are built on top of the previous CNN to encode the temporal patterns. Then, two fully-connected dense layers are stacked together to process the outputs of LSTMs. Finally, a linear regression layer is adopted to predict the tool wear depth. The whole structure of the proposed CBLSTM is shown in Figure 2.
3.1. Local Feature Extractor: CNN
The adopted one-layer CNN consists of two layers: one convolutional layer and one pooling layer, which directly process the raw input sequence. The convolutional layer slides the filters over the whole input sequence to generate feature maps. Each feature map can be regarded as the convolutional activation of the corresponding filter over the whole sequence. It is assumed that k filters with a window size of m are used in the convolutional layer. Then, the pooling layer is applied to compress each generated feature map to produce significant features. The details of these two layers are presented in the following subsections:
Convolution: The convolution operation is defined as a multiplication operation between a filter vector and a concatenation vector representation given by:
(1) |
where represents a window of m continuous time steps starting from the i-th time step. In addition, a bias term b is also added into the convolution operation, so that the final operation is given as:
(2) |
where denotes the transpose of a matrix * and g is a non-linear activation function that is set to Rectified Linear Units(ReLu)in our model [43].
Each vector can be regarded as a filter, and the single value can be regarded as the activation of the window.
Max-pooling: The convolution operation over the whole sequence is applied by sliding the filtering window from the beginning time step to the ending time step. It is easily shown that a feature map is a vector denoted as follows:
(3) |
where the index j denotes the j-th filter. It corresponds to multi-windows as . The pooling layer is able to reduce the length of the feature map, which can further minimize the number of model parameters. The hyper-parameter of pooling layer is the pooling length denoted as s. The max operation is taking a max over the s consecutive values in feature map .
Then, the compressed feature vector can be obtained as:
(4) |
where . Generally, multiple filters are applied with different initialized weights to derive the output of the CNN layer.
Generally, the size of the input sequence in the CNN layer is , and n is the number of data samples. The size of the corresponding outputs is . It is easily shown that after the convolutional and pooling operation, the length of sequence data can be compressed from l to . Compared to the original representation is raw sensory data with a dimensionality of d that is usually the number of sensors in each time step; more abstract and informative representation can be learned after CNN, and the corresponding dimensionality is k, which is the number of filters. Therefore, CNN plays the role of feature extractor to feed better sequential representation into the subsequent LSTM models compared to the raw sequential data. To give a clear illustration, the framework for the local feature extractor based on CNN has been displayed in Figure 3.
3.2. Temporal Encoder: Bi-Directional LSTMs
Here, a two-layer bi-directional LSTM is built on the top of the CNN. This recurrent model is applied to summarize the temporal information. After introducing the basic LSTMs, deep bi-directional LSTM is presented.
3.2.1. Basic LSTMs
The core idea behind LSTMs lies in that at each time step, a few gates are used to control the passing of information along the sequences that can capture long-range dependencies more accurately. In our paper, we adopt one popular LSTM framework proposed in [44]. In LSTM, at each time step t, hidden state is updated by current data at the same time step , the hidden state at the previous time step , the input gate , the forget gate , the output gate and a memory cell . The following updating equations are given as follows:
(5) |
where model parameters including all , and are shared by all time steps and learned during model training, σ is the sigmoid activation function, ⊙ denotes the element-wise product and k is a hyper-parameter that represents the dimensionality of hidden vectors. Here, Equation (5) defines the hidden layer function .
Firstly, the basic LSTM is constructed to process the sequential data in time order. Additionally, the output at the terminal time step is used to predict the output by a linear regression layer, as shown in the following equation.
(6) |
where and z is the dimensionality of the output. In our tasks, the output is the flank wear width, so that . For model training, the predicted tool wear value is compared with the true tool wear value y to derive the Mean Squared Error (MSE) as model loss.
(7) |
where n is the training sample size. The corresponding LSTMs architecture is shown in Figure 4a.
3.2.2. Bi-Directional LSTMs
It is easily shown that basic LSTMs are only able to access the previous context of each specific time step. However, in machine health monitoring systems, the sequential sensory data have strong temporal dependencies. It is meaningful to consider the future context. Therefore, the bi-directional LSTM is applied here. Bi-directional LSTMs are able to process the sequence data in two directions including forward and backward ways with two separate hidden layers and then feed forward to the same output layer. The following equations define the corresponding hidden layer function, and the → and ← denote the forward and backward process, respectively.
(8) |
(9) |
Then, the complete BLSTM hidden element representation is the concatenated vector of the outputs of forward and backward processes as follows:
(10) |
Deep bi-directional LSTM: During recent years, deep architectures have been shown to be successful in representation learning [45,46]. Therefore, it is meaningful to stack multiple LSTM layers to form a deep LSTM neural network. The core idea behind the deep neural network is that inputs to the model should go through multiple non-linear layers. When it comes to deep LSTMs, the input to the model can be passed through multiple LSTM layers. As shown in Figure 4c, the hidden output of one LSTM layer is not only propagated through time, but also used as the input data to the next LSTM layer. The output sequence of one layer is fed into the next layer. In the framework of bi-directional LSTM, every hidden layer receives an input sequence that consists of the output sequences of forward and backward layers at the level below.
For Layer 1, the input data are the sequence output of the previous CNN model. Additionally, the output of the last LSTM layer at the terminal time step is adopted as the output of our deep bi-directional LSTM. The advantages of stacking of LSTM layers are two-fold. One is that stacking layers enable the model to learn the characteristic of the raw signal at different time scales. The other is that parameters can be distributed over the space, i.e., layers, instead of increasing memory size, which can contribute to more effective non-linear operations of the input raw signal. The architectures of bi-directional LSTMs and deep bi-directional LSTMs are shown in Figure 4b,c.
3.3. Fully-Connected and Linear Regression Layers
In this part, the output representation of the temporal encoder, a two-layer bi-directional LSTMs, is fed into multiple hidden layers to seek a higher-level representation. Two fully-connected dense layers are stacked together, in which the output of one layer is used as the input into the next layer. The computation in each layer is given by:
(11) |
where and denote the output and input of the i-th fully-connected layer, respectively. and represent the transformation matrix and the bias term in the i-th fully-connected layer, respectively. The function is also set to be ReLu. Additionally, the output of the last layer, , is regarded as the final representation of the input sequence, assuming c fully-connected dense layers are successively stacked.
The final learned representation of the raw signal is fed into the final linear regression layer, which predicts the actual tool wear.
3.4. Training and Regularization for CBLSTMs
Given the predicted outputs and true targets, the mean squared errors over training data can be calculated and back-propagated to update model parameters. The optimization method named Root Mean Square Propagation(RMSprop)that utilizes the magnitude of recent gradients to normalize the gradients is adopted to optimize model parameters over the objective function [47].
Due to the model complexity of deep learning methods, the large scale of training data is vital for the model’s robust performance. In machine monitoring problems, it is hard to obtain a large scale of training data. Therefore, the regularization technique is applied for our proposed models. Dropout was introduced during model training [48]. Via dropout, parts of the hidden outputs are randomly masked so that these neurons will not influence the forward propagation during training procedures. When it comes to testing phases, the dropout will be turned off, and the outputs of all hidden neurons will have effects on model testing. From another view, dropout can be regarded as an approach to enlarge the training data size. During each training epoch, the application of random masking noise creates novel variants of data samples. In our models, we adopted one dropout layer between LSTM models and the first fully-connected layer and another dropout layer between the first fully-connected layer and the second fully-connected layer. Their masking probabilities are both set to 0.5.
4. Experiments
In this section, we empirically evaluated the performances of our proposed CBLSTM. The tool wear monitoring task is conducted. Firstly, the dataset descriptions are given. Then, details about the experimental setup are provided. Finally, the comparison results are shown and discussed.
4.1. Descriptions of Datasets
To experimentally verify the performance of CBLSTM, a high speed CNC machine was run under dry milling operations [49]. The schematic diagram of the experimental platform is shown in Figure 5. The operation parameters are as follows: the running speed of the spindle was 10,400 rpm; the feed rate in the x direction was 1555 mm/min; the depth of cut (radial) in the y direction was 0.125 mm; the depth of cut (axial) in the z direction was 0.2 mm. To acquire data related to this CNC machine’s operation condition, a Kistler quartz 3-component platform dynamometer was mounted between the workpiece and the machining table to measure cutting forces, while three Kistler piezo accelerometers were mounted on the workpiece to measure the machine tool vibration in the x, y and z directions, respectively [49]. Therefore, six different signals acquired by these corresponding six sensors were used in our experiments. DAQ NI PCI1200 was adopted to perform in-process measurements, including force and vibration in three directions (x, y, z) with a continuous sampling frequency of 50 kHz during the tool wear test. Considering that the sampling frequency is quite high such that each data sample has over 100 thousands time steps, the whole sequence is divided into 100 sections, and the max and mean values of each section are kept to form a new time step. By doing this, each data sample is converted into a sequential datum whose length is 100, and the dimensionality of each time step is 12. The corresponding flank wear of each individual flute was measured offline using a LEICA MZ12 microscope after finishing each surface, which is considered to be one cut number in the following data analysis, which will be the target value. Finally, three tool life tests named C1, C4 and C6 were selected as our dataset. Each test contains 315 data samples, while each data sample has a corresponding flank wear. For training/testing splitting, a three-fold setting is adopted such that two tests are used as the training domain and the other one is used as the testing domain. For example, when C4 and C6 are used as the training datasets, C1 will be adopted as the testing dataset. This splitting is denoted as c1. The details about training/testing splitting are shown in Table 1. Our task is defined as the prediction of tool wear depth based on the sensory input. To facilitate the training, the target value of training data is firstly scaled into a range [0, 1]. Additionally, the predicted value of testing data will be inverse transformed and then compared to ground-truth values.
Table 1.
Symbol | Training Domain | Testing Domain |
---|---|---|
c1 | C4, C6 | C1 |
c4 | C1, C6 | C4 |
c6 | C1, C4 | C6 |
4.2. Experimental Setup
The following methods will be compared:
-
*
LR: Linear Regression on extracted features of raw signal;
-
*
SVR: Support Vector Regression on extracted features of the raw signal;
-
*
MLP: Multi-layer neural network on extracted features of the raw signal;
-
*
RNN: Basic RNN on the raw signal;
-
*
Deep RNN: A two-layer RNN on the raw signal;
-
*
LSTM: A one-layer LSTM with dropout on the raw signal;
-
*
Deep LSTM: A two-layer LSTM with dropout on the raw signal;
-
*
BLSTM: A bi-directional LSTM with dropout on the raw signal;
-
*
Deep BLSTM: A two-layer bi-directional LSTM with dropout on the raw signal.
Since LR, SVR and MLP cannot address sequential data, feature extraction is conducted firstly. The same setting in [50] was adopted here, and 9 measures (e.g., RMS, variance, wavelet energy, etc.) that are illustrated in Table 2 were extracted from the six sensors. Then, each machine condition can be represented by a 54-dimensional vector, which is fed into the subsequent regression models, including LR, SVR and MLP. LR has no hyperparameter. In SVR, we search the best regularization parameter C from {0.001, 0.01, 0.1, 1, 10}. For MLP, three fully-connected layers with layer sizes of [162, 162, 108] are designed with the activation function as the sigmoid.
Table 2.
Domain | Features | Expression |
---|---|---|
Statistical | RMS | |
Variance | ||
Maximum | ||
Skewness | ||
Kurtosis | ||
Peak-to-Peak | ||
Frequency | Spectral Skewness | |
Spectral Kurtosis | ||
Time-Frequency | Wavelet Energy |
Six recurrent models, including RNN, deep RNN, LSTMs, deep LSTMs, bi-directional LSTMs and deep bi-directional LSTMs, are compared. These models can directly process time series data, so that feature extraction is not required here. The input sequence has 100 time steps, and each time step is described by a vector with a dimensionality of 12. For RNN and deep RNN, one-layer with a size of 28 and two layers with sizes of [28,56] recurrent models are fed into two fully-connected layers with a size of [80,100] and the final linear regression layer, respectively. For LSTMs and deep LSTMs, the LSTM layer is replaced with the RNN layer, while the other setting is kept the same as RNN and deep RNN models. Compared to LSTM and deep LSTM, bi-directional processing of input sequences is adopted, and the other settings are kept unchanged.
For our proposed CBLSTM, one-layer CNN is firstly designed, whose filter number, filter size and pooling size are set to 150, 10 and 5. Therefore, the shape of the raw sensory sequence is changed from to after CNN. Then, a two-layer bi-directional LSTMis built on top of the CNN. Backward and forward LSTMs share the same layer sizes as [150, 200]. Therefore, the output of the LSTM module is the concatenated vector of the representations learned by backward and forward LSTMs, and its dimensionality is 400. Then, before feeding the representation into the linear regression layer, two fully-connected layers with a size of [500, 600] are adopted. The nonlinearity activation functions in our proposed CBLSTM are all set to ReLu.
It should be stated that in our experiment, the selection of the hyperparameter is cross-validated in a portion of training data. To quantitatively evaluate the performances of all compared models, two measures are adopted, including Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE). MAE is the average value of the absolute values of the errors. RMSE is the square root of the average of the square of all of the errors. The corresponding equations for the calculations of these two measures are given as follows:
(12) |
(13) |
where and are true and predicted tool wear depth.
4.3. Experimental Results on Tool Wear Prediction
In this section, we show a comparison of LSTMs with several benchmark methods. Additionally, MAE and RMSE of all methods on three different datasets are shown in Table 3 and Table 4, respectively.
Table 3.
Category | Methods | Datasets | ||
---|---|---|---|---|
c1 | c4 | c6 | ||
Regression Models | LR | 24.4 | 16.3 | 24.4 |
SVR | 15.6 | 17.0 | 24.9 | |
MLP | 24.5 | 18.0 | 24.8 | |
Recurrent Models | RNN | 13.1 | 16.7 | 25.5 |
Deep RNN | 7.8 | 9.4 | 19.3 | |
LSTMs | 19.6 | 15.6 | 25.3 | |
Deep LSTMs | 8.3 | 8.7 | 15.2 | |
BLSTMs | 9.9 | 10.8 | 15.7 | |
Deep BLSTMs | 8.7 | 6.7 | 14.4 | |
Our Model | CBLSTMs |
Table 4.
Category | Methods | Datasets | ||
---|---|---|---|---|
c1 | c4 | c6 | ||
Regression Models | LR | 31.1 | 19.3 | 30.9 |
SVR | 18.5 | 19.6 | 31.5 | |
MLP | 31.2 | 20.0 | 31.4 | |
Recurrent Models | RNN | 15.6 | 19.7 | 32.9 |
Deep RNN | 12.5 | 11.8 | 22.9 | |
LSTMs | 23.9 | 20.8 | 32.4 | |
Deep LSTMs | 12.1 | 10.2 | 18.9 | |
BLSTMs | 12.3 | 14.7 | 20.8 | |
Deep BLSTMs | 11.5 | 9.1 | 18.9 | |
Our Model | CBLSTMs |
We firstly observed that among regression models, including LR, SVR and MLP based on expert features, LR performs worst due to the limitation of linear models. SVR with the RBF kernel and MLP with the sigmoid activation function are both nonlinear models that can capture the nonlinear relationships between the expert features and the tool wear. However, recurrent models based on the raw input signal all outperform these models based on expert features. It has been shown that recurrent neural network models are able to learn meaningful representations from the raw signal without any feature engineering.
Among recurrent models, deep models that contain two hidden recurrent layers always perform better than their corresponding normal models that only contain one hidden recurrent layer. It is shown that deep models are able to learn more abstract and discriminative representation due to stacked hidden layers. When it comes to different recurrent units, including basic recurrent, LSTM and bi-directional LSTM ones, bi-directional LSTM performs the best, and the basic recurrent model performs the worst. It is obvious that basic LSTMs perform slightly better than RNN. The reasons may be the fact that gate functions employed in LSTMs can enable it to capture long-term dependency better than RNN. Additionally, the bi-directional structure can discover future information compared to the forward recurrent structure.
It is shown that our proposed CBLSTM model achieves the best performance among all compared methods. Compared to the most competitive model, deep BLSTMs, CBLSTM adopts convolutional neural network to address the raw signal and then builds recurrent modules on top of CNN. The experimental results have verified the effectiveness of the convolutional operation in our proposed CBLSTMs. The CNN is adopted to extract local features, which can filter the noise in the raw signal effectively. In addition, CNN can also reduce the length of sequential data. In our case, the length of raw sequential data is reduced from 100 to 19, and the short sequential data can be more easily captured by the following recurrent model.
At last, to qualitatively demonstrate the effectiveness of CBLSTM models, the predicted tool wears under different datasets are illustrated in Figure 6. The actual tool wear conditions measured offline by a microscope are also displayed, respectively. It is found that the predicted tool wear overall is able to follow the trend of the groundtruth data well.
Our proposed CBLSTM was conducted using a NVIDIA Tesla K40c GPU on a Windows Server with a dual-2.70-GHz CPU and a RAM of 512 GB. The training time for one epoch is 5 s, and the testing time for each sample of our algorithm is only 0.027 s, so that CBLSTM can be an effective solution for real-time machine monitoring.
4.4. Effects of Dropout and Bi-Directional Modules on the Performances of CBLSTM
In the CBLSTM model, two key modules are adopted, including dropout and bi-directional recurrent structures. In the following, their effectiveness is experimentally investigated.
Dropout module: The applied dropout layer can relieve the possible overfitting problem. To verify the effectiveness of the dropout module, CBLSTM without dropout was run on these three datasets. The MAE and RMSE measures are shown in Figure 7 and Figure 8. Additionally, the predicted tool wears under different datasets are illustrated in Figure 9. It is shown that the dropout operations are able to reduce the regression error of our CBLSTM method.
Bi-directional module: The adopted bi-directional module can enable our CBLSTM method to consider the previous and future context of each time step. Here, the performances of Convolutional LSTM networks (CLSTM) were evaluated, and the corresponding MAE and RMSE measures are shown in Figure 7 and Figure 8. Compared to CBLSTM, CLSTM uses normal LSTM as the temporal encoder, and other hyperparameters are set to be the same as CBLSTM. Further, the tool wears predicated by CLSTM under different datasets are illustrated in Figure 10. Experimental results state that the bi-directional structure is able to improve the regression performance.
5. Conclusions
In this work, CBLSTM has been proposed to address tool wear prediction tasks. In CBLSTM, CNN is firstly designed to extract local features from raw sequential data. Then, a bi-directional LSTM is adopted to encode the temporal information. As an advanced recurrent model, bi-directional LSTMs are able to capture long-term dependencies in forward and backward ways. Additionally, the stacked LSTM layers can enable our module to learn more abstract and deep features. It is shown that CBLSTM does not require any expert knowledge and feature engineering. In the task of tool wear prediction, experimental results have verified the superior performance of our CBLSTM method. Therefore, our proposed CBLSTM is able to capture and discover meaningful features under the sensory signal for machine health monitoring.
In future work, we plan to introduce wavelet transformation, which is an effective tool to analyze the machine sensory signal, into the deep neural network models. The combination of shallow feature extraction and deep feature extraction methods may be more effective and efficient in the area of machine health monitoring.
Acknowledgments
This work has been supported in part by the National Natural Science Foundation of China (51575102).
Author Contributions
Rui Zhao implemented the algorithm and wrote the manuscript, Rui Zhao and Jinjang Wang analyzed the data, Ruqiang Yan and Kezhi Mao designed the research and revised the manuscript. All authors have read and approved the final manuscript.
Conflicts of Interest
The authors declare no conflict of interest.
References
- 1.Yin S., Li X., Gao H., Kaynak O. Data-Based Techniques Focused on Modern Industry: An Overview. IEEE Trans. Ind. Electron. 2015;62:657–667. doi: 10.1109/TIE.2014.2308133. [DOI] [Google Scholar]
- 2.Kothamasu R., Huang S.H., VerDuin W.H. Handbook of Maintenance Management and Engineering. Springer; Berlin/Heidelberg, Germany: 2009. System health monitoring and prognostics–a review of current paradigms and practices; pp. 337–362. [Google Scholar]
- 3.Chen X., Yan R., Liu Y. Wind turbine condition monitoring and fault diagnosis in China. IEEE Instrum. Meas. Mag. 2016;19:22–28. doi: 10.1109/MIM.2016.7462789. [DOI] [Google Scholar]
- 4.Zhang C., Yao X., Zhang J., Jin H. Tool Condition Monitoring and Remaining Useful Life Prognostic Based on a Wireless Sensor in Dry Milling Operations. Sensors. 2016;16:795. doi: 10.3390/s16060795. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Lasheras F., Nieto P., de Cos Juez F., Bayón R., Suárez V. A Hybrid PCA-CART-MARS-Based Prognostic Approach of the Remaining Useful Life for Aircraft Engines. Sensors. 2015;15:7062–7083. doi: 10.3390/s150307062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Yan R., Gao R.X., Chen X. Wavelets for fault diagnosis of rotary machines: A review with applications. Signal Process. 2014;96:1–15. doi: 10.1016/j.sigpro.2013.04.015. [DOI] [Google Scholar]
- 7.Qian Y., Yan R. Remaining Useful Life Prediction of Rolling Bearings Using an Enhanced Particle Filter. IEEE Trans. Instrum. Meas. 2015;64:2696–2707. doi: 10.1109/TIM.2015.2427891. [DOI] [Google Scholar]
- 8.Yang F., Habibullah M.S., Zhang T., Xu Z., Lim P., Nadarajan S. Health Index-Based Prognostics for Remaining Useful Life Predictions in Electrical Machines. IEEE Trans. Ind. Electron. 2016;63:2633–2644. doi: 10.1109/TIE.2016.2515054. [DOI] [Google Scholar]
- 9.Wang J., Gao R.X., Yan R. Integration of EEMD and ICA for wind turbine gearbox diagnosis. Wind Energy. 2014;17:757–773. doi: 10.1002/we.1653. [DOI] [Google Scholar]
- 10.Zhao R., Yan R., Gao R.X. Dual-scale cascaded adaptive stochastic resonance for rotary machine health monitoring. J. Manuf. Syst. 2013;32:529–535. doi: 10.1016/j.jmsy.2013.05.009. [DOI] [Google Scholar]
- 11.Jiang W., Xie C., Zhuang M., Shou Y., Tang Y. Sensor Data Fusion with Z-Numbers and Its Application in Fault Diagnosis. Sensors. 2016;16:1509. doi: 10.3390/s16091509. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Yu M., Wang D., Luo M. Model-based prognosis for hybrid systems with mode-dependent degradation behaviors. IEEE Trans. Ind. Electron. 2014;61:546–554. doi: 10.1109/TIE.2013.2244538. [DOI] [Google Scholar]
- 13.Jardine A.K., Lin D., Banjevic D. A review on machinery diagnostics and prognostics implementing condition-based maintenance. Mech. Syst. Signal Process. 2006;20:1483–1510. doi: 10.1016/j.ymssp.2005.09.012. [DOI] [Google Scholar]
- 14.Li Y., Kurfess T., Liang S. Stochastic prognostics for rolling element bearings. Mech. Syst. Signal Process. 2000;14:747–762. doi: 10.1006/mssp.2000.1301. [DOI] [Google Scholar]
- 15.Oppenheimer C.H., Loparo K.A. Physically based diagnosis and prognosis of cracked rotor shafts. Proc. SPIE. 2002;4733 doi: 10.1117/12.475502. [DOI] [Google Scholar]
- 16.Taborri J., Scalona E., Palermo E., Rossi S., Cappa P. Validation of Inter-Subject Training for Hidden Markov Models Applied to Gait Phase Detection in Children with Cerebral Palsy. Sensors. 2015;15:24514–24529. doi: 10.3390/s150924514. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Ke W., Wu L. Mobile Location with NLOS Identification and Mitigation Based on Modified Kalman Filtering. Sensors. 2011;11:1641–1656. doi: 10.3390/s110201641. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Yang H.D. Sign Language Recognition with the Kinect Sensor Based on Conditional Random Fields. Sensors. 2014;15:135–147. doi: 10.3390/s150100135. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Auli M., Galley M., Quirk C., Zweig G. Joint Language and Translation Modeling with Recurrent Neural Networks; Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing; Seattle, Washington, USA. 18–21 October 2013; pp. 1044–1054. [Google Scholar]
- 20.Graves A., Liwicki M., Fernández S., Bertolami R., Bunke H., Schmidhuber J. A novel connectionist system for unconstrained handwriting recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2009;31:855–868. doi: 10.1109/TPAMI.2008.137. [DOI] [PubMed] [Google Scholar]
- 21.Karpathy A., Fei-Fei L. Deep visual-semantic alignments for generating image descriptions; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Boston, MA, USA. 7–12 June 2015; pp. 3128–3137. [Google Scholar]
- 22.Schuster M., Paliwal K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997;45:2673–2681. doi: 10.1109/78.650093. [DOI] [Google Scholar]
- 23.D’Addona D., Matarazzo D., Ullah A.S., Teti R. Tool wear control through cognitive paradigms. Procedia CIRP. 2015;33:221–226. doi: 10.1016/j.procir.2015.06.040. [DOI] [Google Scholar]
- 24.D’Addona D.M., Ullah A.S., Matarazzo D. Tool-wear prediction and pattern-recognition using artificial neural network and DNA-based computing. J. Intell. Manuf. 2015:1–17. doi: 10.1007/s10845-015-1155-0. [DOI] [Google Scholar]
- 25.Le Cun B.B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D. Advances in neural information processing systems 2. Morgan Kaufmann Publishers; San Francisco, CA, USA: Handwritten digit recognition with a back-propagation network; pp. 396–404. [Google Scholar]
- 26.Jarrett K., Kavukcuoglu K., Lecun Y. What is the best multi-stage architecture for object recognition?; Proceedings of the 2009 IEEE 12th International Conference on Computer Vision; Kyoto, Japan. 29 September–2 October 2009; pp. 2146–2153. [Google Scholar]
- 27.Krizhevsky A., Sutskever I., Hinton G.E. Imagenet classification with deep convolutional neural networks; Proceedings of the 2012 Advances in Neural Information Processing Systems; Lake Tahoe, NV, USA. 3–6 December 2012; pp. 1097–1105. [Google Scholar]
- 28.Gao Y., Lee H. Local Tiled Deep Networks for Recognition of Vehicle Make and Model. Sensors. 2016;16:226. doi: 10.3390/s16020226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Abdel-Hamid O., Mohamed A.R., Jiang H., Penn G. Applying convolutional neural networks concepts to hybrid NN-HMM model for speech recognition; Proceedings of the 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); Kyoto, Japan. 25–30 March 2012; pp. 4277–4280. [Google Scholar]
- 30.Kim Y. Convolutional neural networks for sentence classification. arXiv. 2014.
- 31.Zhao R., Mao K. Topic-Aware Deep Compositional Models for Sentence Classification. IEEE/ACM Trans. Audio Speech Lang. Process. 2017;25:248–260. doi: 10.1109/TASLP.2016.2632521. [DOI] [Google Scholar]
- 32.Su H., Chong K.T. Induction Machine Condition Monitoring Using Neural Network Modeling. IEEE Trans. Ind. Electron. 2007;54:241–249. doi: 10.1109/TIE.2006.888786. [DOI] [Google Scholar]
- 33.Yoon H., Park C.S., Kim J.S., Baek J.G. Algorithm learning based neural network integrating feature selection and classification. Expert Syst. Appl. 2013;40:231–241. doi: 10.1016/j.eswa.2012.07.018. [DOI] [Google Scholar]
- 34.Rafiee J., Arvani F., Harifi A., Sadeghi M. Intelligent condition monitoring of a gearbox using artificial neural network. Mech. Syst. Signal Process. 2007;21:1746–1754. doi: 10.1016/j.ymssp.2006.08.005. [DOI] [Google Scholar]
- 35.Sun W., Shao S., Zhao R., Yan R., Zhang X., Chen X. A sparse auto-encoder-based deep neural network approach for induction motor faults classification. Measurement. 2016;89:171–178. doi: 10.1016/j.measurement.2016.04.007. [DOI] [Google Scholar]
- 36.Malhi A., Yan R., Gao R.X. Prognosis of Defect Propagation Based on Recurrent Neural Networks. IEEE Trans. Instrum. Meas. 2011;60:703–711. doi: 10.1109/TIM.2010.2078296. [DOI] [Google Scholar]
- 37.Tse P., Atherton D. Prediction of machine deterioration using vibration based fault trends and recurrent neural networks. Jo. Vib. Acoust. 1999;121:355–362. doi: 10.1115/1.2893988. [DOI] [Google Scholar]
- 38.Wang J., Zhuang J., Duan L., Cheng W. A multi-scale convolution neural network for featureless fault diagnosis; Proceedings of the 2016 International Symposium of Flexible Automation (ISFA); Cleveland, OH, USA. 1–3 August 2016; pp. 1–6. [Google Scholar]
- 39.Zhao R., Wang J., Yan R., Mao K. Machine health monitoring with LSTM networks; Proceedings of the 2016 10th International Conference on Sensing Technology (ICST); Nanjing, China. 11–13 November 2016; pp. 1–6. [Google Scholar]
- 40.Li K., Chen P., Wang S. An Intelligent Diagnosis Method for Rotating Machinery Using Least Squares Mapping and a Fuzzy Neural Network. Sensors. 2012;12:5919–5939. doi: 10.3390/s120505919. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Cerrada M., Sánchez R., Cabrera D., Zurita G., Li C. Multi-Stage Feature Selection by Using Genetic Algorithms for Fault Diagnosis in Gearboxes Based on Vibration Signal. Sensors. 2015;15:23903–23926. doi: 10.3390/s150923903. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Zhu D., Bai J., Yang S.X. A Multi-Fault Diagnosis Method for Sensor Systems Based on Principle Component Analysis. Sensors. 2009;10:241–253. doi: 10.3390/s100100241. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Nair V., Hinton G.E. Rectified Linear Units Improve Restricted Boltzmann Machines; Proceedings of the 27th International Conference on Machine Learning (ICML-10); Haifa, Israel. 21–24 June 2010; pp. 807–814. [Google Scholar]
- 44.Gers F.A., Schmidhuber J., Cummins F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000;12:2451–2471. doi: 10.1162/089976600300015015. [DOI] [PubMed] [Google Scholar]
- 45.Hinton G.E. Learning multiple layers of representation. Trends Cognit. Sci. 2007;11:428–434. doi: 10.1016/j.tics.2007.09.004. [DOI] [PubMed] [Google Scholar]
- 46.Bengio Y. Learning deep architectures for AI. Found. Trends® Mach. Learn. 2009;2:1–127. doi: 10.1561/2200000006. [DOI] [Google Scholar]
- 47.Tieleman T., Hinton G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA Neural Netw. Mach. Learn. 2012;4:2. [Google Scholar]
- 48.Hinton G.E., Srivastava N., Krizhevsky A., Sutskever I., Salakhutdinov R.R. Improving neural networks by preventing co-adaptation of feature detectors. arXiv. 2012.
- 49.Li X., Lim B., Zhou J., Huang S., Phua S., Shaw K., Er M. Fuzzy neural network modelling for tool wear estimation in dry milling operation; Annual conference of the prognostics and health management society; San Diego, CA, USA. 27–30 September 2009; pp. 1–11. [Google Scholar]
- 50.Wang J., Xie J., Zhao R., Zhang L., Duan L. Multisensory fusion based virtual tool wear sensing for ubiquitous manufacturing. Robot. Comput. Integr. Manuf. 2017;45:47–58. doi: 10.1016/j.rcim.2016.05.010. [DOI] [Google Scholar]