Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2023 Apr 8;23(8):3827. doi: 10.3390/s23083827

Multiscale Convolutional Neural Network Based on Channel Space Attention for Gearbox Compound Fault Diagnosis

Qinghong Xu 1, Hong Jiang 1,*, Xiangfeng Zhang 1, Jun Li 1, Lan Chen 1
Editor: Hossam A Gabbar1
PMCID: PMC10141628  PMID: 37112168

Abstract

Gearboxes are one of the most widely used speed and power transfer elements in rotating machinery. Highly accurate compound fault diagnosis of gearboxes is of great significance for the safe and reliable operation of rotating machinery systems. However, traditional compound fault diagnosis methods treat compound faults as an independent fault mode in the diagnosis process and cannot decouple them into multiple single faults. To address this problem, this paper proposes a gearbox compound fault diagnosis method. First, a multiscale convolutional neural network (MSCNN) is used as a feature learning model, which can effectively mine the compound fault information from vibration signals. Then, an improved hybrid attention module, named the channel–space attention module (CSAM), is proposed. It is embedded into the MSCNN to assign weights to multiscale features for enhancing the feature differentiation processing ability of the MSCNN. The new neural network is named CSAM-MSCNN. Finally, a multilabel classifier is used to output single or multiple labels for recognizing single or compound faults. The effectiveness of the method was verified with two gearbox datasets. The results show that the method possesses higher accuracy and stability than other models for gearbox compound fault diagnosis.

Keywords: gearboxes, compound faults, multiscale feature extraction, attentional mechanisms, multilabel classification

1. Introduction

Gearboxes are one of the most widely used speed and power transfer elements in rotating machinery and play a vital role in the manufacturing industry. Gearbox failures can lead to unplanned shutdowns, causing substantial economic losses and even significant casualties. Therefore, in order to prevent accidents and ensure the efficient operation of mechanical systems, it is essential to carry out fault diagnosis of gearboxes [1,2].

Among the gearbox fault diagnosis methods, intelligent fault diagnosis is one of the most widely used methods. Hajnayeb et al. [3] designed a fully connected neural network to classify four operating conditions using gearbox vibration signals. D.J. Bordoloi et al. [4] used different optimization algorithms to optimize support vector machine parameters for the multi fault classification of gears using time-frequency vibration data. Hu et al. [5] proposed a fault diagnosis method based on multiscale dimensionless metrics and random forest that can identify different types of faults with an average accuracy of 95.58%. Traditional intelligent fault methods perform feature extraction using signal processing [6,7,8] and then use shallow machine learning [9,10,11] models for fault identification. However, these methods require extensive diagnostic expertise and signal-processing calculations during feature extraction, which are time-consuming and dependent on expert experience [12]. Moreover, as its feature extraction and classification processes are designed independently, unsynchronized model optimization will consume significant time and limit its diagnostic performance [13].

Driven by artificial intelligence technology, deep learning is also widely used in fault diagnosis. Compared with traditional intelligent fault methods, deep learning-based methods are free from reliance on expert knowledge and signal pre-processing methods and have obvious advantages in the face of massive data. Zhang et al. [14] proposed a wide kernel convolutional neural network that uses wide convolutional kernels in the first convolutional layer to suppress high-frequency noise and uses small convolutional kernels to extract deep features. The results showed that the model is robust to variable load conditions and noise. Xia et al. [15] proposed a multi-sensor-based convolutional neural network diagnosis method to fuse multiple sensor signals for gearbox compound fault diagnosis. These methods have achieved fruitful results in fault diagnosis. However, fault-related information contained in vibration signals is usually distributed on multiple scales, and this must be considered. Most only perform single-scale feature extraction and cannot obtain multiscale fault information.

In recent years, some scholars have attempted to construct multiscale convolutional neural network models for capturing the multiscale features of vibration signals. Chen et al. [16] combined a multiscale convolutional neural network with a long- and short-term memory network for fault diagnosis of rolling bearings. Jiang et al. [17] constructed a multiscale convolutional neural network to identify different health states of wind turbine gearboxes. K.N. Ravikumar et al. [18] proposed a multilayer long and short-term memory multiscale deep residual learning fault diagnosis model in internal combustion engine gearboxes. Although these multiscale convolutional neural networks are enriched with fault feature information, they combine the captured multiscale features; precisely, it ignores the different importance of the scale features learned in different branches for the final diagnosis decision, and even some of this misinformation can mislead the diagnosis decision [19].

Attention mechanisms consider weighting effects, i.e., mapping relationships between inputs and outputs, which can enhance critical features and weaken redundant components [20,21]. Therefore, introducing the attention mechanism into the diagnostic model can improve the validity and reliability of the method. Jang et al. [22] introduced spatial attention into autoencoders and designed attentional self-encoders to learn and calibrate the location information in the potential space. Plakias et al. [23] introduced spatial attention into a dense convolutional neural network to improve the feature extraction capability of the model and achieve recognition of bearings with different loss levels. Li et al. [24] designed a fusion strategy based on a channel attention mechanism to obtain more fault-related information during the fusion of multisensor data features. Xie et al. [25] constructed an improved convolutional neural network incorporating a channel attention mechanism for fault diagnosis of diesel engine systems. Huang et al. [26] designed a hybrid attention method to adaptively select important features through tandem spatial and channel attention. They proposed a shallow multiscale CNN with mixed attention for bearing fault diagnosis. Ye et al. [27] presented the same spatial attention as the SE-Net process but with different dimensions. They connected the channel attention in series with the proposed spatial attention for bearing fault diagnosis. Despite the success of traditional intelligent diagnosis methods, there are still some limitations, mainly in the form of:

(1) For spatial attention, the information interaction between channels is ignored because the features of each channel are treated equally. Similarly, channel attention is directly global to the information within the channel and tends to ignore the information interactions in space [28].

(2) In the face of compound faults, conventional classifiers can only output one label instead of multiple labels, meaning it is impossible to classify compound faults into two or more single faults. Compound faults are considered a special independent failure mode in fault diagnosis [29]. For single-class faults and normal states, there exist 2nn+1 classes of compound fault forms [30]. The increase in fault types causes the dimensionality of the deep learning model’s output layer to increase, leading to more network parameters and training difficulties. To address the above problems, this paper proposes a channel–space attention multiscale convolutional neural network (CSAM-MSCNN) based on the composite fault diagnosis method for gearboxes.

The main contributions of this paper are as follows:

(1) A channel–space attention module (CSAM) is proposed to solve the side effects of the bottleneck structure in hybrid attention. Then a CSAM-MSCNN is designed based on the CSAM and multiscale convolutional neural network (MSCNN), which can effectively solve the feature redundancy problem brought by multiscale feature fusion. In addition, the CSAM-MSCNN also takes into account the interaction of channel information and the interaction of spatial information.

(2) For solving the deficiency of a traditional fault diagnosis mode in a compound fault, a gearbox compound fault diagnosis method is proposed by combining a multilabel classifier and CSAM-MSCNN, which can decouple the compound fault into multiple single faults.

(3) Two gearbox datasets are leveraged to validate the compound fault diagnosis capability of the proposed method. The experimental results show that the proposed method has higher prediction accuracy for compound fault than the classical diagnosis methods.

The rest of the paper is organized as follows: Section 2 details the basic theory of convolutional neural networks and multilabel classification. Section 3 presents the proposed CSAM-MSCNN and the intelligent fault diagnosis method. In Section 4, the feasibility and superiority of the method are verified by two experimental studies, and the results are compared with other intelligent diagnostic methods using the same dataset. Conclusions and future research work are discussed in Section 5.

2. Theoretical Backgrounds

2.1. Convolutional Neural Network

A CNN, a supervised deep learning method, has achieved excellence in speech recognition, image classification, and target detection. Typically, CNNs contain a convolutional layer, an activation layer, a pooling layer, a classification layer, and some techniques to improve the model’s generalization ability, such as batch normalization.

2.2. Convolutional Layer

The convolutional layer is the core component of the CNN model, enabling local connectivity and weight sharing through convolutional kernels. The convolution kernel slides along the input feature map and performs convolution operations with the data in the perceptual field to achieve feature extraction. The specific arithmetic formula is as follows:

yjl+1=ikxilwijl+bjl (1)

where xil is the input at the first level, wijl is the weight matrix, bjl is the bias, and yjl+1 is the output in the first level.

2.3. Batch Normalization Layer

Batch normalization (BN) is used to reduce the internal covariance bias and accelerate the training of deep neural networks. The BN layer is usually added after the convolutional or fully connected layer and before the activation function. The mathematical formula for the BN layer is as follows:

μ=1Nbatchs=1Nbatchxs (2)
σ2=1Nbatchs=1Nbatchxsμ2 (3)
x^s=xsμσ2+ε (4)
ys=γx^s+β (5)

where Nbatch denotes the number of batch data, xs denotes the sth input, μ and σ2 denote the mean and variance of the batch data, respectively, ε denotes a constant close to 0 but greater than 0, x^s denotes the result of data normalization, γ and β denote the parameters that the network can learn, and ys denotes the sth output of the data after BN.

2.4. Activation Function

The activation function enables the network to gain nonlinear representation capability, enhances the feature representation of the model, and maps otherwise linearly indistinguishable multidimensional features to another space, making the learned features easier to distinguish. In recent years, rectified linear units (ReLU) have been widely used as activation functions to accelerate the convergence of neural networks. The ReLU is calculated as follows:

fx=max0,x (6)

2.5. Pooling Layer

The pooling layer is usually placed after the convolutional layer to reduce the number of feature dimensions and parameters and to prevent network overfitting. A typical pooling layer includes maximum pooling and average pooling with the following operations:

pil+1j=maxj1W+1tjWatlt (7)
pil+1j=avgj1W+1tjWatlt (8)

where Pil+1 denotes the value of the neuron after the output of the pooling layer, atlt denotes the value of the tth neuron in the lth layer, tj1W+1,jW, W indicates the width of the pooling area.

2.6. Fully Connected Layer

After multiple convolutional and pooling layers, a few fully connected layers are usually secured and used to integrate the extracted features. The activation function in the fully connected layer usually uses the ReLU function. For a one-dimensional input xiL1 of length M, the fully connected layer has N neurons and the output of each neuron can be expressed as:

zjL=ReLUi=1MxiL1wj,iL+bjL,j=1,2,,N (9)

where wj,iL is the connection weight from the ith neuron in the L1th layer to the jth neuron in the Lth layer, xiL1 denotes the input of the jth neuron of the Lth layer, bjL denotes the bias of the jth neuron of the Lth layer, M and N denote the number of neurons in layer L1 and layer L, respectively.

2.7. Multilabel Classifier

The SoftMax classifier (SC) is trained using a cross-entropy loss function and is widely used in multiclassification tasks. Given a training set xi,yii=1M, where xi and yi are the ith sample and label, respectively, and M is the total number of samples. The SoftMax function has the following operator equation:

si=softmaxyiFC=eyiFCc=1CeycFC (10)

where yiFC is the output of the ith neuron in the last fully connected layer, sis=s1,s2,,sC, C is the number of categories. The label of SC can be calculated by the following equation:

labelsoftmax=argmaxs (11)

where argmaxs denotes the index value of the maximum value in the vector s.

The cross-entropy loss function has a more vital global optimization ability and faster convergence than other loss functions. It is widely used in deep convolutional neural networks with the following equation:

J=1Mm=1Mc=1C1ym=clogscm (12)

where 1· is the indicator function that returns 1 if 1· turns out to be true and 0 if it is false.

Since the traditional SC cannot handle the multilabel problem, some changes are made in the output layer activation function, loss function, and labels. First, the Sigmoid function is chosen as the activation function for the last fully connected layer. Similarly, given the training set xi,yii=1M, where xi and yi are the ith sample and label, respectively, and M is the total number of samples, the Sigmoid function is calculated as follows:

σi=σyiFC=11+eycFC (13)

where yiFC is the output of the ith neuron in the previous fully connected layer and σi denotes the confidence probability of the ith category, σiσ=σ1,σ2,,σC, C is the number of categories. The Sigmoid function ensures that the value of the last fully connected layer is converted to between 0 and 1, and that the confidence probabilities of each category do not interact with each other.

Then, in multiclassification, a sample has only one label, while a sample in multilabel classification may contain more than one label. The multilabel classification problem is usually converted into an N binary classification problem, so the binary cross-entropy function is used as the loss function to train the CSAM-MSCNN network, which is calculated as follows:

J=1Ci=1Cyilogσi+1yilog1σi (14)

Finally, a class-unique thermal coding is used for the labels, i.e., the number of all single-fault types is used as the length of the label vector, and the compound fault is considered as a combination of multiple single faults so that its label exists as one at various indices.

The SC takes out the fault type with the highest confidence probability as its prediction label by the argmax function. The multilabel classifier (MLC) filters the confidence probabilities by a threshold ϕ to output predicted labels. If the confidence probability of the fault type is greater than the threshold value, the value at the index of its predicted label vector is set to 1. Otherwise, it is set to 0. The threshold value in this paper is set to 0.7.

The differences between SC and MLC are shown in Figure 1. Squares, circles, and triangles indicate three different single faults, and the overlapping forms of triangles and circles indicate compound faults. The MLC can accurately identify and decouple compound faults by outputting labels for two or more single fault types. Still, the SC classifies compound faults as error categories because it cannot decouple them.

Figure 1.

Figure 1

SoftMax classifier and multilabel classifier.

3. The Proposed Method

3.1. Channel–Space Attention Module

The convolutional block attention module (CBAM) [21] is a typical hybrid attention method used for lifetime prediction and machinery fault diagnosis. The CBAM structure is shown in Figure 2, and it consisted of a channel attention module and a spatial attention module connected in series. CBAM combines the advantages of channel and spatial attention and considers the information interaction in the channel and space. It is better than a single channel or spatial attention module. The channel attention module in the CBAM is shown in Figure 3. First, the spatial information in the average pooling and maximum pooling aggregated feature maps was used. Then, two sets of channel attention weights were obtained through a shared fully connected layer, and the two sets of weights were combined. Finally, the final channel attention weights were received after the Sigmoid function. Although channel attention implements interactions between channels using a shared fully connected layer, the dimensionality reduction due to its bottleneck structure has side effects on channel attention prediction. It is not efficient or necessary to capture the interactions between all channels [31].

Figure 2.

Figure 2

Structure of the CBAM.

Figure 3.

Figure 3

Channel Attention Module in the CBAM. C stands for channel and L stands for length.

Inspired by the spatial attention module structure in the CBAM, an improved channel attention mechanism was proposed using a one-dimensional convolution instead of the original shared fully connected layer. The side effects of bottleneck structures were avoided while ensuring cross-channel information interaction and reducing model complexity. The design of the specific improved channel attention module is shown in Figure 4, and the use space attention module in this paper is shown in Figure 5. The improved Channel Attention Module (CAM) was connected in series with the original Spatial Attention Module (SAM) to form the Channel–Spatial Attention Module (CSAM) used in this paper.

Figure 4.

Figure 4

Improved Channel Attention Module.

Figure 5.

Figure 5

Spatial Attention Module in the CSAM.

The channel attention module in the CSAM was processed as follows: First, the input feature map Fin is fed into the channel attention module. Then the average pooling and maximum pooling are used to aggregate the spatial information of Fin, and two different vectors Favgc, Fmaxc describing the spatial information are obtained. The Favgc,Fmaxc deformations are stitched together to obtain the spatial information synthesis description vector. After the one-dimensional convolution process, the weights are then compressed to 0~1 by the Sigmoid activation function. The channel attention weights Mc are obtained after deformation reduction. The channel attention weight Mc is multiplied with the input feature map Fin to obtain the channel attention corrected feature map Fc.

Favgc=AvgpoolFin (15)
Fmaxc=MaxpoolFin (16)
Fsumc=ConcatenatereshapeFavgc,reshapeFmaxc (17)
Mc=ReshapeσconvFsumc (18)
Fc=McFin (19)

The spatial attention module in the CSAM handles the process as follows: First, the feature map Fc corrected by channel attention is input into the spatial attention module. Then two different channel information description vectors Favgs, Fmaxs are obtained using the average pooling and maximum pooling aggregation Fc of channel information, respectively. The spatial information synthesis description vector Fsums is obtained by stitching Favgs, Fmaxs together. The spatial attention weight Ms is obtained after a one-dimensional convolution process and then by a Sigmoid activation function. The spatial attention Ms is multiplied by Fc to obtain the spatial attention corrected feature map Fs, which is the output feature Fout after correction by the CSAM module.

Favgs=AvgpoolFc (20)
Fmaxs=MaxpoolFc (21)
Fsums=ConcatenateFavgs,Fmaxs (22)
Ms=σconvFsums (23)
Fout=Fs=MsFc (24)

3.2. Channel–Spatial Attention Multiscale Convolutional Neural Networks

The network architecture constructed in this paper is shown in Figure 6, which mainly included raw signal input, multiscale feature extraction based on channel–space attention, and fault diagnosis. First, the original data samples of the same length X1,X2,X3,,Xn were sampled from the vibration signal and fed into the network. The fault information was extracted through the feature extraction layer, and finally, all the fault information was aggregated in the classification layer, and diagnostic decisions were made.

Figure 6.

Figure 6

CSAM-MSCNN architecture diagram.

The feature extraction layer performed feature extraction on the input sample X using three different scales. The first scale obtained the scale information of the vibration signal by averaging pooling and subsequently mined its deeper fault information using four convolution blocks. Each convolutional block contained a one-dimensional convolutional layer, a batch normalization layer, an activation function ReLU, and a highest pooling layer. The second scale directly used four convolutional blocks for feature extraction of the original scale signal. The difference is that the first layer convolution kernel used a wide convolution kernel with a sizeable sensory field to obtain global information and shallow features. Then three convolution blocks were used to mine deep features. Unlike the first scale, the third scale used maximum pooling to obtain the scale information of the vibration signal, followed by the same four convolutional blocks to learn the fault information in the scale signal. The extracted features at three scales are f1,f2, f3, and fused to obtain the fused feature F.

f1=convconvconvconvavgpoolX (25)
f2=convconvconvconvX (26)
f3=convconvconvconvmaxpoolX (27)
F=Concatenatef1,f2,f3 (28)

To solve the feature redundancy problem caused by feature fusion, the CSAM assigned different weights to the fused features. Attention-corrected features were obtained by enhancing critical fault information and suppressing useless information. Finally, the feature map was downscaled by global average pooling to get the feature map Fa. The feature map obtained from the feature extraction layer was flattened and sent to the fully connected layer. After two fully connected layers, the confidence probability P for each fault type was calculated using the Sigmoid function.

Fa=GAPAttentionF (29)
P=σLinearLinearFlattenFa (30)

The main advantages of the CSAM-MSCNN proposed in this paper are as follows: (1) The feature extraction layer of the network can significantly enrich the fault feature information obtained by feature extraction of the input fault samples from three different scales. (2) Using the hybrid channel–space attention mechanism, the large amount of fault feature information extracted from the multiscale structure is weight mapped to enhance key features and weaken redundant CSAM, effectively solving the feature redundancy problem caused by multiscale feature fusion and improving the efficiency of fault diagnosis. (3) CSAM optimizes the bottleneck structure of the channel attention module in CBAM. While ensuring cross-channel information interaction, the side effects of the bottleneck structure are avoided, and the model complexity is also reduced. (4) The CSAM-MSCNN uses a Sigmoid function with a binary cross-entropy loss, which provides independent probabilities for each class and an unrestricted sum of possibilities for all classes, solving the problem that traditional intelligent fault diagnosis cannot classify compound faults into two or more single faults.

3.3. Fault Diagnosis Method Based on CSAM-MSCNN

The CSAM-MSCNN proposed in this paper consisted of three gearbox compound fault diagnosis parts. The first part was to collect the gearbox vibration signal using the digital acquisition system, sample it and divide the dataset. The second part was to train the CSAM-MSCNN model. The third part was the performance test evaluation of the completed training model to realize the compound fault diagnosis of the gearbox. In this method, there was no need to extract and select features manually. The original vibration signal could be used directly to achieve fault diagnosis in an end-to-end manner. Figure 7 shows the diagnostic flow chart of the CSAM-MSCNN, and the detailed steps of fault diagnosis are as follows:

Figure 7.

Figure 7

Fault diagnosis process based on the CSAM-MSCNN.

(1) Acquisition of gearbox vibration signals using acceleration sensors.

(2) The sample dataset is constructed by the overlapping sampling of the original vibration signals, and the training set, validation set, and test set are divided randomly in proportion.

(3) Construct the CSAM-MSCNN model, set the network hyperparameters such as learning rate, sample batch size, and the number of training iterations, and initialize the network weights.

(4) The CSAM-MSCNN is trained using the training set samples, and the error between the network output and the sample labels is calculated using the binary cross-entropy loss function. The error is back-propagated according to the Adam gradient descent algorithm, and the network weight parameters are updated up to the set number of training iterations. The network is evaluated using the validation set for each training round, but backpropagation of errors and network weight updates are not performed.

(5) Performance test evaluation of the completed trained CSAM-MSCNN using the test set. The network prediction labels of the test set samples are compared with the sample labels to obtain the test set sample accuracy.

4. Experimental Verification

4.1. Experiment 1: Parallel Bearing Gearbox Fault Diagnosis

4.1.1. Description of Experimental Data

To verify the effectiveness and superiority of the method in this paper, we used SQI’s wind turbine drive system fault diagnosis test bench to conduct compound fault diagnosis experiments and collect relevant fault data. The structure of the test bench is shown in Figure 8, which mainly contained a motor, a motor controller, a two-stage parallel shaft gearbox, a planetary gearbox, and a magnetic powder brake. A sketch of the structure of the secondary parallel shaft gearbox is shown in Figure 9. The number of gear teeth on the input and output shafts was 36 and 100, respectively, and the number of teeth of the two gears on the idler shaft was 90 and 28, respectively. The gear modulus was 1.5, and the gearbox ratio was 8.92.

Figure 8.

Figure 8

Wind turbine drive system fault diagnosis test bench.

Figure 9.

Figure 9

Structure diagram of parallel shaft gearbox.

The experimental data were collected at a gearbox with 0 loads and a fixed speed of 1500 rpm. The sensor used to collect the data was a piezoelectric triaxial acceleration sensor. The sensor’s sensitivity was 95.8 mV/g, and the output bias was 10.9 VDC. The gearbox failure parts used in the experiment are shown in Figure 10, containing seven single failure types, such as bearing rolling body failure, bearing outer ring failure, bearing inner ring failure, broken teeth, missing teeth, tooth surface wear, and tooth root cracking. When the data were collected, the faulty gear and bearing were installed in the second parallel shaft gearbox. The data contained 12 gearbox states, including normal conditions, seven single-fault, and four compound-fault states. The status information of the compound fault is shown in Table 1.

Figure 10.

Figure 10

Faulty parts in the gearbox. (a) Bearing rolling body wear failure; (b) Bearing inner ring wear failure; (c) Bearing outer ring pitting failure; (d) Gear tooth breakage failure; (e) Missing gear teeth failure; (f) Gear tooth wear failure; (g) Gear tooth root crack failure.

Table 1.

The information on the compound fault status.

Compound Fault (A-B) Outer-Chipped Tooth Ball-Missing Tooth Inner-Cracked Tooth Ball-Worn Tooth
Single fault types Outer ring pitting,
chipped tooth
Worn ball,
missing tooth
Worn inner ring,
cracked tooth
Worn ball,
worn tooth
Fault information Outer ring pitting: moderate pitting, with a pitting diameter of 1 mm, located in the center of the raceway.
Chipped tooth: A gear tooth is missing a quarter, i.e., half in each direction of tooth height and tooth width.
Worn ball:
moderate wear, the wear area is about 3 mm long by 1 mm wide irregular shape.
Missing tooth: a wheel tooth is completely broken, i.e., one tooth is missing.
Worn inner ring: Moderate wear, the wear area is an irregular shape about 3 mm long and 1 mm wide, located in the center of the raceway.
cracked tooth: Crack depth 1.5 mm
Worn ball: moderate wear, the wear area is about 3 mm long by 1 mm wide irregular shape.
worn tooth: Moderate grinding of all gear teeth
Fault location Faulty bearing in position 6; The faulty gear is located in the first-stage driven wheel The faulty gear is a first-class driven wheel. Faulty bearing in position 6; The faulty gear is located in the primary active wheel. Faulty bearing in position 6; The faulty gear is located in the secondary driven wheel. Faulty bearing in position 6; The faulty gear is located in the secondary active wheel.

The gearbox vibration signal was collected by an acceleration sensor placed on the secondary parallel shaft gearbox, with a sampling frequency of 20,480 Hz and a sampling time of 8 s. For each gearbox state, the data sample contained 163,840 points. Two thousand and forty-eight points were selected as the sampling window, with five hundred overlapping points between adjacent samples. Nine hundred samples were obtained for each state sampling, and their time domain plots are shown in Figure 11. There were 10,800 samples of 12 gearbox states, and the sample details and labels are shown in Table 2. The sample dataset was randomly divided into the training, validation, and test sets according to the ratio of 8:1:1, where the training set contained 8640 samples, the validation set included 1080 samples, and the test set had 1080 samples. The training set was used to train the network and update the network parameters. The validation set was used to evaluate the model after each training round to check whether the model was overfitted on the training set. The test set was then used to assess the fault classification performance of the final model.

Figure 11.

Figure 11

Time domain diagram of the vibration signal.

Table 2.

Experimental samples and labels.

Work Condition Length of Sample Number of Samples Label Vector
Normal 2048 900 [1, 0, 0, 0, 0, 0, 0, 0]
Ball 2048 900 [0, 1, 0, 0, 0, 0, 0, 0]
Inner 2048 900 [0, 0, 1, 0, 0, 0, 0, 0]
Outer 2048 900 [0, 0, 0, 1, 0, 0, 0, 0]
Chipped tooth 2048 900 [0, 0, 0, 0, 1, 0, 0, 0]
Missing tooth 2048 900 [0, 0, 0, 0, 0, 1, 0, 0]
Worn tooth 2048 900 [0, 0, 0, 0, 0, 0, 1, 0]
Cracked tooth 2048 900 [0, 0, 0, 0, 0, 0, 0, 1]
Outer with chipped tooth 2048 900 [0, 0, 0, 1, 1, 0, 0, 0]
Ball with missing tooth 2048 900 [0, 1, 0, 0, 0, 1, 0, 0]
Inner with cracked tooth 2048 900 [0, 0, 1, 0, 0, 0, 0, 1]
Ball with worn tooth 2048 900 [0, 1, 0, 0, 0, 0, 1, 0]

4.1.2. Experimental Results and Analysis

To verify the effectiveness of the improved attention module CSAM proposed in this paper, the CSAM-MSCNN was compared with the CBAM-MSCNN, convolutional neural network (CNN) [30], fully-connected neural network (FNN) [30], support vector machine (SVM) [30], random forest (RF) [30], and wavelet transform and multilabel convolutional neural network (WT-MLCNN) [30]. To ensure the reliability of the comparison experiments, the network parameters of the MSCNN, the primary network structure of the CSAM-MSCNN and CBAM-MSCNN, are shown in Table 3, and the position of the attention module in the network is shown in Figure 6. The batch size was 32, the maximum number of training rounds was 100, the gradient descent optimization algorithm was chosen as Adam, the learning rate was 0.001, and the learning rate decay coefficient was 0.0001. Ten parallel experiments were performed considering the effect of random initialization parameters of the deep learning network.

Table 3.

Detailed network parameters of the CSAM-MSCNN.

Block Layer Name Kernel Size/Stride/Channel Output Shape (C, L)
\ Input layer \ 1 × 2048
\ Adaptive average pool \ 1 × 512
Convolutional block1-1 Conv1d 16/4/16 16 × 128
Max pool 2/2/16 16 × 64
Convolutional block1-2 Conv1d 3/1/32 32 × 64
Max pool 2/2/32 32 × 32
Convolutional block1-3 Conv1d 3/1/64 64 × 32
Max pool 2/2/64 64 × 16
Convolutional block1-4 Conv1d 3/1/64 64 × 16
Max pool 2/2/64 64 × 8
Convolutional block2-1 Conv1d 64/16/16 16 × 128
Max pool 2/2/16 16 × 64
Convolutional block2-2 Conv1d 3/1/32 32 × 64
Max pool 2/2/32 32 × 32
Convolutional block2-3 Conv1d 3/1/64 64 × 32
Max pool 2/2/64 64 × 16
Convolutional block2-4 Conv1d 3/1/64 64 × 16
Max pool 2/2/64 64 × 8
\ Adaptive max pool \ 1 × 512
Convolutional block3-1 Conv1d 16/4/16 16 × 128
Max pool 2/2/16 16 × 64
Convolutional block3-2 Conv1d 3/1/32 32 × 64
Max pool 2/2/32 32 × 32
Convolutional block3-3 Conv1d 3/1/64 64 × 32
Max pool 2/2/64 64 × 16
Convolutional block3-4 Conv1d 3/1/64 64 × 16
Max pool 2/2/64 64 × 8
\ Feature fusion \ 192 × 8
CSAM \ \ 192 × 8
\ Global average pool \ 192 × 1
Classification block Flatten \ 192
Linear \ 64
Linear \ 8

Table 4 records the average classification accuracy and standard deviation of each failure mode in the test sample of the method in this paper. This paper showed excellent performance in gearbox compound faults, and the classification accuracy of all fault types was higher than 99%. The average accuracy of the overall test set was 99.61%, with a standard deviation of only 0.13%, indicating that the CSAM-MSCNN has good stability for compound fault diagnosis. Comparing the performance of the CSAM-MSCNN with the rest of the models in Table 5, it is easy to see that the classification accuracy of the proposed method in this paper is more competitive.

Table 4.

Results of ten parallel experiments of CSAM-MSCNN.

Gearbox Status 1 2 3 4 5 6 7 8 9 10 Accuracy ± Std (%)
Normal 100 96.51 100 100 100 100 100 100 100 98.78 99.52 ± 1.07
Ball 100 100 100 100 100 100 98.84 100 100 100 99.88 ± 0.34
Inner 100 100 100 98.90 100 97.7 100 98.96 100 100 99.55 ± 0.74
Outer 100 100 100 100 97.62 98.8 100 98.97 98.89 100 99.42 ± 0.78
Chipped tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Missing tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Worn tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Cracked tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Outer with chipped tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Ball with missing tooth 98.90 100 97.96 100 95.06 100 100 100 100 100 99.19 ± 1.52
Inner with cracked tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Ball with worn tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Testing set 99.90 99.72 99.81 99.9 99.44 99.72 99.90 99.81 99.90 99.81 99.79 ± 0.13
Table 5.

Mean test accuracy and standard deviation of the proposed method versus other methods.

Methods Classification Accuracy ± Std (%)
CNN SC 76.50 ± 1.51
FNN SC 21.58 ± 4.81
SVM SC 24.29 ± 2.73
RF SC 44.62 ± 1.83
WT-MLCNN MLC 95.48 ± 0.81
CBAM-MSCNN MLC 98.62 ± 0.38
CSAM-MSCNN MLC 99.79 ± 0.13

4.1.3. Supplementary Experiments

To supplement the effectiveness of the proposed method, simulation data were used to test the diagnostic performance of the proposed model. The structure of the gearbox simulation model is shown in Figure 9. The simulated gearbox was set to operate at 0 loads and 1500 rpm, and the information on the dataset used for the experiment is shown in Table 6. The data had four gearbox states, including the normal state, two single-fault states, and the compound fault state coupled by a single fault, and the vibration signal time domain diagram is shown in Figure 12. The original vibration data were sampled by sliding windows with 2048 and 500 window shift lengths. Five hundred samples were sampled for each state to form the sample dataset. The dataset was divided into the training set, validation set, and test set according to the ratio of 8:1:1, i.e., the training set contained 1600 samples, and the validation set and test set had 200 samples each.

Table 6.

Description of gearbox failure simulation dataset.

Work Condition Fault Information Length of
Sample
Number of
Samples
Label
Vector
Normal \ 2048 500 [1, 0, 0]
Cracked tooth The gear is a first-class passive wheel, and the crack depth is 2 mm. 2048 500 [0, 1, 0]
Pitting tooth The gear is a first-class active wheel with severe pitting. 2048 500 [0, 0, 1]
Cracked tooth-
pitting tooth
Compound failure of tooth root crack and tooth surface pitting 2048 500 [0, 1, 1]
Figure 12.

Figure 12

Time domain diagram of the simulated signal.

The output layer dimension of the network was changed to three, and the rest of the model parameters are shown in Table 3. The maximum number of training rounds was 20. The rest of the hyperparameter settings were the same as in Experiment 1. The training process of the CSAM-MSCNN is shown in Figure 13, and the network converged after five training rounds. Since depth model parameters are affected by random initialization, ten parallel experiments were conducted to calculate the average accuracy and standard deviation. The results are shown in Table 7. Through 10 parallel experiments, the average accuracy of the CSAM-MSCNN on the test set was 99.46%, with a standard deviation of 0.67%. To visualize the fault classification effect on the test set, t-distributed stochastic neighbor embedding (t-SNE) was used to map the output layer features of the network to a two-dimensional plane, as shown in Figure 14. It can be seen that the CSAM-MSCNN can easily distinguish the four gearbox operating states. The results of this experiment show that the CSAM-MSCNN can perform the diagnosis task well for both the actual working condition data and the simulation data, which further verifies the effectiveness of the proposed method.

Figure 13.

Figure 13

Simulation experimental training process of CSAM-MSCNN: (a) Loss function curve; (b) accuracy curve.

Table 7.

Parallel experimental results of simulation data for CSAM-MSCNN.

Gearbox Status 1 2 3 4 5 6 7 8 9 10 Accuracy ± Std (%)
Normal 100 100 100 100 100 100 100 100 100 100 100 ± 0
Cracked tooth 100 100 96.61 100 100 98 100 100 100 100 99.88 ± 1.12
pitting tooth 100 100 100 100 100 100 100 100 100 100 100 ± 0
Cracked tooth with pitting tooth 92.86 100 100 100 93.10 100 100 98 100 100 98.39 ± 2.77
Testing set 98.21 100 99.15 100 98.27 99.5 100 99.5 100 100 99.46 ± 0.67
Figure 14.

Figure 14

Visualization of test set diagnostic results. Labels 0, 1, 2, 3 represent normal, cracked tooth, pitting tooth, and gear crack-pitting compound failure, respectively.

4.2. Experiment 2: 2009 PHM Gearbox Fault Diagnosis

4.2.1. Description of Experimental Data

The data for this experiment originated from the 2009 PHM Challenge, which are representative of general industrial gearbox data. The gearbox contained three shafts: the input shaft; the idler shaft; and the output shaft. It also included six bearings and four gears, divided into spur and helical gears. In this paper, the data of the spiral gear set was selected, and only the first channel of the vibration signal was used. A sketch of the gearbox structure is shown in Figure 15, with 16 and 40 teeth on the input and output shafts, respectively, and 48 and 24 teeth on the two gears of the idler shaft.

Figure 15.

Figure 15

Schematic of the gearbox used in the PHM 2009 Challenge Data [32].

The experimental dataset was collected from the gearbox at low load and 30, 40, and 50 Hz speed and contained six different health states. The six states of the gearbox are shown in Table 8, having eight independent states of the normal, broken tooth, missing tooth, bearing compound fault, bearing inner ring fault, bearing outer ring fault, shaft bending, and shaft unbalance, so the number of neurons in the final output layer of the CSAM-MSCNN was 8. The sampling frequency was 66.7 kHz, the sampling time was 4 s, the selected sample window length was 2048, and the window moving step was 500. Six hundred samples were collected for each health state at 30, 40, and 50 Hz speed as the network dataset, so the total dataset contained 10,800 samples. The dataset was randomly divided into the training set, validation set, and test set in the ratio of 8:1:1, and the samples and labels are described in detail, as shown in Table 9.

Table 8.

Gearbox Health Status Description.

Case Gear Bearing Shaft
16T 48T 24T 40T IS:IS ID:IS OS:IS IS:OS ID:OS OS:OS Input Output
Case1 Good Good Good Good Good Good Good Good Good Good Good Good
Case2 Good Good Chipped Good Good Good Good Good Good Good Good Good
Case3 Good Good Broken Good Good Good Good Combination Inner Good Bent Shaft Good
Case4 Good Good Good Good Good Good Good Combination Ball Good Imbalance Good
Case5 Good Good Broken Good Good Good Good Good Inner Good Good Good
Case6 Good Good Good Good Good Good Good Good Good Good Bent Shaft Good
Table 9.

Sample and Label Instructions.

Work Condition Length of Sample Number of Samples Label Vector
Case 1 2048 1800 [1, 0, 0, 0, 0, 0, 0, 0]
Case 2 2048 1800 [0, 1, 0, 0, 0, 0, 0, 0]
Case 3 2048 1800 [0, 0, 1, 1, 1, 0, 1, 0]
Case 4 2048 1800 [0, 0, 0, 1, 0, 1, 0, 1]
Case 5 2048 1800 [0, 0, 1, 0, 1, 0, 0, 0]
Case 6 2048 1800 [0, 0, 0, 0, 0, 0, 1, 0]

4.2.2. Experimental Results and Analysis

To further show the superiority of the method in this paper for composite fault diagnosis, it was compared with the CBAM-MSCNN, CNN, FNN, SVM, RF, and WT-MLCNN. The detailed network parameters of the CSAM-MSCNN are shown in Table 3. Similarly, the network parameters of the CBAM-MSCNN remained the same as in the CSAM-MSCNN, except for the attention module. The batch size was 32, the maximum number of training rounds was 200, the gradient descent optimization algorithm was chosen as Adam, the learning rate was 0.001, and the learning rate decay coefficient was 0.0001. The training process of the CSAM-MSCNN is shown in Figure 16.

Figure 16.

Figure 16

The training process of the CSAM-MSCNN: (a) Loss function curve; (b) accuracy curve.

Ten parallel experiments were conducted to avoid specificity and chance, and the experimental results are shown in Table 10 and Table 11. The results when comparing the method in this paper with the CBAM-MSCNN and the rest of the models are shown in Figure 17.

Table 10.

Parallel experimental results of the CSAM-MSCNN.

Gearbox Status 1 2 3 4 5 6 7 8 9 10 Accuracy ± Std (%)
Case 1 100 100 100 100 100 100 100 100 99.40 100 99.94 ± 0.18
Case 2 100 98.94 99.47 100 99.47 100 99.47 97.35 99.47 99.47 99.36 ± 0.74
Case 3 100 99.41 100 100 100 100 100 100 99.41 100 99.88 ± 0.23
Case 4 100 100 100 100 100 100 100 100 100 100 100 ± 0
Case 5 100 100 100 99.02 99.51 99.51 99.51 99.51 100 99.02 99.61 ± 0.36
Case 6 98.84 98.84 97.11 99.42 98.84 98.84 99.84 99.42 100 98.84 98.99 ± 0.75
Testing set 99.81 99.54 99.44 99.72 99.63 99.72 99.63 99.35 99.72 99.54 99.61 ± 0.13
Table 11.

Parallel experimental results of the CBAM-MSCNN.

Gearbox Status 1 2 3 4 5 6 7 8 9 10 Accuracy ± Std (%)
Case 1 100 100 100 98.81 100 100 98.81 97.02 99.4 100 99.40 ± 0.92
Case 2 97.35 98.94 97.88 100 100 98.41 96.3 98.41 99.47 94.71 98.14 ± 1.59
Case 3 100 100 100 100 100 100 100 100 99.41 100 99.94 ± 0.17
Case 4 100 100 100 100 100 100 100 99.44 100 100 99.94 ± 0.16
Case 5 100 100 99.51 100 99.51 100 100 98.53 100 100 99.75 ± 0.16
Case 6 93.06 98.27 98.27 97.11 98.27 98.84 99.42 100 99.42 98.84 98.15 ± 1.86
Testing set 98.43 99.54 99.26 99.35 99.63 99.54 99.07 98.89 99.63 98.89 99.22 ± 0.37
Figure 17.

Figure 17

Comparison of accuracy and standard deviation between the CSAM-MSCNN and other models.

To demonstrate the advantages of the proposed CSAM-MSCNN more intuitively, feature visualization was implemented using t-SNE. The features extracted at the first scale, the features extracted at the second scale, the features extracted at the third scale, the features obtained by feature fusion, the features corrected by the attention module, and the last fully connected layer features of the output layer were mapped to two-dimensional space, in the model. The output features of each network layer are shown in Figure 18.

Figure 18.

Figure 18

t-SNE visualization of network layer features: (a) Features extracted at the first scale; (b) Features extracted at the second scale; (c) Features extracted at the third scale; (d) Features obtained from feature fusion; (e) Features after CSAM correction; (f) The last layer features of the output layer.

Analysis of the above experimental results led to the following:

(1) In Figure 16, after 200 training iterations, the loss function and accuracy curves of the training and validation set finally converged. This indicates that CSAM-MSCNN can thoroughly explore the composite fault-related information in the gearbox vibration signal and achieve accurate fault classification.

(2) Comparing the results of the 10 parallel experiments of the CSAM-MSCNN and CBAM-MSCNN in Table 10 and Table 11. The average test accuracy of the CSAM-MSCNN was 99.61%, with a standard deviation of 0.13. In comparison, the average test accuracy of the CBAM-MSCNN was 99.22% with a standard deviation of 0.37%, which indicates that the improvement in the CSAM is effective. The main reason is that the CSAM uses a shared one-dimensional convolution instead of the shared fully connected layer in CBAM. This reduces the model parameters, solves the side effects generated by the original bottleneck structure, and improves the model’s robustness.

(3) The advantages of the proposed method in terms of diagnostic accuracy and stability can be visualized in Figure 17. Notably, traditional models such as CNN, FNN, SVM, and RF use single-label classifiers and perform poorly when diagnosing raw vibration data. The proposed method used a multilabel classifier to diagnose end-to-end faults using primary vibration data, which can decouple compound faults into multiple single fault modes while obtaining high diagnostic accuracy. The fault diagnosis task in this paper was more complex than traditional fault diagnosis tasks because we were diagnosing multiple faults simultaneously rather than treating them as a new failure mode. In contrast, compared with WT-MLCNN and CBAM-MSCNN, which also use multilabel classification, the CSAM-MSCNN showed more competitive diagnostic accuracy and standard deviation.

(4) As seen from the plots (a), (b), and (c) in Figure 18, the clustering effect of the proposed features in the CSAM-MSCNN varied among the three scales, which indicates that the use of multiscale network extraction results can enrich the fault information learned by the network. Observing plot (d), it can be found that performing multiscale feature fusion can yield better clustering results than single-scale feature extraction. This validates that fusing features from the three scales is effective and can complement each other’s feature information that needs to be focused on each scale. By comparing (d) and (c) plots, we found that the feature maps corrected by the CSAM were better clustered than those obtained by feature fusion. This indicates that the CSAM can effectively deal with the feature redundancy problem brought by multiscale feature fusion, focus the network attention on critical fault information, highlight the fault features that play a crucial role in diagnostic decisions, and suppress the features that are meaningless for fault classification or even lead to classification errors. The final (f) plot shows that the model can achieve good fault signal classification and that the CSAM-MSCNN can achieve end-to-end fault diagnosis using raw time-domain vibration signals.

5. Conclusions

This paper proposed a new gearbox composite fault diagnosis method, CSAM-MSCNN. The technique combined an improved channel–space attention module, a multiscale network structure, and a multilabel classifier, which can take full advantage of the feature extraction capability of the multiscale system, the feature enhancement capability of the attention mechanism, and the edges of the multilabel classifier in handling compound faults. The proposed method’s validity was verified using gearbox data collected in the laboratory and the data from the 2009 PHM Challenge. In Experiment 1, the average accuracy of the CSAM-MSCNN was 99.79%, with a standard deviation of only 0.13%. In Experiment 2, the average accuracy of the CSAM-MSCNN was 99.61%, with a standard deviation of only 0.13%. In two experiments, the proposed method in this paper effectively diagnosed compound faults in gearboxes and showed a more competitive fault diagnosis performance than other models. The method proposed in this paper is mainly used for severe gearbox failures. The intention is to apply the method to early faint fault diagnosis in future research.

Author Contributions

Conceptualization, Q.X. and H.J.; methodology, Q.X. and H.J.; formal analysis, X.Z.; writing—original draft preparation, Q.X.; writing—review and editing, J.L. and L.C.; funding acquisition, X.Z. and H.J. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationship that could have appeared to influence the work reported in this paper.

Funding Statement

This research was funded by the National Natural Science Foundation of China (grant number 52265016) and Xinjiang agricultural machinery R & D, manufacturing, promotion and application of integrated projects (grant number YTHSD2022-07).

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Zhao Z., Wu J., Li T., Sun C., Yan R., Chen X. Challenges and opportunities of AI-enabled monitoring, diagnosis & prognosis: A review. Chin. J. Mech. Eng. 2021;34:56. [Google Scholar]
  • 2.Zhou J., Qin Y., Luo J., Zhu T. Remaining useful life prediction by distribution contact ratio health indicator and consolidated memory GRU. IEEE Trans. Ind. Inform. 2022 doi: 10.1109/TII.2022.3218665. [DOI] [Google Scholar]
  • 3.Hajnayeb A., Ghasemloonia A., Khadem S., Moradi M.H. Application and comparison of an ANN-based feature selection method and the genetic algorithm in gearbox fault diagnosis. Expert Syst. Appl. 2011;38:10205–10209. doi: 10.1016/j.eswa.2011.02.065. [DOI] [Google Scholar]
  • 4.Bordoloi D., Tiwari R. Support vector machine based optimization of multi-fault classification of gears with evolutionary algorithms from time–frequency vibration data. Measurement. 2014;55:1–14. doi: 10.1016/j.measurement.2014.04.024. [DOI] [Google Scholar]
  • 5.Hu Q., Si X.-S., Zhang Q.-H., Qin A.-S. A rotating machinery fault diagnosis method based on multi-scale dimensionless indicators and random forests. Mech. Syst. Signal Process. 2020;139:106609. doi: 10.1016/j.ymssp.2019.106609. [DOI] [Google Scholar]
  • 6.Kang M., Kim J., Wills L.M., Kim J.-M. Time-varying and multiresolution envelope analysis and discriminative feature analysis for bearing fault diagnosis. IEEE Trans. Ind. Electron. 2015;62:7749–7761. doi: 10.1109/TIE.2015.2460242. [DOI] [Google Scholar]
  • 7.Qin Y., Mao Y., Tang B., Wang Y., Chen H. M-band flexible wavelet transform and its application to the fault diagnosis of planetary gear transmission systems. Mech. Syst. Signal Process. 2019;134:106298. doi: 10.1016/j.ymssp.2019.106298. [DOI] [Google Scholar]
  • 8.Lv Y., Yuan R., Song G. Multivariate empirical mode decomposition and its application to fault diagnosis of rolling bearing. Mech. Syst. Signal Process. 2016;81:219–234. doi: 10.1016/j.ymssp.2016.03.010. [DOI] [Google Scholar]
  • 9.Merainani B., Rahmoune C., Benazzouz D., Ould-Bouamama B. A novel gearbox fault feature extraction and classification using Hilbert empirical wavelet transform, singular value decomposition, and SOM neural network. J. Vib. Control. 2018;24:2512–2531. doi: 10.1177/1077546316688991. [DOI] [Google Scholar]
  • 10.Li Z., Yan X., Tian Z., Yuan C., Peng Z., Li L. Blind vibration component separation and nonlinear feature extraction applied to the nonstationary vibration signals for the gearbox multi-fault diagnosis. Measurement. 2013;46:259–271. doi: 10.1016/j.measurement.2012.06.013. [DOI] [Google Scholar]
  • 11.Praveenkumar T., Sabhrish B., Saimurugan M., Ramachandran K. Pattern recognition based on-line vibration monitoring system for fault diagnosis of automobile gearbox. Measurement. 2018;114:233–242. doi: 10.1016/j.measurement.2017.09.041. [DOI] [Google Scholar]
  • 12.Qian Q., Qin Y., Wang Y., Liu F. A new deep transfer learning network based on convolutional auto-encoder for mechanical fault diagnosis. Measurement. 2021;178:109352. doi: 10.1016/j.measurement.2021.109352. [DOI] [Google Scholar]
  • 13.Xu Z., Li C., Yang Y. Fault diagnosis of rolling bearings using an improved multi-scale convolutional neural network with feature attention mechanism. ISA Trans. 2021;110:379–393. doi: 10.1016/j.isatra.2020.10.054. [DOI] [PubMed] [Google Scholar]
  • 14.Zhang W., Peng G., Li C., Chen Y., Zhang Z. A new deep learning model for fault diagnosis with good anti-noise and domain adaptation ability on raw vibration signals. Sensors. 2017;17:425. doi: 10.3390/s17020425. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Xia M., Mao Z., Zhang R., Jiang B., Wei M. A new compound fault diagnosis method for gearbox based on convolutional neural network; Proceedings of the 2020 IEEE 9th Data Driven Control and Learning Systems Conference (DDCLS); Liuzhou, China. 20–22 November 2020; Piscataway, NJ, USA: IEEE; 2020. pp. 1077–1083. [Google Scholar]
  • 16.Chen X., Zhang B., Gao D. Bearing fault diagnosis base on multi-scale CNN and LSTM model. J. Intell. Manuf. 2021;32:971–987. doi: 10.1007/s10845-020-01600-2. [DOI] [Google Scholar]
  • 17.Jiang G., He H., Yan J., Xie P. Multiscale convolutional neural networks for fault diagnosis of wind turbine gearbox. IEEE Trans. Ind. Electron. 2018;66:3196–3207. doi: 10.1109/TIE.2018.2844805. [DOI] [Google Scholar]
  • 18.Ravikumar K., Yadav A., Kumar H., Gangadharan K., Narasimhadhan A. Gearbox fault diagnosis based on Multi-Scale deep residual learning and stacked LSTM model. Measurement. 2021;186:110099. doi: 10.1016/j.measurement.2021.110099. [DOI] [Google Scholar]
  • 19.Hao Y., Wang H., Liu Z., Han H. Multi-scale CNN based on attention mechanism for rolling bearing fault diagnosis; Proceedings of the 2020 Asia-Pacific International Symposium on Advanced Reliability and Maintenance Modeling (APARM); Vancouver, BC, Canada. 20–23 August 2020; Piscataway, NJ, USA: IEEE; 2020. pp. 1–5. [Google Scholar]
  • 20.Zhou J., Qin Y., Luo J., Wang S., Zhu T. Dual-thread gated recurrent unit for gear remaining useful life prediction. IEEE Trans. Ind. Inform. 2022 doi: 10.1109/TII.2022.3217758. [DOI] [Google Scholar]
  • 21.Woo S., Park J., Lee J.-Y., Kweon I.S. Cbam: Convolutional block attention module; Proceedings of the European Conference on Computer Vision (ECCV); Munich, Germany. 8–14 September 2018; pp. 3–19. [Google Scholar]
  • 22.Jang G.-B., Cho S.-B. Feature space transformation for fault diagnosis of rotating machinery under different working conditions. Sensors. 2021;21:1417. doi: 10.3390/s21041417. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Plakias S., Boutalis Y.S. Fault detection and identification of rolling element bearings with Attentive Dense CNN. Neurocomputing. 2020;405:208–217. doi: 10.1016/j.neucom.2020.04.143. [DOI] [Google Scholar]
  • 24.Li X., Wan S., Liu S., Zhang Y., Hong J., Wang D. Bearing fault diagnosis method based on attention mechanism and multilayer fusion network. ISA Trans. 2022;128:550–564. doi: 10.1016/j.isatra.2021.11.020. [DOI] [PubMed] [Google Scholar]
  • 25.Xie Y., Niu T., Shao S., Zhao Y., Cheng Y. Attention-based Convolutional Neural Networks for Diesel Fuel System Fault Diagnosis; Proceedings of the 2020 International Conference on Sensing, Measurement & Data Analytics in the Era of Artificial Intelligence (ICSMD); Xi’an, China. 15–17 October 2020; Piscataway, NJ, USA: IEEE; 2020. pp. 210–214. [Google Scholar]
  • 26.Huang T., Fu S., Feng H., Kuang J. Bearing fault diagnosis based on shallow multi-scale convolutional neural network with attention. Energies. 2019;12:3937. doi: 10.3390/en12203937. [DOI] [Google Scholar]
  • 27.Ye Z., Yu J. AKSNet: A novel convolutional neural network with adaptive kernel width and sparse regularization for machinery fault diagnosis. J. Manuf. Syst. 2021;59:467–480. doi: 10.1016/j.jmsy.2021.03.022. [DOI] [Google Scholar]
  • 28.Lv H., Chen J., Pan T., Zhang T., Feng Y., Liu S. Attention mechanism in intelligent fault diagnosis of machinery: A review of technique and application. Measurement. 2022;199:111594. doi: 10.1016/j.measurement.2022.111594. [DOI] [Google Scholar]
  • 29.Huang R., Liao Y., Zhang S., Li W. Deep decoupling convolutional neural network for intelligent compound fault diagnosis. IEEE Access. 2018;7:1848–1858. doi: 10.1109/ACCESS.2018.2886343. [DOI] [Google Scholar]
  • 30.Liang P., Deng C., Wu J., Yang Z., Zhu J., Zhang Z. Compound fault diagnosis of gearboxes via multi-label convolutional neural network and wavelet transform. Comput. Ind. 2019;113:103132. doi: 10.1016/j.compind.2019.103132. [DOI] [Google Scholar]
  • 31.Wang Q., Wu B., Zhu P., Li P., Zuo W., Hu Q. ECA-Net: Efficient channel attention for deep convolutional neural networks; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Seattle, WA, USA. 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  • 32.Phm Data Challenge 2009. [(accessed on 10 April 2009)]. Available online: https://phmsociety.org/data-analysis-competition.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES