Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2023 Mar 14;23(6):3108. doi: 10.3390/s23063108

Detection of Corona Faults in Switchgear by Using 1D-CNN, LSTM, and 1D-CNN-LSTM Methods

Yaseen Ahmed Mohammed Alsumaidaee 1, Chong Tak Yaw 2, Siaw Paw Koh 2,3,*, Sieh Kiong Tiong 2,3, Chai Phing Chen 3, Talal Yusaf 4, Ahmed N Abdalla 5, Kharudin Ali 6, Avinash Ashwin Raj 7
Editor: Giovanni Betta
PMCID: PMC10059847  PMID: 36991819

Abstract

The damaging effects of corona faults have made them a major concern in metal-clad switchgear, requiring extreme caution during operation. Corona faults are also the primary cause of flashovers in medium-voltage metal-clad electrical equipment. The root cause of this issue is an electrical breakdown of the air due to electrical stress and poor air quality within the switchgear. Without proper preventative measures, a flashover can occur, resulting in serious harm to workers and equipment. As a result, detecting corona faults in switchgear and preventing electrical stress buildup in switches is critical. Recent years have seen the successful use of Deep Learning (DL) applications for corona and non-corona detection, owing to their autonomous feature learning capability. This paper systematically analyzes three deep learning techniques, namely 1D-CNN, LSTM, and 1D-CNN-LSTM hybrid models, to identify the most effective model for detecting corona faults. The hybrid 1D-CNN-LSTM model is deemed the best due to its high accuracy in both the time and frequency domains. This model analyzes the sound waves generated in switchgear to detect faults. The study examines model performance in both the time and frequency domains. In the time domain analysis (TDA), 1D-CNN achieved success rates of 98%, 98.4%, and 93.9%, while LSTM obtained success rates of 97.3%, 98.4%, and 92.4%. The most suitable model, the 1D-CNN-LSTM, achieved success rates of 99.3%, 98.4%, and 98.4% in differentiating corona and non-corona cases during training, validation, and testing. In the frequency domain analysis (FDA), 1D-CNN achieved success rates of 100%, 95.8%, and 95.8%, while LSTM obtained success rates of 100%, 100%, and 100%. The 1D-CNN-LSTM model achieved a 100%, 100%, and 100% success rate during training, validation, and testing. Hence, the developed algorithms achieved high performance in identifying corona faults in switchgear, particularly the 1D-CNN-LSTM model due to its accuracy in detecting corona faults in both the time and frequency domains.

Keywords: corona discharge, energy, switchgear, 1D-CNN-LSTM, faults

1. Introduction

Switchgear plays a crucial role in power distribution networks [1]. The term “switching gear” refers to a variety of switching devices that perform functions such as controlling, protecting, and isolating power systems. This definition also encompasses devices used to regulate and meter power systems, as well as circuit breakers and other similar technologies [2]. Switchgear plays a critical role in ensuring the safety of downstream maintenance, including fault clearing, upkeep, equipment substitutions, and a range of other related concerns, especially in power-generating systems. It functions by isolating and de-energizing specific electrical components and networks to prevent potential hazards.

Furthermore, switchgear is typically classified into low, medium, and high voltage groups and is primarily constructed using insulating mediums such as air or oil [3]. It is crucial to closely monitor the performance and condition of switchgear during operation. Early detection of malfunctioning switchgear and prioritization of corrective maintenance are essential [4]. Failure to do so could result in negative outcomes for the distribution network, operational employees, and end users, leading to disruptive and unsettling fluctuations in customer data and regulatory perceptions [5].

Breakdowns in switchgear can be classified into several types, including arcing, tracking, surface discharge, and mechanical failure [6]. Corona discharge is a common issue that occurs in switchgear [7]. It is considered one of the primary causes of equipment breakdown, aging, and power loss in steady power systems [8]. In the event of an incident involving high-voltage systems, corona discharge faults can damage Medium-Voltage (MV) switchgear, which may result in equipment failure. MV switchgear is a crucial component of the power distribution network [9].

To maintain switchgear safety and ensure a stable power supply, it is crucial to detect and closely monitor corona discharge activities [10]. A corona discharge in switchgear results in the release of three types of energy: electromagnetic radiation, ultrasonic waves, and gases such as ozone and Nitrogen Oxides (NOx) [11]. Electromagnetic and ultrasonic techniques have been utilized to detect and monitor corona discharge activity in medium-voltage switchgear with reduced interference and improved efficiency [11]. In particular, electricity-based methods are commonly employed to measure discharge activities.

Offline detection procedures can be costly, and it can be challenging to identify the substances that promote corona discharge. To address this, research has investigated measuring ozone emissions post-discharge for fault identification and monitoring [12,13]. As a result, deep learning has gained significant traction and advancements in recent years and is now widely used across a wide range of industries, particularly in image and natural language analysis [14,15,16].

Convolutional neural networks (CNNs) are a popular type of deep learning model that can extract many features using various convolutional layers, including max pooling, normalization, and fully connected layers. [17]. However, CNNs are not suitable for remembering past time-series patterns. Recurrent neural networks (RNNs) are a specific type of neural network that can remember information from the past by using previous outputs as inputs [18,19].

Lately, several research studies have employed RNNs in the fields of speech recognition and natural language processing, among others. Specifically, the long short-term memory (LSTM) RNN architecture has been acknowledged and applied in time series processing. Its distinctive design explicitly resolves the problem of gradient vanishing in traditional RNNs and enables the learning of long-term dependencies, leading to more effective acquisition of the chronological aspects of sequential data [16,17,18,19,20,21].

The combination of CNN and RNN in a convolutional LSTM neural network is essential to model chronological sequences and improve deep neural network modeling capabilities. This network has been used in various large vocabulary tasks and has been shown to provide a relative improvement of 4–6% over an LSTM [22,23]. Additionally, there are many other studies focused on extracting temporal and spatial characteristics by combining CNN and LSTM models [24]. Table 1 presents the advantages and disadvantages of three methods that are commonly used in deep learning: 1D-CNN, LSTM, and 1D-CNN-LSTM. Each method has its own strengths and weaknesses, and researchers often choose one over the others depending on the specific task at hand. By comparing and contrasting these methods, researchers can determine which one is the best fit for their particular problem.

Table 1.

Summary of the advantages and disadvantages of the proposed methods.

Ref Methods Advantages Disadvantages
[25]
[26]
[27]
[28]
1D-CNN
  1. Can learn complex patterns in data through feature extraction;

  2. Effective for processing sequential data;

  3. Can handle high-dimensional inputs;

  4. Can be computationally efficient.

  1. Not ideal for handling variable-length inputs;

  2. May not perform as well as LSTM for tasks that require long-term memory.

[29]
[30]
[31]
[32]
LSTM
  1. Effective for processing sequential data and modeling long-term dependencies;

  2. Can handle variable-length inputs;

  3. Can retain and selectively forget information over long time intervals;

  4. Can be combined with other models, such as CNNs, for improved performance.

  1. Can be computationally expensive;

  2. May be overfit on small datasets.

[33]
[21]
[34]
[35]
1DCNN-LSTM
  1. Can combine the strengths of both 1D-CNN and LSTM models for improved performance;

  2. Effective for processing sequential data and modeling long-term dependencies;

  3. Can handle high-dimensional and variable-length inputs.

  1. Can be computationally expensive;

  2. May be overfit on small datasets;

  3. Overall, the choice of model will depend on the specific characteristics of the problem being solved and the available data.

The following points summarize the study’s motivations and contributions:

  1. This study uses deep learning techniques to detect corona faults in switchgear. The techniques employed include 1D-CNN, LSTM, and a hybrid model of 1D-CNN and LSTM;

  2. The purpose of the research is to determine the most effective method for detecting corona faults. The hybrid method (1D-CNN-LSTM) was found to be the most effective compared to other methods;

  3. The study is unique in that it was conducted in both the time and frequency domains, which had not previously been explored in this field;

  4. The proposed models are able to quickly identify and differentiate corona faults from other types of faults. Overall, the hybrid model is considered the best model for detecting corona faults in both the time and frequency domains.

The categorization of the article can be seen; thus, Section 2 comprises the theoretical study and methods. Section 3 comprises performance metrics, Section 4 comprises results and discussion, and Section 5 concludes the paper.

2. Theoretical Study and Methods

2.1. The Theory of Corona Discharge

A corona discharge refers to the phenomenon where a fluid, such as the air surrounding an electrically charged conductor, ionizes, and creates an electrical discharge. Without measures to reduce the electric field strength, high-voltage equipment may experience spontaneous corona discharges. A corona is formed when the electric field strength around a conductor is strong enough to create a conductive zone but not strong enough to cause electrical breakdown or arcing. Nevertheless, corona discharge can have detrimental effects on high-voltage equipment and cause a loss of power [2]. Figure 1a–c illustrates the corona discharge phenomenon and how the conductance and avalanche processes in Medium-Voltage (MV) switchgear are influenced by the electrode material.

Figure 1.

Figure 1

(ac) A corona discharge is a phenomenon that occurs in the air [2].

Figure 1 shows that when a neutral atom or molecule is exposed to a high electric potential gradient, it becomes ionized and initiates the corona discharge process. High-conductivity materials can lead to severe corona discharge flaws due to the strong electric field they produce, which attracts a lot of free electrons. Copper (Cu) and Stainless Steel (SS) are two popular metals used in MV switchgear due to their durability and conductivity. Copper has a higher conductivity than SS, with a value of 5.96107 Siemens per meter (S/m) compared to 1.45 × 106 (S/m) for SS. However, copper can produce higher corona discharges in MV switchgear than SS due to its higher conductivity.

Furthermore, when subjected to high-voltage stress, air, which contains electrically neutral molecules, acts as an insulator. When powerful energy travels as electrons from an electrode and strikes molecules, the molecules change [36].

The energy absorbed by the electrons surrounding atomic nuclei can lead to high-energy transformations, causing them to dissociate from the atoms or molecules they were attached to. This results in the formation of a positively charged ion and a dissociated free electron. Through the avalanche process, new electrons are produced as a result of electrons being struck by other electrons, leading to a mixture of electrons and ions [37], whereas the appearance of light signals from locally applied electric fields that permit sustained discharge into the atmosphere is another cause of corona discharge [38].

2.2. Proposed Methods

The proposed method for detecting corona faults involves collecting and preprocessing data in both the time and frequency domains. Figure 2 is used to illustrate the mechanism of data collection and preprocessing. The data is cleaned and collected, and then converted from sound to image using a mel-spectrogram. The data includes both corona and non-corona faults that occur in the switchgear, and it is classified into three phases: the training phase, the validation phase, and the testing phase. Features are then extracted using 1D-CNN, and classification is applied using the LSTM algorithm to achieve the highest possible accuracy. These methods are aimed at detecting corona faults in switchgear in both the time and frequency domains.

Figure 2.

Figure 2

General flowchart diagram of the research method.

2.2.1. Data Collection

The test findings received using airborne ultrasonic testing (AUT) equipment were converted into raw data. The waveform audio file format, also known as wav, or the Moving Picture Experts Group (MPEG), was used to record the ultrasonic data. A technique for data preprocessing that consolidates or transforms the data into a format appropriate for a particular machine learning algorithm will be introduced as a result of this. The data, which were in the wav and mp3 file formats, were transformed into a matrix format to meet the requirements of the MATLAB software.

2.2.2. Preprocessing

The preprocessing step involves several stages, such as converting the raw audio data in wav or mp3 format to a matrix format suitable for the MATLAB programming language. The next step is to concatenate the corona with non-corona data, which includes other fault types like arcing, tracking mechanical, and normal, to differentiate between corona and other faults. Once the data is prepared, it is classified into three phases: training, validation, and testing, with 70%, 15%, and 15% of the data, respectively. Additionally, the “Mel-Spectrogram” technique is used to convert the voice data into an image format.

2.2.3. D CNN

A one-dimensional convolutional neural network (1D-CNN) is a neural network that operates on one-dimensional data, such as time series or sequence data. In this study, 1D CNNs are used to extract representative properties from corona and non-corona faults in both the time and frequency domains. This is achieved through 1D convolution operations using filters. Figure 3 shows the CNN structure, which includes convolutional layers, filters, a MaxPooling1D layer, a FC layer, and a classification layer with a Rectified Linear Unit (RELU) as the activation function. Dropout and batch size are used to avoid overfitting. In Equations (1–3), these equations are stated. (RELU), which can cause nonlinearity, is used as a CNN activation function. While deep CNNs were initially used for image classification tasks, they have since been used for video analysis and recognition. However, the use of 1D sequence data for classification and prediction is relatively new [2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39]. The corona and non-corona categories can be seen as a sequential modeling effort, making 1D CNNs a suitable choice. Compressed 1D-CNNs are preferred for real-time applications due to their low processing requirements [40].

xo,fll=f(im xil1kio,fll+bl) (1)
xol=f[max(im xil1)+bl] (2)
x0l=f(xil1diol+bl) (3)

where x is a one-dimensional input matrix (n×1), f (.) is an activation function, kio,fll is a kernel filter (k×1),  lth is the number of filters F, the output of the lth convolutional layer is xo,fll. b, denoted as bias vectors, and d are learnable parameters.

Figure 3.

Figure 3

The basic structure of 1D-CNN.

2.2.4. RNN(LSTM) Structure

LSTMs (Long Short-Term Memory) are a type of RNN (Recurrent Neural Network) that overcome the vanishing gradient problem in traditional RNNs and are capable of capturing long-term dependencies in sequential data. In LSTMs, the hidden layers of RNNs are replaced with memory cells that are capable of selectively retaining or discarding information using gates. There are three types of gates in an LSTM: the input gate, the output gate, and the forget gate, which regulate the flow of information into and out of the memory cell [41]. For consecutive modeling applications like text categorization and time series modeling, LSTM is very useful [29] LSTM’s structural layout is drawn in Figure 4. The mathematical expression for an LSTM block presented by Equations (4)–(9) is as follows:

it=σ( xtUi+ht1Wi+bi) (4)
ft=σ( xtUf+ht1Wf+bf) (5)
Ot=σ( xtUo+ht1Wo+bo) (6)
Ct=σ( ftCt1+itCˇ tUi+bc) (7)
Cˇ=tanh( xtUg+ht1Wg) (8)
ht=tanh( Ct)Ot (9)

where xt is the network input matrix, ht is the output of the hidden layer, and σ indicates the SoftMax function. C is a memory cell that facilitates recalling a significant case, which is counted using Equation (7). The elderly cell case Ct1 is updated to the modern cell case Ct. New nominee values Cˇ and the output of the present LSTM block ht makes use of the hyperbolic tangent function as shown in Equations (8) and (9). The parameters for observation are weights (Wi, Wf, Wo, and Wg) and (Ui, Uf, Uo, and Ug) that create the neural network’s internal parameters and bias vectors (bi, bf, bo, and bc). The model improves weights and biases by reducing the goal function. The operator denotes the multiplication of elements.

Figure 4.

Figure 4

The basic structure of LSTM.

2.2.5. D-CNN-LSTM

After the feature extraction stage by 1D-CNN, the output is fed into an LSTM layer, which can capture the long-term dependencies of the input sequence. The LSTM layer helps in learning the temporal relationship between the features extracted by the 1D-CNN model. The output of the LSTM layer is then passed through a fully connected layer for classification into corona or non-corona faults in both domains. The proposed hybrid model benefits from the strengths of both 1D-CNN and LSTM and is expected to achieve better performance in identifying corona faults compared to using either model alone as shown in Figure 5.

Figure 5.

Figure 5

An instance of a 1D-CNN-LSTM model.

In the time domain analysis, the recording sampling comprises a multi-input, series-dimensional matrix. The second stage involves training and classifying corona and non-corona cases using the LSTM neural network. The Adam optimizer was used before the optimization process, which is a popular method for weight optimization. The model’s accuracy and losses have been studied by varying the number of convolutional layers, kernel size, MaxPooling1D, and Fully Connected (FC) layer as well as the model parameters, which have been optimized to improve efficiency. The samples of datasets for corona and non-corona in the time domain are in Table 2, which has a size of 17.5 megabytes (MB) in the time domain.

Table 2.

The datasets sample corona and non-corona in the time domain.

No Name of Samples No of Samples
1 Arcing faults 54 ∗ 20,001
2 Corona faults 41 ∗ 20,001
3 Tracking faults 313 ∗ 20,001
4 Mechanical faults 17 ∗ 20,001
5 Normal faults 13 ∗ 20,001

As a result, the 1D CNN-LSTM structure consists of three 1D convolutional layers and three global MaxPooling1D layers, with a 0.5% dropout for RELU to accelerate processing. MaxPooling1D is set to 2, and the kernel size is 3. Following these are two LSTM layers with 128 and 32 units, respectively, and a 0.5% dropout. The FC layer, with a SoftMax activation function, is shown below, and the learning rate is 0.0001 with a batch size of 16 and an epoch number of 100. Figure 6 depicts the CNN-LSTM architecture as designed.

Figure 6.

Figure 6

The suggested 1D CNN-LSTM model’s structure (time domain).

On the one hand, the 1D CNN architecture previously described consists of two one-dimensional convolutional layers, each with 16 and 32 filters, respectively. A drop-out rate of 0.4 is applied after each layer, followed by two max pooling layers with a size and kernel established at 3 and 2. The fully connected layers use the SoftMax activation function with a learning rate of 0.0001 and batch size of 16, and the number of epochs is 60. The LSTM architecture for corona and non-corona fault detection consisted of two connected LSTM units with 32 and 32 LSTM units, resulting in a significant increase in training time. As a result, a fully connected layer of 32 neurons with a SoftMax activation function and a learning rate of 0.0001 was constructed, with a batch size of 16 and a number of epochs of 50. The detailed structural design of the deep learning algorithms can be found in Table 3.

Table 3.

The structural design for deep learning algorithms in the time domain.

1D-CNN LSTM 1D-CNN-LSTM
1DConv (16) (RELU) LSTM (32) 1DConv (16) (RELU)
Drop-out Drop-out MaxPooling1D
MaxPooling1D LSTM (32) Drop-out
1DConv (32) (RELU) Drop-out 1DConv (32) (RELU)
Drop-out Flatten MaxPooling1D
MaxPooling1D Dense (32) (RELU) Drop-out
Flatten Dense (2) (SoftMax) 1DConv (256) (RELU)
Dense (32) (RELU) MaxPooling1D
Dense (2) (SoftMax) Drop-out
LSTM (128)
Drop-out
LSTM (32)
Dense (2) (SoftMax)
The number of total parameters
6738 16,226 245,170

In the case of frequency domain analysis, the suggested method involves recording samples as dimensional matrices in the frequency domain. The samples are then used to teach an LSTM neural network how to tell the difference between cases of corona and non-corona, and the same steps are taken in the time domain where the Adam optimizer is used to optimize the weights of the neural network during training. The number of convolutional layers, kernel size, MaxPooling1D, and FC layer are all model parameters that are investigated to achieve the best possible accuracy and loss. By tweaking these parameters, the model’s performance can be improved. The samples of datasets for corona and non-corona in the frequency domain are shown in Table 4, which has a size of 11.3 megabytes (MB) in the frequency domain.

Table 4.

The samples of datasets for corona and non-corona in the frequency domain.

No Name of Samples No of Samples
1 Arcing faults 53 ∗ 10,001
2 Corona faults 39 ∗ 10,001
3 Tracking faults 40 ∗ 10,001
4 Mechanical faults 16 ∗ 10,001
5 Normal faults 12 ∗ 10,001

The 1D-CNN-LSTM model structure consists of two global max pooling layers and two 1D convolution layers. To save processing time, a dropout of 0.3 is employed after each layer from the global max pooling layers, with RELU activation used for the CNN layers. The MaxPooling1D is set to 2, and the kernel size is equal to 3. After the two CNN layers, there are two LSTM layers with 128 and 32 units, respectively, followed by a dropout of 0.3. The subsequent fully connected (FC) layer has a learning rate of 0.0001, a batch size of 16, and 100 epochs with a SoftMax activation function. The structure of the 1D-CNN-LSTM is depicted in Figure 7.

Figure 7.

Figure 7

The suggested 1D CNN-LSTM model’s structure (frequency domain).

The 1D-CNN model architecture consists of three one-dimensional convolutional layers with 64, 128, and 256 filters for each layer. Three max pooling layers are also employed, with a dropout of 0.3 between each of them to prevent overfitting. The sizes of the kernel and MaxPooling1D are set to 3 and 2, respectively. Finally, a SoftMax activation function and learning rate of 0.0001 are used for the last layer of the network, as well as a batch size of 16 and a number of epochs of 60. Overall, this architecture appears to be a deep learning approach for classifying corona and non-corona cases based on time series data.

On the other side, the LSTM model consists of two connected LSTM units with 64 and 32 units, respectively, and a fully connected layer of 32 neurons. A fully connected layer with a SoftMax activation function, a learning rate of 0.0001, a batch size of 16, and a number of epochs of 50. The training time for this model was found to be significantly higher than the hybrid 1D-CNN-LSTM model. Table 5 provides detailed information on the designs of the deep learning algorithms used in this study. Overall, both the hybrid 1D-CNN model and the LSTM model and the hybrid 1D-CNN-LSTM model are deep learning approaches that are designed to classify corona and non-corona cases based on time series data. However, the specifics of the architectures differ, and their effectiveness and generalizability would depend on the specifics of the dataset and training process.

Table 5.

The structural design for deep learning algorithms in the frequency domain.

1D CNN LSTM 1D CNN-LSTM
1DConv (32) (RELU) LSTM (64) 1DConv (64) (RELU)
Drop-out Drop-out MaxPooling1D
MaxPooling1D LSTM (32) Drop-out
1DConv (64) (RELU) Drop-out 1DConv (128) (RELU)
Drop-out Flatten MaxPooling1D
MaxPooling1D Dense (32) (RELU) Drop-out
1DConv (1024) (RELU) Dense (2) (SoftMax) LSTM (128)
Drop-out Drop-out
MaxPooling1D LSTM (32)
Flatten Dense (2) (SoftMax)
Dense (32) (RELU)
Dense (2) (SoftMax)
The number of total parameters
143,650 35,298 197,250

3. Performance Metrics

Because of this, it is difficult to assess the effectiveness of the suggested models in Table 6 using the paper confusion matrix. The situations of corona and non-corona are represented, respectively, by 1 and 0. Equations (8)–(11) provide a number of indices, including classification accuracy, dependability, sensitivity, and F1-Measure (11). The confusion matrix makes use of them to evaluate the model’s performance. The recall is the portion of a pertinent example that is remembered out of all the pertinent samples. Precision is the portion of the recovered samples that are pertinent samples. As a separate performance indicator, the cross-entropy loss is used to assess how accurately the model predicts the target data as provided in Equation (12). The reduction of the difference between the corona and non-corona condition probability distributions is a crucial performance parameter. Most issues involving binary classification employ it. For a decent model, the loss value is 0.

Classification Accuracy (%)=100×TP+TNTP+FP+FN+TN , (10)
Recall (Sensitivity)(%)=100×TPTP+FN , (11)
Precision (Dependability)(%)=100×TPTP+FP , (12)
F1 measure (%)=100×2×(Precision×Recall) (Precision+Recall)  , (13)
Loss (L)=1NI=1M TIlog (xi). (14)

Table 6.

Confusion matrix for corona and non-corona cases.

Predictive Corona
Findings (1)
Predictive Non-Corona
Findings (0)
Actual corona findings (1) TP FP
Actual non-corona findings (0) FN TN

The right forecast is shown as True Positive (TP) and True Negative (TN), while the wrong forecast is expressed as False Positive (FP) and False Negative (FN).

In an instance in which Xi is the predicting response, Ti will be the target value, M will be the total number of responses predicting in X (all findings and classes combined), and N will be the cumulative number of observations in X. The binary cross entropy missed is totaled, inclusive of divisions and surveillance, and normalized by the number of states.

4. Results and Discussion

On the basis of the sound waves collected when the switchgear is in operation, experiments are carried out to illustrate how well the created algorithms function in locating corona faults in the switchgear. The tests are conducted in the time domain and the frequency domain, which are two distinct domains. For 1DCNN, LSTM, and 1D-CNN-LSTM, in order to distinguish between corona and non-corona faults in the time domain, a total of 438 data samples were created and split into training, validation, and testing datasets. The test dataset contained only positive or negative cases of corona. Seventy percent of the data was used for training, while fifteen percent was used for both validation and testing.

Each phase of the study consisted of 306 cases for testing, 66 cases for validation, and 66 cases for training. The confusion matrix presented in the tables below shows the true and false detection percentages based on the best results from repeated trials. Table 7 presents the results of the 1D-CNN-LSTM training phase in the time domain, using a total of 306 datasets. Among these, 26 cases of corona and 278 cases of non-corona were correctly identified. Only two corona cases were misclassified as non-corona, resulting in an accuracy of 99.3% and an error rate of 0.7%. On the other hand, the CNN accurately identified 274 non-corona cases and 26 corona cases, while misclassifying five corona cases as non-corona, resulting in an accuracy of 98.3% and an error rate of 1.7%. Similarly, the LSTM identified 26 corona cases and 272 non-corona cases, with eight corona cases misclassified as non-corona, resulting in an accuracy of 97.3% and an error rate of 2.7%.

Table 7.

Time domain corona fault classification output matrix for the training phase.

1D-CNN LSTM 1D-CNN-LSTM
Corona Non-Corona Corona Non-Corona Corona Non-Corona
Actual corona 26 5 26 8 26 2
Actual non-corona 0 275 0 272 0 278

During the validation stage of the time domain analysis for 1D-CNN-LSTM, 66 sets of data were used, including seven cases of corona and fifty-eight cases of non-corona. Unfortunately, one corona case was misclassified as a non-corona case, resulting in an overall accuracy of 98.4% with a 1.6% error rate. Both the CNN and LSTM models were employed during this analysis, and the results were similar. Seven corona cases and fifty-eight non-corona cases were successfully identified by both models, with one corona case being incorrectly classified as non-corona. The overall accuracy rate for both models was determined to be 98.4%, with a 1.6% error rate. Table 8 shows the detailed matrix results for this analysis.

Table 8.

Time domain corona fault classification output matrix for the validation phase.

1D-CNN LSTM 1D-CNN-LSTM
Corona Non-Corona Corona Non-Corona Corona Non-Corona
Actual corona 7 1 7 1 7 1
Actual non-corona 0 58 0 58 0 58

The results of the testing are shown in Table 9. According to the information provided, the algorithm has used a dataset of 66 sets of data for testing and has achieved an accuracy of 98.4% with a 1.6% error rate in identifying non-corona faults and corona faults using 1D-CNN-LSTM. Specifically, out of the 66 sets of data, the algorithm has correctly identified fifty-seven non-corona faults and eight corona cases but has mistakenly recognized one corona case as non-corona. Furthermore, the algorithm has also been tested using only CNN and LSTM, separately. When using CNN alone, the algorithm has achieved an accuracy of 93.9% with a 6.1% error rate, and when using LSTM alone, the algorithm has achieved an identical accuracy and error rate. In both cases, the algorithm has identified fifty-three non-corona faults and eight cases of corona, but has misidentified five cases of corona as non-corona, while for CNN, the algorithm has identified fifty-four non-corona faults and eight cases of corona, but has misidentified four cases of corona as non-corona.

Table 9.

Time domain corona fault classification output matrix for the testing phase.

1D-CNN LSTM 1D-CNN-LSTM
Corona Non-Corona Corona Non-Corona Corona Non-Corona
Actual corona 8 4 8 5 8 1
Actual non-corona 0 54 0 53 0 57

According to the information given in the frequency domain analysis, the dataset used in this analysis consisted of 160 samples, with 70% being used for training and 15% for both validation and testing. The training dataset comprised 112 samples, while the validation and testing datasets consisted of 24 and 24 samples, respectively.

The results of the validation phase are presented in confusion matrices for each method, showing the best and worst results from repeated experiments as well as the percentage of correct and false detections.

Table 10 shows the output matrix for the training phase of a deep learning algorithm using 1D-CNN-LSTM for corona and non-corona fault detection in the frequency domain. According to the information provided, the training dataset consisted of 112 sets of data, and the algorithm has identified 87 non-corona cases and 25 corona cases with 100% accuracy and a 0% error rate. The algorithm has also been tested using only CNN and LSTM, separately. When using CNN alone, the algorithm has identified 83 non-corona cases and 29 corona cases with 100% accuracy and a 0% error rate. When using LSTM alone, the algorithm has identified 83 non-corona cases and 29 corona cases with 100% accuracy and a 0% error rate.

Table 10.

Frequency domain corona fault classification output matrix for the training phase.

1D-CNN LSTM 1D-CNN-LSTM
Corona Non-Corona Corona Non-Corona Corona Non-Corona
Actual corona 29 0 29 0 25 0
Actual non-corona 0 83 0 83 0 87

The validation dataset in the frequency domain analysis consisted of 24 sets of data, and the 1D-CNN-LSTM algorithm has identified seventeen non-corona cases and seven corona cases with 100% accuracy and a 0% error rate. When using only CNN, the algorithm has identified nineteen non-corona cases and four corona cases with 95.8% accuracy and a 1.2% error rate, where one corona case was mistakenly classified as a non-corona case. When using only LSTM, the algorithm has identified sixteen non-corona cases and five corona cases with 100% accuracy and a 0% error rate. The resulting matrix is displayed in Table 11.

Table 11.

Frequency domain corona fault classification output matrix for the validation phase.

Actual Class 1D-CNN LSTM 1D-CNN-LSTM
Corona Non-Corona Corona Non-Corona Corona Non-Corona
Actual corona 4 1 5 0 7 0
Actual non-corona 0 19 0 19 0 17

Based on the information provided, the 1D-CNN-LSTM algorithm achieved 100% accuracy and a 0% error rate in the testing phase of the frequency domain analysis with 24 sets of data, identifying seventeen non-corona cases and seven corona cases. When using only CNN, the algorithm identified nineteen non-corona cases and four corona cases with 95.8% accuracy and a 4.2% error rate, where one corona case was wrongly classified as a non-corona case. When using only LSTM, the algorithm identified nineteen non-corona cases and five corona cases with 100% accuracy and a 0% error rate, as shown in Table 12.

Table 12.

Frequency domain corona fault classification output matrix for testing the phase.

1D-CNN LSTM 1D-CNN-LSTM
Corona Non-Corona Corona Non-Corona Corona Non-Corona
Actual corona 4 1 5 0 7 0
Actual non-corona 0 19 0 19 0 17

A thorough analysis using performance metrics, including sensitivity, dependability, loss, accuracy, and F1-Measure in Table 13 and Table 14, has been done to further study the truthfulness of CNN-LSTM.

Table 13.

The assessment of various metrics in accordance with the DL structure’s testing-phase performance results for the time domain.

Case Corona Outcome (1)
Approaches Accuracy Sensitivity Dependability F1-Measure
1D-CNN 92.4 100 62 76
LSTM 92.4 100 62 76
1D-CNN-LSTM 98.4 98 89 94
Case Non-Corona Outcome (0)
Approaches Accuracy Sensitivity Dependability F1-Measure
1D-CNN 92.4 91 100 95
LSTM 92.4 91 100 95
1D-CNN-LSTM 98.4 98 100 99

Table 14.

The assessment of different metrics in light of the best performance outcomes of DL structures during testing for the frequency domain.

Case Corona Outcome (1)
Approaches Accuracy Sensitivity Dependability F1-Measure
1D-CNN 95.8 100 95 97
LSTM 100 100 100 100
1D-CNN-LSTM 100 100 100 100
Case Non-Corona Outcome (0)
Approaches Accuracy Sensitivity Dependability F1-Measure
1D-CNN 95.8 100 80 89
LSTM 100 100 100 100
1D-CNN-LSTM 100 100 100 100

Based on the results presented in Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12, it can be concluded that both LSTM and 1D-CNN-LSTM models are effective in detecting corona and non-corona faults in both the time and frequency domains. In the testing phase, the 1D-CNN-LSTM model achieved higher accuracy (98.4%) compared to the LSTM model (92.4%) in the time domain analysis, while both models achieved 100% accuracy in the frequency domain analysis.

However, it is worth noting that in the time domain analysis, the 1D-CNN-LSTM model wrongly identified one corona case as a non-corona case, which led to a slightly lower accuracy compared to the frequency domain analysis. This suggests that further improvements can be made to the 1D-CNN-LSTM model to increase its accuracy in identifying corona cases.

Overall, the results indicate that the 1D-CNN-LSTM model is a promising approach for fault detection in power systems and has the potential to outperform traditional methods. The training process for the 1D-CNN-LSTM model in both the time and frequency domains is more time-consuming than the other techniques due to the large number of model parameters, as shown in Table 2 and Table 3. The 1D-CNN model is considered the most efficient and quickest for training time, with a total of 6738 parameters in the time domain. On the other hand, the LSTM model is considered the most efficient and quickest for training time in the frequency domain, with a total of 35,298 parameters.

It should be noted that the total time required for each model differs between the time domain and frequency domain analyses, as shown in Table 15 below. It is important to mention that the models were implemented on Google Colab, and the LSTM model required the least amount of time for implementation in the time domain analysis, while the 1D-CNN model required the least amount of time for implementation in the frequency domain analysis.

Table 15.

The calculation times of the methods in the time and frequency domains.

Methods Time Domain Frequency Domain
1D-CNN 0:00:42 s 0:00:18 s
LSTM 0:00:26 s 0:00:21 s
1D-CNN-LSTM 0:00:39 s 0:00:28 s

Upon analyzing the outcomes of the three phases of training, validation, and testing, we can conclude that the three models (1D-CNN, LSTM, and 1D-CNN-LSTM) exhibit superior performance compared to the results obtained by the study of [42]. Based on the results obtained from the three phases of training, validation, and testing, we can observe that the three models (1D-CNN, LSTM, and 1D-CNN-LSTM) outperform the results obtained by the researcher, who utilized the extreme learning machine approach to detect corona in the same data in both the time and frequency domains. In the time domain analysis, the results obtained were 90.63%, 87.5%, and 87.5%, while in the frequency domain analysis, the results were 89.84%, 83.33%, and 87.5%. Therefore, we can conclude that the hybrid model, which is the 1D-CNN-LSTM model, has proven its effectiveness in terms of accuracy and error rate in comparison to the previous techniques, and it is a suitable model for detecting corona in switchgear.

5. Conclusions

In conclusion, switchgear plays a critical role in ensuring safety and protection, especially during downstream repairs. Inadequate monitoring, inspection, and evaluation can cause switchgear malfunctions, making it necessary to have a reliable flaw identification system to prevent manual inspection and aid in proper operation. This research proposes various essential deep learning methodologies for corona detection. CNN and LSTM algorithms have previously been used to classify datasets automatically, and 1D CNN, LSTM, and CNN-LSTM architectures were presented to determine defect detection systems based on sound waves in this study. Time and frequency domain analyses were conducted, and the success rates in differentiating corona and non-corona instances in the training, validation, and testing phases were evaluated. The 1D-CNN achieved success rates of 98.3%, 98.4%, and 93.9% in the time domain analysis, while the LSTM achieved 97.3%, 98.4%, and 92.4%. The best-performing method was 1D-CNN-LSTM, with success rates of 99.3%, 98.4%, and 98.4%. In the frequency domain analysis, the 1D-CNN achieved success rates of 100%, 95.8%, and 95.8%, while the LSTM achieved 100%, 100%, and 100%. The accuracy for 1D-CNN-LSTM and the success rates in the training, validation, and testing phases were 100%, 100%, and 100%, respectively. The proposed algorithm performed well and could be enhanced and implemented in real-time simulations, particularly in industrial sectors.

Acknowledgments

This work was supported by the University Tenaga Nasional, BOLDREFREASH 2025, and the AAIBE Chair of Renewable Energy (ChRe) for providing all the laboratory support.

Author Contributions

Conceptualization, S.P.K. and S.K.T.; methodology, Y.A.M.A., C.T.Y., S.P.K. and S.K.T.; software, Y.A.M.A.; validation, Y.A.M.A., C.T.Y., S.P.K. and C.P.C.; formal analysis, Y.A.M.A., C.T.Y. and S.P.K.; investigation, Y.A.M.A., C.T.Y., S.P.K. and C.P.C.; resources, Y.A.M.A., C.T.Y., S.P.K., S.K.T. and C.P.C.; data curation, Y.A.M.A., C.T.Y. and C.P.C.; writing—original draft preparation, Y.A.M.A., C.T.Y. and S.P.K.; writing—review and editing, Y.A.M.A., C.T.Y., S.P.K., S.K.T., C.P.C., T.Y., A.N.A., K.A. and A.A.R.; visualization, S.P.K., C.T.Y., S.K.T. and C.P.C.; supervision, S.P.K. and C.T.Y.; project administration, S.P.K. and S.K.T.; funding acquisition, S.P.K. and S.K.T. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Funding Statement

This research was funded by 202101KETTHA and BOLDREFRESH2025 (J510050002 (IC-6C)).

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Velásquez R.M.A., Lara J.V.M., Melgar A. Reliability model for switchgear failure analysis applied to ageing. Eng. Fail. Anal. 2019;101:36–60. doi: 10.1016/j.engfailanal.2019.03.004. [DOI] [Google Scholar]
  • 2.Alsumaidaee Y.A.M., Yaw C.T., Koh S.P., Tiong S.K., Chen C.P., Ali K. Review of Medium-Voltage Switchgear Fault Detection in a Condition-Based Monitoring System by Using Deep Learning. Energies. 2022;15:6762. doi: 10.3390/en15186762. [DOI] [Google Scholar]
  • 3.Subramaniam A., Sahoo A., Manohar S.S., Raman S.J., Panda S.K. Switchgear Condition Assessment and Lifecycle Management: Standards, Failure Statistics, Condition Assessment, Partial Discharge Analysis, Maintenance Approaches, and Future Trends. IEEE Electr. Insul. Mag. 2021;37:27–41. doi: 10.1109/MEI.2021.9399911. [DOI] [Google Scholar]
  • 4.Durocher D.B., Loucks D. Infrared Windows Applied in Switchgear Assemblies: Taking Another Look. IEEE Trans. Ind. Appl. 2015;51:4868–4873. doi: 10.1109/TIA.2015.2456064. [DOI] [Google Scholar]
  • 5.Courtney J., McDonnell A. Impact on Distribution System Protection with the Integration of EG on the Distribution Network; Proceedings of the 2019 54th International Universities Power Engineering Conference (UPEC); Bucharest, Romania. 3–6 September 2019; [DOI] [Google Scholar]
  • 6.Riba J.-R., Moreno-Eguilaz M., Ortega J.A. Arc Fault Protections for Aeronautic Applications: A Review Identifying the Effects, Detection Methods, Current Progress, Limitations, Future Challenges, and Research Needs. IEEE Trans. Instrum. Meas. 2022;71:3504914. doi: 10.1109/TIM.2022.3141832. [DOI] [Google Scholar]
  • 7.Chang J.-S., Phil A.L., Toshiaki Y. Corona discharge processes. IEEE Trans. Plasma Sci. 1991;19:1152–1166. doi: 10.1109/27.125038. [DOI] [Google Scholar]
  • 8.Ryan H.M., editor. High Voltage Engineering and Testing. The Institution of Electrical Engineers; New York, NY, USA: 2001. No. 32. Iet. [Google Scholar]
  • 9.Javed H., Li K., Guoqiang Z. The Study of Different Metals Effect on Ozone Generation Under Corona Discharge in MV Switchgear Used for Fault Diagnostic; Proceedings of the 2019 IEEE Asia Power and Energy Engineering Conference (APEEC); Chengdu, China. 29–31 March 2019; [DOI] [Google Scholar]
  • 10.Hussain G.A., Zaher A.A., Hummes D., Safdar M., Lehtonen M. Hybrid Sensing of Internal and Surface Partial Discharges in Air-Insulated Medium Voltage Switchgear. Energies. 2020;13:1738. doi: 10.3390/en13071738. [DOI] [Google Scholar]
  • 11.Ishak S., Yaw C.T., Koh S.P., Tiong S.K., Chen C.P., Yusaf T. Fault Classification System for Switchgear CBM from an Ultrasound Analysis Technique Using Extreme Learning Machine. Energies. 2021;14:6279. doi: 10.3390/en14196279. [DOI] [Google Scholar]
  • 12.Javed H., Li K., Zhang G.Q., Plesca A.T. Online Monitoring of Partial Discharge by Measuring Air Decomposition By-Products under Low and High Humidity; Proceedings of the International Conference on Energy, Power and Environmental Engineering (ICEP); Shanghai, China. 23–24 April 2017. [Google Scholar]
  • 13.Gui Y., Zhang X., Zhang Y., Qiu Y., Chen L. Study on the characteristic decomposition components of air-insulated switchgear cabinet under partial discharge. AIP Adv. 2016;6:075106. doi: 10.1063/1.4958883. [DOI] [Google Scholar]
  • 14.Acharya U.R., Sree S.V., Swapna G., Martis R.J., Suri J.S. Automated EEG analysis of epilepsy: A review. Knowl. Based Syst. 2013;45:147–165. doi: 10.1016/j.knosys.2013.02.014. [DOI] [Google Scholar]
  • 15.LeCun Y., Yoshua B., Geoffrey H. Deep learning. Nature. 2015;521:436–444. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
  • 16.Jiang Y., Chung F.-L., Wang S., Deng Z., Wang J., Qian P. Collaborative Fuzzy Clustering From Multiple Weighted Views. IEEE Trans. Cybern. 2014;45:688–701. doi: 10.1109/TCYB.2014.2334595. [DOI] [PubMed] [Google Scholar]
  • 17.Radenović F., Giorgos T., Ondřej C. Fine-tuning CNN image retrieval with no human annotation. IEEE Trans. Pattern Anal. Mach. Intell. 2018;41:1655–1668. doi: 10.1109/TPAMI.2018.2846566. [DOI] [PubMed] [Google Scholar]
  • 18.Choi E., Schuetz A., Stewart W.F., Sun J. Using recurrent neural network models for early detection of heart failure onset. J. Am. Med. Inform. Assoc. 2017;24:361–370. doi: 10.1093/jamia/ocw112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Kong W., Dong Z.Y., Jia Y., Hill D.J., Xu Y., Zhang Y. Short-Term Residential Load Forecasting Based on LSTM Recurrent Neural Network. IEEE Trans. Smart Grid. 2017;10:841–851. doi: 10.1109/TSG.2017.2753802. [DOI] [Google Scholar]
  • 20.Hanson J., Yang Y., Paliwal K., Zhou Y. Improving protein disorder prediction by deep bidirectional long short-term memory recurrent neural networks. Bioinformatics. 2017;33:685–692. doi: 10.1093/bioinformatics/btw678. [DOI] [PubMed] [Google Scholar]
  • 21.Ozcanli A.K., Baysal M. Islanding detection in microgrid using deep learning based on 1D CNN and CNN-LSTM networks. Sustain. Energy Grids Netw. 2022;32:100839. doi: 10.1016/j.segan.2022.100839. [DOI] [Google Scholar]
  • 22.Sainath T.N., Vinyals O., Senior A., Sak H. Convolutional, Long Short-Term Memory, Fully Connected Deep Neural Networks; Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP); South Brisbane, QLD, Australia. 19–24 April 2015; pp. 4580–4584. [DOI] [Google Scholar]
  • 23.Sainath T.N., Senior A.W., Vinyals O., Sak H. Convolutional, Long Short—Term Memory, Fully Connected Deep Neural Networks. No. 10,783,900. [(accessed on 4 January 2023)];U.S. Patent. 2022 September 22; Available online: https://patentimages.storage.googleapis.com/7f/9e/62/c0fa633cefd6df/US10783900.pdf.
  • 24.Jiang Y., Zhao K., Xia K., Xue J., Zhou L., Ding Y., Qian P. A Novel Distributed Multitask Fuzzy Clustering Algorithm for Automatic MR Brain Image Segmentation. J. Med. Syst. 2019;43:118. doi: 10.1007/s10916-019-1245-1. [DOI] [PubMed] [Google Scholar]
  • 25.Zhao B., Lu H., Chen S., Liu J., Wu D. Convolutional neural networks for time series classification. J. Syst. Eng. Electron. 2017;28:162–169. doi: 10.21629/JSEE.2017.01.18. [DOI] [Google Scholar]
  • 26.Cui Z., Wenlin C., Yixin C. Multi-scale convolutional neural networks for time series classification. arXiv. 20161603.06995 [Google Scholar]
  • 27.Zheng Y., Liu Q., Chen E., Ge Y., Zhao J.L. Time series classification using multi-channels deep convolutional neural networks; Proceedings of the Web-Age Information Management: 15th International Conference, WAIM 2014; Macau, China. 16–18 June 2014; [DOI] [Google Scholar]
  • 28.Ismail Fawaz H., Forestier G., Weber J., Idoumghar L., Muller P.-A. Deep learning for time series classification: A review. Data Min. Knowl. Discov. 2019;33:917–963. doi: 10.1007/s10618-019-00619-1. [DOI] [Google Scholar]
  • 29.Van Houdt G., Carlos M., Gonzalo N. A review on the long short-term memory model. Artif. Intell. Rev. 2020;53:5929–5955. doi: 10.1007/s10462-020-09838-1. [DOI] [Google Scholar]
  • 30.Olah C. Understanding LSTM Networks. 2015. [(accessed on 4 January 2023)]. Available online: https://research.google/pubs/pub45500/
  • 31.Yu Y., Si X., Hu C., Zhang J. A review of recurrent neural networks: LSTM cells and network architectures. [(accessed on 4 January 2023)];Neural Comput. 2019 31:1235–1270. doi: 10.1162/neco_a_01199. Available online: https://direct.mit.edu/neco/article-abstract/31/7/1235/8500/A-Review-of-Recurrent-Neural-Networks-LSTM-Cells. [DOI] [PubMed] [Google Scholar]
  • 32.Lipton Z.C., John B., Charles E. A critical review of recurrent neural networks for sequence learning. arXiv. 2015:arXiv:1506.00019.1506.00019 [Google Scholar]
  • 33.Ordóñez F.J., Daniel R. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors. 2016;16:115. doi: 10.3390/s16010115. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Shorten C., Taghi M.K. A survey on image data augmentation for deep learning. J. Big Data. 2019;6:1–48. doi: 10.1186/s40537-019-0197-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Zhao J., Xia M., Lijiang C. Speech emotion recognition using deep 1D & 2D CNN LSTM networks. Biomed. Signal Process. Control. 2019;47:312–323. doi: 10.1016/j.bspc.2018.08.035. [DOI] [Google Scholar]
  • 36.Lu B., Huang W., Xiong J., Song L., Zhang Z., Dong Q. The Study on a New Method for Detecting Corona Discharge in Gas Insulated Switchgear. IEEE Trans. Instrum. Meas. 2021;71:9000208. doi: 10.1109/TIM.2021.3129225. [DOI] [Google Scholar]
  • 37.Bandi M.M., Ishizu N., Kang H.-B. Electrocharging face masks with corona discharge treatment. Proc. R. Soc. A. 2021;477:20210062. doi: 10.1098/rspa.2021.0062. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Schoenau L., Steinpilz T., Teiser J., Wurm G. Corona discharge of a vibrated insulating box with granular medium. Granul. Matter. 2021;23:1–6. doi: 10.1007/s10035-021-01132-3. [DOI] [Google Scholar]
  • 39.Krizhevsky A., Ilya S., Geoffrey E.H. Imagenet classification with deep convolutional neural networks. Commun. ACM. 2017;60:84–90. doi: 10.1145/3065386. [DOI] [Google Scholar]
  • 40.Kiranyaz S., Avci O., Abdeljaber O., Ince T., Gabbouj M., Inman D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021;151:107398. doi: 10.1016/j.ymssp.2020.107398. [DOI] [Google Scholar]
  • 41.Hochreiter S., Jürgen S. Long short-term memory. Neural Comput. 1997;9:1735–1780. doi: 10.1162/neco.1997.9.8.1735. [DOI] [PubMed] [Google Scholar]
  • 42.Ishak S., Koh S.P., Tan J.D., Tiong S.K., Chen C.P. Corona fault detection in switchgear with extreme learning machine. Bull. Electr. Eng. Inform. 2020;9:558–564. doi: 10.11591/eei.v9i2.2058. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Not applicable.


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES