Skip to main content
PLOS One logoLink to PLOS One
. 2023 Feb 9;18(2):e0281397. doi: 10.1371/journal.pone.0281397

Study on recognition of coal and gangue based on multimode feature and image fusion

Lijuan Zhao 1,2, Liguo Han 1,*, Haining Zhang 1, Zifeng Liu 3, Feng Gao 3, Shijie Yang 1, Yadong Wang 1
Editor: Ayan Seal4
PMCID: PMC9910643  PMID: 36758046

Abstract

Aiming at the problems of low accuracy of coal gangue recognition and difficult recognition of mixed gangue rate, a coal rock recognition method based on modal fusion of RGB and infrared is proposed. A fully mechanized coal gangue transportation test bed is built, RGB images are obtained by camera, and infrared images are obtained by industrial microwave heating system and infrared thermal imager. the image data of the whole coal, whole gangue, and coal gangue with different gangue mixing as training and test samples, identify the released coal gangue and its mixing rate. The AlexNet, VGG-16, ResNet-18 classification networks and their convolutional neural networks with modal feature fusion are constructed. results: The classification accuracy of ResNet networks on RGB and infrared image data is higher than AlexNet and VGG-16 networks. The early convergence network performance of ResNet is verified through the convergence of different models. The recognition rate of the network is 97.92 the confusion matrix statistics, which verifies the feasibility of the application of modal fusion method in the field of coal gangue recognition. The fusion of modal features and early models of ResNet coal gangue, which is the basic premise for realizing intelligent coal caving.

1. Introduction

Fully mechanized mining and top coal mining have become two of the main methods of thick coal seam mining in China, and intelligent coal mining is the only way to the high-quality development of the coal industry [1, 2]. The basic premise of intelligent coal mining is to achieve intelligent coal caving, which must have the capability of automatic recognition of gangue. Coal gangue recognition is the key technology of fully mechanized mining and top coal mining and has a significant impact on the coal extraction rate and coal recovery quality. In recent years, many experts and scholars have conducted a lot of in-depth research on the coal gangue recognition problem. Liu Wei et al. [3] proposed a method for detecting the interface of coal gangue to identify the Hilbert spectrum information entropy feature of the vibration of coal gangue. Zhang Ningbo et al. [4] proposed a method for measuring and identifying the mixed coal gangue of coal gangue in the process of natural coal discharge, providing a basis for judging the appearance of coal gangue using natural coal radiation technology. Liu Chuang [5] [Unpublished] proposed an active gangue recognition method for microwave heating-infrared (IR) detection, studied the recognition mechanism, and analyzed its feasibility. Yuan Yuan et al. [6] designed a feature extraction and classification method for coal acoustic signals based on the wavelet packet decomposition and random forest (RF) algorithm. Dou Xijie et al. [7] proposed a gangue recognition method based on an intrinsic mode function energy moment and support vector machine, and this method effectively recognized the data of multiple gangue vibration samples. Xue Guanhui et al. [8] proposed a coal gangue image recognition method based on the RF algorithm. The RF model’s performance improved after dimension reduction. Jiang Lei et al. [9] based on convolutional neural networks (CNNs) and lightweight dilated CNNs, established a fully dilated CNN-based coal and coal gangue intelligent recognition model, using the hydraulic support tail beam vibration signal Mel frequency cepstral coefficient feature matrix as the CNN input. They realized the structure optimization of the recognition model, significantly improved the operation speed, reduced the use of resources, and revealed the recognition mechanism and classification basis of the model. Shan Pengfei et al. [10] investigated the optimization method of the attention mechanism fusion in the ResNet50 backbone feature extraction network, which determined the best fusion position with the coal-gangue falling state detection as the target, increasing the ability to extract the weight information of coal-gangue. Zhang Jinwang et al. [11] proposed a new concept of coal gangue recognition of "liquid intervention + infrared detection," and a recognition test of liquid intervention under different mixing degrees was conducted. By selecting a reasonable liquid type, temperature, and intervention amount, the coal rock recognition accuracy can be effectively improved. Combined with the virtual prototype technology, a project team proposed a coal rock cut state recognition scheme based on the cyber-physical system concept and integrated heterogeneous data such as acquisition, processing, and recognition data from multiple fields, realizing the adaptive height adjustment of the shearer [12, 13]. Due to the poor working conditions and complex environment, the occurrence conditions of the top coal, the coal discharge mode of the tail beam, the kinematic parameters of the scraper conveyor, the gradient characteristics of the hydraulic system, and the interaction between the roof beam and roof will directly or indirectly affect the top coal caving process of the top coal caving support under the conditions of existing coal gangue. Although the recognition technology of integrated gangue based on vibration and sound characteristics can realize the recognition of gangue, available signals can be easily polluted by environmentally stimulated signals during the underground production process; thus, the ability to effectively extract gangue signals from the mixed signals is critical [14] [Unpublished]; Although research on natural radiation methods and other technologies is not limited by the underground coal discharge environment, it is difficult to apply them to a working face that does not contain radioactive elements or has a low content of radioactive elements and contains too much gangue in coal gangue [15] [Unpublished]. Therefore, how to quickly and accurately identify the rapidly moving coal gangue and its gangue content on a rear scraper conveyor remains a technical bottleneck to realizing intelligent coal discharge. The image-based recognition method of coal gangue primarily obtains information about the coal caving status using cameras based on the coal caving mechanism and employs machine learning algorithms to complete the recognition and classification task, which have the advantages of high reliability and fast response. However, the fast movement of back scraper conveyors carrying coal gangue, the low brightness of the underground environment, and the high dust level remain some of the technical bottlenecks of using images to identify coal gangue and realize intelligent coal caving.

In this study, a multimodal information acquisition and test platform of gangue microwave excitation is constructed. By simulating the motion state of a scraper conveyor at the back of the working face, the RGB image data of gangue are obtained by cameras, and the IR image data of gangue are collected by an IR thermal imager. A two-input CNN based on ResNet is constructed, and the modal fusion of RGB and IR images by feature fusion is adopted to realize the accurate recognition of gangue and gangue content, verify the feasibility of the modal fusion method in the field of gangue recognition, and provide a new method for intelligent coal release. Accurate recognition of coal gangue is achieved by combining multimodal features and ResNet model technology, which replaces the eye recognition of coal miners, improves mining safety, increases the top coal extraction rate, and provides a new method for intelligent coal caving and coal mine intellectualization.

2. Research methods

In recent years, modal fusion techniques have been extensively used in medical imaging omics [1618], self-driving [19], mass detection [20], and other fields. Combined with the characteristics of coal gangue recognition in a comprehensive discharge working face, mode fusion technology is applied to research on intelligent top discharge coal discharge support with low illumination, high dust, and small space. Through torch.cat channel splicing, the features of gangue RGB and IR images are integrated, a multimodal fusion network based on ResNet-18 is constructed, and the CrossEntropyLoss loss function is used for gangue classification while improving the recognition accuracy and ensuring system stability.

2.1 Convolutional neural network

A CNN is one of the representative algorithms of deep learning, which introduces the local connection and weight sharing between the network layers, reducing the model parameters and solving the problems of difficult convergence and overfitting. A CNN is very suitable for image processing. During each convolution operation, an image with an input size of Hi * Wi is subjected to a convolution kernel of Fh * Fh. The dilated convolution value is D, the filling is Ph * Ph, and the step size is Sh * Sw. The output image size Ho * Wo is as follows:

Ho=[Hi+2×PhD×(Fh1)1Sh+1], (1)
Wo=[Wi+2×PwD×(Fw1)1Sw+1]. (2)

The main structure of the CNN includes the convolutional, pooling, and fully connected layers, and a typical CNN classification network structure is depicted in Fig 1.

Fig 1. CNN structure.

Fig 1

2.2 Modal fusion method based on ResNet

The deep residual network (DRN) was launched in 2015 [21, 22]. Compared with traditional deep neural networks, a jump connection mode is used in DRN, which eliminates harmful data features and solves the problem of error gradient disappearance with an increase in network depth [23]. Taking the classical ResNet-18 network, as an example, it is composed of multiple residue blocks, one of which is shown in Fig 2, and the network structure is depicted in Fig 3.

Fig 2. The ResNet residual module.

Fig 2

Fig 3. ResNet-18 network structure diagram.

Fig 3

Modal data include visible light, IR, and depth images [24, 25]. According to the data level, the modal fusion mode can be divided into pixel-level fusion, feature-level fusion, and decision-level fusion. In this study, a feature-based mode fusion scheme is proposed, and an early mode fusion method is constructed. The model structure is depicted in Fig 4. The pseudocode of the ResNet early fusion algorithm is shown in Table 1. The two types of original mode data with three channels are fused into six-channel data that are input into the ResNet network for training to accurately obtain the corresponding category by torch.cat channel splicing. The model structure of late mode fusion is shown in Fig 5. The pseudocode of the ResNet late fusion algorithm is shown in Table 2. The original modal data of two types of three channels are passed through the ResNet network to form 512 channels of data, which are then spliced into 1024 channels by torch.cat. Finally the data is trained through a fully connected layer to obtain the accuracy of the corresponding category. In addition, the RGB and IR image data of coal gangue are input to the network in one-to-one correspondence. The RGB and IR image data names of coal gangue at the same time have the same part. By retrieving the name of the same part, the data at the same time will be input into the network one by one. The feature-based mode fusion method can reduce the heterogeneity difference between modes while maintaining the integrity of specific features of each mode, effectively overcoming the problem of a low single-mode recognition rate and significantly improving the generalization ability of the model.

Fig 4. The early fusion network.

Fig 4

Table 1. ResNet early fusion pseudocode.

Algorithm 1: ResNet early modal feature fusion algorithm
Input: Infrared dataset Dh, RGB dataset Dz, Dataset label L, ResNet model M, Iteration number N;
Output: Accuracy of image classification of test set;
1 Divide the data set {Dh, Dz, L} into training set {Dh1, Dz1, L1} and test set {Dh2, Dz2, L2}; The corresponding batch is divided into B1 and B2;
2 for j = 0:N
3 foreach b1B1
4 Fusion of training data on channel dimension, D1_b1 = concat (Dh1_b1, Dz1_b1);
5 Send D1_b1 to model M for training;
6 foreach b2B2
7 Fusing the test data on the channel dimension, D2_b2 = concat (Dh2_b2, Dz2_b2);
8 Test model M;
9 Get Accuracy

Fig 5. Late fusion network.

Fig 5

Table 2. ResNet late fusion pseudocode.

Algorithm 2: ResNet late modal feature fusion algorithm
Input: Infrared dataset Dh, Natural light data set Dz, Dataset label L, ResNet model about Mh and Mz without full connection layer FC1, Iteration number N;
Output: Accuracy of image classification of test set;
1 Divide the data set {Dh, Dz, L} into training set {Dh1, Dz1, L1} and test set {Dh2, Dz2, L2}; The corresponding batch is divided into B1 and B2;
2 for j = 0:N
3 foreach b1B1
4 Training model Mh Dh1_b1_f = Mh(Dh1_b1);
5 Training model Mz Dz1_b1_f = Mz(Dz1_b1);
6 Fusion of two modal features in channel dimension D1_b1_f = concat(Dh1_b1_f, Dz1_b1_f);
7 Send D1_b1_f to the full connection layer FC2 for testing;
8 foreach b2B2
9 Dh2_b2_f = Mh(Dh2_b2);
10 Dz2_b2_f = Mz(Dz2_b2);
11 Fusion of two modal features in channel dimension D2_b2_f = concat(Dh2_b2_f, Dz2_b2_f);
12 Send D2_b2_f to the full connection layer FC2 for testing;
13 Get Accuracy

2.3 Loss function and evaluation index

In PyTorch, for multiclassification problems, nn.Crossentropyloss is used as the loss function, which calculates the cross entropy loss between the predicted value and the target. Here, the CrossEntropyLoss loss function is applied, and it is no longer necessary to use the Softmax classifier for the probability mapping of input features. The memory network input is (N, C), where N represents the minibatch number, and C represents the total number of categories. Its loss function is defined by the following Eq (3):

ln=logexp(xn,yn)c=1Cexp(xnc), (3)

where xnc denotes the output value of the network, n = 1, 2,…, N, c = 1, 2, …, C, and the true category of the nth sample is yn.

In the actual sample acquisition process, there may be data imbalance between different categories. Therefore, it is necessary to introduce a weight K = (k1, k2, , kC) to ensure the data balance; the loss function after adding the weight is defined by Eq (4):

ln=kyn(xn,yn+log(c=1Cexp(xnc))). (4)

The actual network training is conducted in minibatch units, and the loss after an iterative training session can be returned in three ways, as shown in Eq (5):

l={(l1,l2,,lN)ifreduction="none"n=1Nlnn=1Nkynifreduction="mean"n=1Nlnifreduction="sum". (5)

In the accuracy evaluation of network models, accuracy (ACC), precision (PPV), recall rate (TPR), and F1 score are typically used to measure the performance of recognition networks [26]. ACC indicates the proportion of the number of samples in the total samples correctly classified as coal gangue, which can be determined by Eq (6).

ACC=TP+TNTP+TN+FP+FN. (6)

PPV represents the weighted average value of the precision (PPVi) of different coal gangue categories, revealing the discrimination ability of the recognition network to negative samples, which can be determined by Eqs (7) and (8). PPVi is proportion of correct numbers predicted to be in that category versus all predicted to be in that category among the gangue prediction category of the classification network model, where i indicates a category, N represents the total number of categories, and K indicates the weight of a gangue sample in the total sample.

PPVi=TPiTPi+FPi, (7)
PPV=iNPPVi×KiN. (8)

TPR represents the weighted average value of different coal gangue categories (TPRi), revealing the discrimination ability of the recognition network to positive samples, which can be determined by Eqs (9) and (10). Here, i indicates a category in the classification network model and TPRi denotes the ratio of the number of samples correctly predicted to the total number of samples in that category.

TPRi=TPiTPi+FNi, (9)
TPR=iNTPRi×KiN. (10)

F1 score, as the weighted average of precision (PPV) and recall (TPR), is used to evaluate the classification network performance and can be determined by Eq (11).

F1=2PPV×TPRPPV+TPR. (11)

3. Simulation test of coal gangue movement state

In the process of gangue identification, the accurate acquisition of multimodal features is a prerequisite for improving the speed and accuracy of gangue identification. Because gangue infrared features were more obvious gangue RGB features after excitation by a heat source, a microwave-heated gangue transport test bench was constructed in this section. Both RGB and infrared images of gangue were acquired simultaneously. Multimodal features of gangue were obtained and preprocessed by scientifically setting the mixing rate and designing multiple groups of test categories.

3.1 Construction of test bench

A multimodal information acquisition test platform of coal gangue microwave excitation is constructed in this study. By simulating the movement state of the scraper conveyor at the back of the fully mechanized top coal caving face, RGB and infrared images excited by microwave are collected. The experimental device is primarily composed of the main conveyor, industrial-grade microwave emission system, RGB image collector, and infrared thermal imager. A schematic of the acquisition equipment and recognition system is depicted in Fig 6. The microwave emission system is primarily composed of a magnetron, waveguide, thermostat, fan, air guide, power supply, and shielding shell. The RGB image collector consists of a triangle holder and a video camera. The resolution of the image taken by the camera on the computer is 720*1280, and the number of frames taken per second is 30 fps. The Haikang Micro Shadow Infrared Thermal Imager under Haikang Video H16 was used in this experiment. Its output image resolution on a computer is 640*480 pixels, and the number of frames taken per second is 25 fps.

Fig 6. Multimodal image acquisition test platform.

Fig 6

(a)Tester; (b) Schematic of the comprehensive discharge and waste mixing rate recognition system.

3.2 Test scheme

Gangue content in the coal release process [27] can be determined using Eq (12):

R=MrMc+Mr×100%. (12)

Here, Mc denotes the volume of the top coal, and Mr denotes the volume of the gangue.

The original coal gangue sample is shown in Fig 7. The mass of the coal gangue sample is measured by electronic scale, as depicted in Fig 8. The samples are divided into whole coal, whole gangue, 10% gangue content [28, 29], 25% gangue content [30], and 50% gangue content [30]. The five groups were tested, as shown in Fig 9.

Fig 7. Coal gangue test sample.

Fig 7

(a) Coal; (b) Gangue.

Fig 8. Mass measurement of coal gangue.

Fig 8

(a) Coal; (b) Gangue.

Fig 9. Broken coal gangue sample.

Fig 9

(a) Coal; (b) Gangue.

3.3 Data acquisition and production

Put the gangue in the coal falling area of the conveyor and run the conveyor. Start the microwave heating system, and simultaneously place the RGB image camera and infrared thermal imager at symmetrical positions on the left and right sides of the conveyor and start them synchronously. After the conveyor runs for 5 min, RGB and infrared video data are obtained, and 7200 RGB images and 6000 infrared images are obtained after decomposing the video data by frames. The video data obtained through five groups of experiments with different mixing rates were copied to the computer and intercepted in the same way, with 36000 RGB images and 30000 infrared images obtained. The multimodal images of the samples obtained are shown in Figs 1014. The figures show that the whole coal and gangue excited by microwave exhibit a state of near full red in the infrared images, and the brightness of the two images (RGB and IR) is slightly different. The infrared image of the gangue containing sample is in a bright state in the coal block area, while the gangue area is in a dark state. The positional relationship of the IR image is consistent with that of the RGB image.

Fig 10. Multimodal image of whole coal.

Fig 10

(a) RGB image; (b) Infrared image.

Fig 14. Multimodal image with 50% gangue content.

Fig 14

(a) RGB image; (b) Infrared image.

Fig 11. Multimodal image of whole gangue.

Fig 11

(a) RGB image; (b) Infrared image.

Fig 12. Multimodal image with 10% gangue content.

Fig 12

(a) RGB image; (b) Infrared image.

Fig 13. Multimodal image with 25% gangue content.

Fig 13

(a) RGB image; (b) Infrared image.

4. Model comparison, validation, and analysis

Under complex and harsh conditions such as low brightness, high dust, and coal gangue stacking, it is difficult to improve the recognition accuracy of coal gangue using a single feature. In this section, the obtained multimodal features of coal gangue were trained and tested on different network models to analyze their recognition performance. Then, the optimal fusion scheme was obtained by different modal feature fusion methods.

4.1 Comparison of different single-mode models

According to the classical image classification network commonly used by a CNN, single-mode classification analysis is performed for RGB and infrared modal images. The early training classification networks of AlexNet, VGG-16, and ResNet-18 are built using Python 3.7 language, python architecture, and the PyCharm platform. The number of random samples for each type of single-mode model is 6000, and the total number of samples is 30000. The training and verification sets have a ratio of 4:1, and the number of iterations is set to 400. Other parameter settings are consistent. The training parameter settings are shown in Table 3, and the computer configuration and relevant application versions are shown in Table 4. Through the operation analysis of the Adam optimizer, the recognition ability of the RGB and infrared modal models obtained by the Tensorboard platform is shown in Table 5.

Table 3. Training parameters.

Parameter Numeric value
Learning rate 0.0001
Smooth constant β1 0.9
Smooth constant β2 0.999
Eps 1e-08

Table 4. Computer configuration and relevant application versions.

Name Model version
Central processing unit AMD Ryzen 9 5950X 16-Core
Graphics processing unit NVIDIA GeForce RTX 3060
System memory 64G
Solid State Disk 1T
PyCharm 2020.1 x64
Python 3.8.13
Tensorboard 2.9.0

Table 5. Recognition and comparison of different models of RGB and infrared modes.

Model of cognition RGB modality of average recognition rate (%) Infrared modality of average recognition rate (%) Time consumed by RGB modality (min) Time consumed by infrared modality (min)
AlexNet 74.5 79.1 10.34 10.43
VGG-16 31.4 56.8 50.40 47.25
ResNet-18 83.9 87.7 10.58 11.15

Table 5 shows that the training time of the same mode on different networks is different. The time consumed on the AlexNet network is the shortest, whereas the time consumed on the VGG-16 network is the longest. This is because the AlexNet network depth is 8, and the required parameters are the least, whereas the VGG-16 network depth is 16, and its parameters are large. The average recognition rates of RGB and infrared modal data on the ResNet-18 network are 83.9% and 87.7%, respectively. The performance of these two modal data on the ResNet-18 network is higher than that of other networks. This is because this network introduces residual blocks, which solves the degradation problem in the deep network. On the same network model, the recognition accuracy of infrared modal data is higher than that of RGB modal data, indicating that its data resolution characteristics are strong and that it has obvious advantages over RGB modal data.

4.2 Fusion results and evaluation

60000 sample data points were obtained through the model information collection platform, each sample type has 12000 data points, and each category contains 6000 data points of infrared and RGB modal data. After dividing the data into the training and test sets in a ratio of 4:1, the two modal image types are input into the feature-based early fusion network and the late fusion network in one-to-one correspondence. The recognition rate of the obtained training process is depicted in Fig 15, and the training loss is shown in Fig 16.

Fig 15. Training recognition rate.

Fig 15

Fig 16. Training loss.

Fig 16

Fig 15 shows that the recognition rate of the two fusion networks increases rapidly with an increase in the number of iterations. After reaching a certain number of iterations, the recognition rate increases slowly. The last two fusion networks are in convergence. The recognition accuracy of the early fusion network is 97.6%, and the recognition accuracy of the late fusion network is 94.4%. This is because the early fusion network performs feature fusion before the input network, and the entire network training process requires fewer parameters, which is easy to converge and has high stability. The late fusion network performs feature fusion after the input network, and its parameters are twice as large as those of the early fusion network. The network training time is long, the stability is low, and the generalization ability is weak. The early fusion network outperforms the late fusion network for the coal gangue classification problem and its samples of RGB and infrared mode fusion.

Fig 16 shows that the training loss of the two fusion networks decreases rapidly with an increase in the number of iterations, and the loss decreases slowly after reaching a certain number of iterations. Finally, the last two fusion networks are in a convergence state. After the number of iterations of the early fusion network reaches 30000, the training loss tends to be stable without large fluctuations. Even if the late fusion network reaches 40000 iterations, the later training loss also exhibits significant fluctuation. The training loss of the early fusion network is 0.1, and the recognition accuracy of the late fusion network is 0.23. The early fusion network can complete the coal gangue classification target task faster and better than the late fusion network.

To further analyze the recognition and classification effect of coal gangue, the confusion matrix is used for visual analysis, and the results are shown in Fig 17. As shown in the figure, the total number of samples in the test set is 60,00. Taking the gangue rate of 25% as an example, the total number of samples under this category is 1200, the correct number of samples is 1173, and 27 data samples are misjudged. This is because the colors of some gangue samples are similar to those of coal blocks, and the image of the gangue rate at a certain time is the same as that of other categories, resulting in interference with the classification results, However, such interference only accounts for 2.25% of the total samples. Through Eqs (6)–(11), the accuracy of this category is determined as 97.75%, the precision is 97.18%, and the recall rate is 97.43%, which reflect the sample category status and can ensure high-precision classification results. Similarly, the accuracy of the modal fusion network model is determined as 97.92%, the precision is 97.85%, the recall rate is 97.85%, and the F1 score is 97.85%. The early modal fusion method based on ResNet is more stable when classifying and recognizing coal gangue images and can accurately identify different gangue rates.

Fig 17. Confusion matrix.

Fig 17

4.3 Analysis of fusion results of different models

To verify the reliability of this model in coal gangue recognition, it is compared with the early fusion methods of other models, as shown in Table 6. The evaluation indicators in the table are obtained by taking the category of 25% gangue as an example. As shown in the table, the ResNet model fusion method is the best. Compared with VGG-16 model fusion, the accuracy of our proposed method is improved by 49.33%. Compared with the AlexNet model fusion method, the proposed method is also significantly improved. The above experimental analysis results show that under the same working and sample data conditions, the coal gangue recognition performance of the proposed ResNet early fusion model is better than that of other model fusion methods.

Table 6. Fusion results of different models.

Method accuracy Precision Recall rate F1-Score
VGG early fusion 48.59 47.90 48.14 48.02
AlexNet early fusion 82.25 81.88 82.21 82.04
ResNet early fusion 97.92 97.85 97.85 97.85

5. Conclusions

Aiming at the problems of low recognition accuracy of coal gangue and difficult recognition of gangue content rates, through the early fusion of dual ResNet models, a coal gangue recognition method based on image multimodality is proposed in this study. First, the captured video data are preprocessed by frame capturing and clipping. Second, the obtained infrared and RGB image features are extracted for channel fusion. Finally, the fused data are input into the ResNet network for training and classification for an improved recognition effect.

The validity of the proposed method is verified by comparing its results with the single-mode experimental results. In addition, from the analysis of different model fusion results, the ResNet early model fusion network has the highest recognition rate, and the recognition accuracy of gangue content can reach 97.92%.

The proposed method, based on RGB and infrared fusion, has the capability of automatic recognition of coal gangue, which is the basic premise for realizing intelligent coal caving. In the future, relying on gangue recognition technology, the mechanism of the intelligent opening and closing of the tail beam caving will be investigated to achieve the maximum top coal extraction rate. In addition, such a study provides technical support for realizing a digital twin of "three machines and one frame" in coal mines. Through the promotion of gangue recognition and intelligent coal drawing technology, unmanned and intelligent mining of gangue will be realized.

Data Availability

All relevant raw data are in the paper and the kaggle database. The minimum data set is included in the kaggle database. Modal data set of coal and gangue are available from the kaggle database. (URL: https://www.kaggle.com/datasets/xiaomangguo/meigan).

Funding Statement

The study was supported by the National Natural Science Foundation of China (project no. 51674134), URL: https://www.nsfc.gov.cn/. The funder had no role in the study design, data collection and analysis, decision to publish, and preparation of the manuscript.

References

  • 1.Wang J.; Pan W.; Zhang G.; Yang S.; Yang K.; Li L. Principles and applications of image-based recognition of withdrawn coal and intelligent control of drawing opening in longwall top coal caving face. J. Journal of China Coal Society, 2022, 47(01), 87–101. doi: 10.13225/j.cnki.jccs.yg21.1530 [DOI] [Google Scholar]
  • 2.Hao X.; Cao W.; Wang X.; Fan H. Numerical Simulation Research on Reasonable Caving Ratio of Extremely Thick Coal Seam with Fully Mechanized Caving Mining. J. Coal Technology, 2022, 41(02), 9–13. doi: 10.13301/j.cnki.ct.2022.02.003 [DOI] [Google Scholar]
  • 3.Liu W.; Hua Z.; Wang S. Vibrational Feature Analysis for Coal Gangue Caving Based on Information Entropy of Hilbert Spectrum. J. China Safety Science Journal, 2011, 21(04), 32–37. doi: 10.16265/j.cnki.issn1003-3033.2011.04.001 [DOI] [Google Scholar]
  • 4.Zhang N.; Liu C.; Chen X.; Chen Y. Measurement analysis on the fluctuation characteristics of low level natural radiation from gangue. J. Journal of China Coal Society, 2015, 40(05): 988–993. doi: 10.13225/j.cnki.jccs.2014.0776 [DOI] [Google Scholar]
  • 5.Liu C. Research on the Method of Synergetic Multi-Windows Top Coal Caving and the Mechanism of Coal-Gangue Identification in Longwall Top Coal Caving Working Face. Ph.D. Thesis, Henan Polytechnic University, 2018. [Google Scholar]
  • 6.Yuan Y.; Wang J.; Zhu D.; Wang J.; Wang T.; Yang K. Feature extraction and classification method of coal gangue acoustic signal during top coal caving. J. Journal of Mining Science and Technology, 2021, 6(06): 711–720. doi: 10.19606/j.cnki.jmst.2021.06.010 [DOI] [Google Scholar]
  • 7.Dou X.; Wang S.; Xie Y.; Xuan T. Coal and gangue identification based on IMF energy moment and SVM. J. Journal of Vibration and Shock, 2020, 39(24), 39–45. doi: 10.13465/j.cnki.jvs.2020.24.006 [DOI] [Google Scholar]
  • 8.Xue G.; Li X.; Qian X.; Zhang Y. Coal-gangue image recognition in fully-mechanized caving face based on random forest. J. Journal of Mine Automation, 2020, 46(05), 57–62. doi: 10.13272/j.issn.1671-251x.2019110064 [DOI] [Google Scholar]
  • 9.Jiang L.; Ma L.; Yang K.; Xu Z. Coal and gangue intelligent separation based on MFCC and FD-CNN convolutional neural network for top coal caving mining. J. Journal of China Coal Society, 2020, 45(S2), 1109–1117. doi: 10.13225/j.cnki.jccs.2020.0738 [DOI] [Google Scholar]
  • 10.Shan P.; Sun H.; Lai X.; Zhu X.; Yang J.; Gao J. Identification method on mixed and release state of coal-gangue masses of fully mechanized caving based on improved Faster R-CNN. J. Journal of China Coal Society, 2022, 47(03), 1382–1394. doi: 10.13225/j.cnki.jccs.xr21.1662 [DOI] [Google Scholar]
  • 11.Zhang J.; He G.; Wang J. Coal /gangue recognition accuracy based on infrared image with liquid intervention under different mixing degree. J. Journal of China Coal Society, 2022, 47(03), 1370–1381. doi: 10.13225/j.cnki.jccs.xr21.1922 [DOI] [Google Scholar]
  • 12.Wang Y.; Zhao L.; Zhang M. Research on self-adaptive height adjustment control strategy of shearer. J. Journal of China Coal Society: 1–19. doi: 10.13225/j.cnki.jccs.2021.1371 [DOI] [Google Scholar]
  • 13.Zhang M.; Zhao L.; Wang Y. Research on Recognition System of Coal-rock Cutting State Based on CPS Perception Analysis. J. Journal of China Coal Society, 2021, 46(12): 4071–4087. doi: 10.13225/j.cnki.jccs.2021.0236 [DOI] [Google Scholar]
  • 14.Song Q. Study and application on the caving automation in fully mechanized top coal caving face. Ph.D. Thesis, China University of Mining and Technology, 2015 [Google Scholar]
  • 15.Wang B. Research on Automatic Recognition of the Coal-rock Interface in Top Coal Caving. Ph.D. Thesis, Shandong University, 2012. [Google Scholar]
  • 16.Fang X.; Fang Y.; Wang D. Brain tumor image segmentation method based on multi-modal feature fusion. J. Chinese Journal of Medical Physics, 2022, 39(06), 682–689. [Google Scholar]
  • 17.Wang L.; Chao Y.; Tian L.; Chen Q.; Guo S.; Zhang J., et al. Adaptive multi-modality fusion network for glioma grading. J. Journal of Image and Graphics, 2021, 26(09): 2243–2256. [Google Scholar]
  • 18.Yang R.; Li Q.; Mao C.; Peng X.; Wang Y.; Guo Y., et al. Multimodal image fusion technology for diagnosis and treatment of the skull base-infratemporal tumors. J. Journal of Peking University(Health Sciences), 2019, 51(01), 53–58. doi: 10.19723/j.issn.1671-167X.2019.01.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Zhang X.; Zou Z.; Li Z.; Liu H.; Li J. Deep multi-modal fusion in object detection for autonomous driving. J. CAAI transactions on intelligent systems, 2020, 15(4), 758–771. [Google Scholar]
  • 20.Zhu B.; Shi Y.; Wang J. Welding Quality Detection Method Based on Multimodal Tensor Fusion. J/OL. Hot Working Technology. 10.14158/j.cnki.1001-3814.20211460 [DOI] [Google Scholar]
  • 21.Zhang X.; Hao R.; Cheng W.; Xia H.; Duan Z. Combination of multi-scale feature fusion and improved ResNet Research on Fault Diagnosis of Gearbox. J. Mechanical Science and Technology for Aerospace Engineering: 1–7. doi: 10.13433/j.cnki.1003-8728.20220146 [DOI] [Google Scholar]
  • 22.He K, Zhang X, Ren S, et al. Deep Residual Learning for Image Recognition. C. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. [Google Scholar]
  • 23.Liu X.; Jin T.; Liu Y.; Gong Z.; Miu H.; Lan M. Open Circuit Fault Diagnosis Method of Electric Vehicle DC Charging Pile Based on Tensor Reshape Fusion Diagnostic Model. J. Proceedings of the CSEE, 1–13. [Google Scholar]
  • 24.Liu L.; Ren Y.; Wang L. Face Anti-spoofing Detection Algorithm Based on a Multi-modal and Multi scale Fusion. J. Journal of Information Security Research, 2022, 8(05), 513–520. [Google Scholar]
  • 25.Zhang D.; Yang Y.; Huang H.; Xie Z.; Li W. Multimodal Fusion Classification Network Based on Distance Confidence Score. J. Journal of Shanghai Jiao Tong University, 2022, 56(01), 89–100. doi: 10.16183/j.cnki.jsjtu.2020.186 [DOI] [Google Scholar]
  • 26.Liu C.; Wang W.; Wang M.; Lv F.; Konan M. An efficient instance selection algorithm to reconstruct training set for support vector machine. J. Knowledge-Based Systems, 2017, 116(1), 58–73. doi: 10.1016/j.knosys.2016.10.031 [DOI] [Google Scholar]
  • 27.Zhang N.; Lu Y., Liu C.; Yang P. Basic study on automatic detection of coal and gangue in the fully mechanized top coal caving mining. J. Journal of Mining & Safety Engineering, 2014, 31(04), 532–536. doi: 10.13545/j.issn1673-3363.2014.04.006 [DOI] [Google Scholar]
  • 28.Liu G.; Song X.; Li H.; Cao J. Research on optimization of top coal caving technology parameters in fully mechanized top coal caving. J. China Mining Magazine, 2021, 30(12), 90–97. [Google Scholar]
  • 29.Zhu D.; Chen Z. A model for top coal recovery ratio assessment and its application in longwall top coal caving. J. Journal of China Coal Society, 2019, 44(09): 2641–2649. doi: 10.13225/j.cnki.jccs.2018.1297 [DOI] [Google Scholar]
  • 30.Huang B.; Liu C.; Cheng Q. Relation between top-coal drawing ratio and refuse content for fully mechanized top coal carving. J. Journal of China Coal Society, 2007(08), 789–793. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All relevant raw data are in the paper and the kaggle database. The minimum data set is included in the kaggle database. Modal data set of coal and gangue are available from the kaggle database. (URL: https://www.kaggle.com/datasets/xiaomangguo/meigan).


Articles from PLOS ONE are provided here courtesy of PLOS

RESOURCES