Abstract
Tomato plants' disease detection and classification at the earliest stage can save the farmers from expensive crop sprays and can assist in increasing the food quantity. Although, extensive work has been presented by the researcher for the tomato plant disease classification, however, the timely localization and identification of various tomato leaf diseases is a complex job as a consequence of the huge similarity among the healthy and affected portion of plant leaves. Furthermore, the low contrast information between the background and foreground of the suspected sample has further complicated the plant leaf disease detection process. To deal with the aforementioned challenges, we have presented a robust deep learning (DL)-based approach namely ResNet-34-based Faster-RCNN for tomato plant leaf disease classification. The proposed method includes three basic steps. Firstly, we generate the annotations of the suspected images to specify the region of interest (RoI). In the next step, we have introduced ResNet-34 along with Convolutional Block Attention Module (CBAM) as a feature extractor module of Faster-RCNN to extract the deep key points. Finally, the calculated features are utilized for the Faster-RCNN model training to locate and categorize the numerous tomato plant leaf anomalies. We tested the presented work on an accessible standard database, the PlantVillage Kaggle dataset. More specifically, we have obtained the mAP and accuracy values of 0.981, and 99.97% respectively along with the test time of 0.23 s. Both qualitative and quantitative results confirm that the presented solution is robust to the detection of plant leaf disease and can replace the manual systems. Moreover, the proposed method shows a low-cost solution to tomato leaf disease classification which is robust to several image transformations like the variations in the size, color, and orientation of the leaf diseased portion. Furthermore, the framework can locate the affected plant leaves under the occurrence of blurring, noise, chrominance, and brightness variations. We have confirmed through the reported results that our approach is robust to several tomato leaf diseases classification under the varying image capturing conditions. In the future, we plan to extend our approach to apply it to other parts of plants as well.
Subject terms: Plant sciences, Diseases, Mathematics and computing
Introduction
According to a prediction by the Food and Agriculture Organization (FAO) of the United Nations, the world population could reach 9.1 billion by 2050, ultimately increasing food requirements1. Whereas the unavailability of large cultivation space and shortage of clean water is the major hindrance to the progression of food quantity. The need of the hour is to increase the growth rate of food in small spaces to cater to the needs of a large population. However, the presence of several plant diseases also causes a significant reduction in both food quality and quantity. Therefore, it is necessary to diagnose these plant abnormalities in the initial phase as these damages have a bad impact on the agriculturalist’s revenue and can also increase food prices. Its more devastating form can introduce starvation situations, especially in poor countries. Mostly, humans performed crop inspections by visiting the cultivation areas. However, visual inspection is a tedious task that is highly dependent on the availability of human resources. Secondly, human crop inspection is not much reliable as it is difficult for them to view individual plants2. Therefore, the identification and categorization of crop diseases at the early stage is a mandatory task for improving the food growth rate and saving the agriculturalists from expensive spray procedures. Therefore, to avoid manual inspection, researchers are designing automated systems for crop disease recognition3.
Although there exist a variety of crop plants, including Potato, Onions, Strawberry, Cheery, etc., however, the Tomato plant has the highest consumption rate with the range of 15 kg per capita in a year and counts for about 15% of the total vegetable consumption in all around the globe4. Therefore, the identification and classification of various tomato leaf diseases are the main focus of this study. Among the world's vegetable crop production, tomatoes rank first with an output of more than 170 million tons annually5. Various countries produce tomatoes globally each year, with India, Turkey, Egypt, and the United States being the leading producers6. According to a study conducted by the FOA, the major hindrance to the production rate of tomatoes is the presence of several diseases, where most of which start from the tomato plant leaves and results in to decrease in the production quantity from 8 to 10% every year7. Computer-aided systems can assist the practitioners in better identifying the tomato leaf diseases at the earliest stages to save the farmers from extensive losses. At first, the research community employed the approaches of molecular biology and immunology areas for tomato leaf disease inspection8,9. However, these works are suffering from high processing complexity and need a lot of human expertise. Mostly, crop cultivation is performed by poor people having a low source of income10, so, such expensive solutions are not feasible for them11. With the progression of machine learning (ML), researchers have applied several hand-crafted techniques in agricultural science12. The easier digital data acquisition methods have assisted to collect extensive data in real-time which is utilized by the ML-based methods to make intelligent decisions. Such approaches namely support vector machines (SVM)13, decision trees (DT)14, Gaussian frameworks15, and K-nearest neighbors (KNN)16,17 are also explored for inspecting plant diseases. These ML-based approaches18,19 are easier to understand and require a small amount of data for model training, and, however, these approaches are time-consuming and highly rely on skilled human resources. Moreover, traditional ML-based feature extraction approaches always require a trade-off with respect to processing complexity and recognition effectiveness20.
To deal with ML-based approaches' limitations, DL-based methods are mainly tested21. Nowadays, several DL methods i.e. CNN22, Recurrent neural networks (RNNs)23, and long short term memory (LSTM)24 are highly evaluated in agricultural science. These approaches are proficient in robustly computing the descriptive set of image features without the need for human expertise. DL methods mimic the way the human brain works as people localize and recognize several objects by viewing their samples, and so do these methods to perform pattern recognition. DL-based frameworks show robust performance in the area of agricultural science and suit well to numerous tasks and various types of deep neural networks (DNNs) show better accuracy over hyperspectral analysis25. Techniques like GoogLeNet26, AlexNet27, VGG28 and ResNet29 are heavily explored in agriculture for tasks like measuring food quantity, identifying crop heads, fruit counting, crop disease recognition, and classification, etc. These methods are empowered to show effective classification results while reducing the computational time due to their ability to exploit the topological data from the input images30.
Although extensive work has been presented by researchers for tomato plant leaf diseases, its identification at the early stages is still a complex task due to the extensive chrominance similarity between the affected and unaffected plant regions31. Moreover, the varying plant leaf sizes, light and intensity variations, presence of noise, and blurring in the suspected images have further complicated the detection process. Therefore, there is still room for performance enhancement in terms of plant disease recognition accuracy and processing time. For this reason, we have modified a DL-based approach named Faster-RCNN where we have introduced the ResNet-34 framework as a feature extractor for tomato plant disease localization and categorization. The proposed work has improved the classification performance with the mAP and accuracy values of 0.981, and 99.97% respectively, and minimized the testing time with the value of 0.23 s. The proposed approach is capable of detecting and classifying the tomato plant leaf disease classification performance under the occurrence of several image artifacts like blurring, noise, light variations, etc. We have presented an extensive performance evaluation comparison with the latest approaches to show its robustness and have confirmed through experimentation that the proposed solution shows an efficient and effective solution to plant leaf disease classification.
Related work
Here, we have presented an analysis of existing approaches employed in the field of agriculture to detect and classify numerous plant diseases.
Le et al.32 introduced a technique for classifying several plant leaf diseases. In the first step, a pre-processing step based on the morphological opening and closing methods was applied to eliminate the unwanted information from the suspected samples. After this, an approach namely the filtered local binary pattern method with contour mask and coefficient k (k-FLBPCM) was applied for keypoints extraction. The calculated feature vector was applied for the SVM classifier training to perform the classification. This work32 is robust to diagnosing plant leaf-related diseases with an accuracy of 98.63%, however, its performance decreases for distorted images. Similarly, a framework called Directional Local Quinary Patterns (DLQP) was presented in33 to calculate the descriptive set of sample features later employed for SVM training. The technique33 exhibits better performance for plant leaves abnormalities categorization with an accuracy of 97.80%, however, it shows degraded results for blurry input images. Sun et al.34 presented a solution for crop disease categorization. Initially, Simple Linear Iterative Cluster (SLIC) was used for dividing the input image into several blocks. Then, the fuzzy salient region contour and Gray Level Co-occurrence Matrix (GLCM) were applied for feature computation from the image blocks. The extracted keypoints vector was applied to train the SVM method to classify numerous plant diseases. The methodology34 demonstrates better plant leaf disease recognition results with an accuracy of 98.50%, however, it suffers from high computational cost. A similar method was introduced in2 where the GrabCut algorithm was used for image segmentation and the LBP method was used for extracting the sample features. Moreover, the SVM was trained over the extracted keypoints to categorize leaf abnormalities. The work presented in2 is effective for crop disease identification with an accuracy of 95%, however, it may not perform well for noisy images. Ramesh et al.35 presented a framework to achieve the classification of plant leaf diseases where the Histogram of Oriented Gradients (HOGs) approach was used to calculate the image features and the Random Forest (RF) classifier was used to perform the classification process. This approach presents a low-cost solution for plant leaf disease classification with an accuracy of 70.14%, however, it needs further classification performance enhancements. In36 a method was proposed to recognize the leaf diseases from the turmeric plant. Initially, the K-means technique was used for segmenting the suspected image sample. After this, the GLCM method was used for computing the image keypoints for the SVM training to execute the disease classification job. This technique36 works well for crop disease recognition with an accuracy of 91%, however, it is not robust to huge brightness changes in the suspected samples. An ML-based approach was introduced in37 for classifying several plant leaf diseases. Numerous techniques, namely, shift-invariant feature transform (SIFT), local binary pattern (LBP), and GLCM were applied to compute the keypoints of the suspected samples. The calculated features were exploited in the next step for training the SVM, KNN, and RF classifiers. This approach37 presents the best classification results with the RF classifier with an accuracy of 82.12%, however, the performance needs more improvements. In38 another framework was proposed to recognize plant diseases. In the first step, the fourteen color spaces were employed for computing the features from the input samples with a vector length of 172. Secondly, the extracted keypoints were utilized for the SVM classifier training with an accuracy of 94.65%. The method in38 demonstrates improved classification performance, however, it is unable to work well for the distorted samples.
Several researchers have explored the DL-based techniques for plant disease categorization and gained better results39. One of such works was presented in40 that employed a DL-based approach named as Few-Shot Learning (FSL) for locating the diseased area of plant leaves and specifying the associated class. For deep features computation, the Inception V3 method was applied while the SVM classifier was employed for the categorization task. The work presented40 shows better disease classification performance with an accuracy of 91.40%; however, the approach needs evaluation over a large and complex dataset. In41, a CNN-based network comprising of 3 convolution layers was proposed to calculate the image features and perform the classification task. This work gives a low-cost solution for the automated detection of crop diseases with an accuracy of 91.20%, however, it suffers from the problem of model over-fitting. Richey et al.42 proposed a lightweight plant disease detection approach that can be deployed to mobile phones. The technique was concerned to classify the maize crop disease. The ResNet50 framework was applied for feature computation and deciding the respective classes. The work demonstrated in42 performs well for plant disease recognition with an accuracy of 99%, however, not suitable for all types of cell phones due to the processing requirements.
Similarly in43, a work was proposed for tomato leaf disease identification. A deep feature extractor framework, AlexNet was used to compute the reliable set of image keypoints from the suspected sample which was later used to train the KNN classifier to perform the categorization job. The framework in43 is robust to tomato leaf disease classification with an accuracy of 76.10%, however, the KNN technique is a tedious and time-taking approach. Furthermore, another framework was presented in44 for tomato leaf disease classification where a residual framework was used for reliable keypoints extraction and a CNN-based approach was to categorize the diseased areas of plant leaves. The methodology in44 exhibits robust leaf disease recognition accuracy of 98%, however, at the expense of increased processing load. Dwivedi et al.45 presented a methodology named region-based CNN (RCNN) to recognize and categorize the several leaf diseases of grapes. Initially, the ResNet18 approach was used for calculating the reliable set of suspected image keypoints. The RCNN model later categorized the computed features into several classes. The work45 exhibits improved classification performance with a value of 99.93%, however, it exhibits lower performance to unseen examples. In46 several frameworks namely, VGG, ResNet, and DenseNet were used for feature computations from the input images. The work46 shows robust accuracy of 98.27% using the DenseNet approach, however, it suffers from a high economic burden. Although the research community has presented extensive work for the automated recognition and categorization of crop leaves contagious region categorization, however, still there is a need of performance improvement.
Saleem et al.47 employed several DL approaches like SSD, Faster-RCNN, and Region-based Fully Convolutional Networks (RFCN) to identify and categorize the diseased portions of plants. The work shows the highest mAP score of 73.07%. However, unable to identify the affected regions of very small sizes from the leaf samples. Zhao et al.48 proposed an approach to classify several tomato plant leaf diseases where the convolutional neural network that integrates an attention mechanism was applied. The approach acquired an accuracy value of 99.24%. Similarly in30 several DL-based approaches like AlexNet, GoogleNet, Inception V3, ResNet-18, and ResNet-50 were evaluated to categorize the tomato plant leaf samples into several classes. The work attains the highest results for the GoogleNet model with an accuracy of 99.39%. Moreover, Bhujel et al.49 proposed a DL-based approach namely the ResNet18 with CBAM for tomato plant leaf disease classification and acquires an accuracy of 99.69%. The works presented in30,48,49 improve the tomato plant leaf diseases classification performance, however, these methods perform the image level classification and are unable to locate the exact diseased region which is an important aspect for the automation of plant abnormality detection methods. To cover the limitations of such works, we have presented an object detection approach capable of detecting and classifying the diseased regions of tomato plant leaves with a small number of model parameters to reduce the computational complexity. Moreover, we are focused on presenting a model that can detect the affected areas of varying shapes and sizes.
A review of work from history used for the identification and categorization of several plants leaves abnormalities is presented in Table 1. Although the research community has presented extensive work for the automated recognition and categorization of crop leaves contagious region categorization, still there is a need for performance improvement.
Table 1.
Review of existing techniques.
| Reference | Technique | Limitation |
|---|---|---|
| ML-based methods | ||
| 32 | K-FLBPCM with the SVM classifier | The work is not robust to distorted samples |
| 33 | DLQP with the SVM classifier | The work does not perform well with blurry images |
| 34 | GLCM with the SVM classifier | The approach is computationally complex |
| 2 | LBP with the SVM classifier | The method is not effective in noisy images |
| 35 | HOGs with the RF classifier | The method needs further performance enhancement |
| 36 | K-means + GLCM with the SVM classifier | The work is not robust to huge brightness changes in the suspected samples |
| 37 | SIFT, LBP, GLCM, along with the SVM, KNN, and RF classifiers | The work needs further classification accuracy improvements |
| 38 | Fourteen color spaces along with the SVM classifier | The work is not workable with distorted images |
| DL-based methods | ||
| 40 | Inception V3 method along with the SVM classifier | The method requires evaluation over a large and complex dataset |
| 41 | CNN | The work is suffering from the model over-fitting |
| 42 | ResNet50 | The work is computationally expensive |
| 43 | AlexNet along with the KNN classifier | The method is slow and time-consuming |
| 44 | CNN | The technique requires extensive data for model training |
| 45 | RCNN | The work is not generalized well to real-world examples |
| 46 | VGG, ResNet and DenseNet | The approach is economically inefficient |
| 47 | SSD, Faster-RCNN, RFCN | The work is unable to identify the affected regions of very small sizes |
| 48 | CNN with an attention mechanism | The work is computationally expensive and unable to locate the exact diseased region on the test sample |
| 30 | AlexNet, GoogleNet, Inception V3, ResNet-18, and ResNet-50 | The work is computationally expensive and performs the image level classification |
| 49 | ResNet18 with CBAM | The method cannot identify the exact location of the diseased area from the suspected images |
Contributions
To cover the existing issues of tomato plant leaf disease classification, we have presented a DL-based approach named Faster-RCNN with ResNet-34 as a feature extractor for tomato plant disease localization and categorization. Following are the main contribution of our work:
Modified an object detection technique namely Faster-RCNN for plant disease classification which enhanced the classification accuracy to 99.97%.
We introduce a cost-effective approach to improve the tomato plant disease classification performance while reducing the testing time with a value of 0.23 s.
Robust detection of the affected regions of plant leaves even under the existence of noise, blurring, color, size, and light alterations due to the efficacy of the Faster-RCNN technique to deal with the over-fitted training data.
The presented framework is empowered to correctly localize the disease area of plant leaves with the mAP score of 0.981 because of the reliable feature extraction capability of the RESNET-34-based Faster-RCNN model.
A thorough analysis has been conducted in comparison to other approaches to plant disease classification methods over a standard database namely PlantVillage to demonstrate the efficacy of the presented work.
Materials and methods
In the presented framework, we have introduced an approach for identifying and categorizing several tomato plant leaf diseases from the suspected images via employing a DL-based approach namely the Faster-RCNN with the ResNet-34 as the base network. Initially, annotations are created to specify the affected region from the input images. The annotated samples are passed to the Faster-RCNN framework where the ResNet-34 base network computes the deep features which are localized and classified by the Faster-RCNN framework. Finally, the trained model is evaluated on the test data to identify and classify several tomato plant diseases. The entire workflow of the introduced solution is exhibited in Fig. 1. Our evaluation results confirm that Faster-RCNN with ResNet-34 feature extractor is robust to tomato plant disease classification due to its high recall ability.
Figure 1.
Workflow representation of the introduced approach.
Material
To assess the classification power of the introduced solution, we have used the PlantVillage dataset50, a standard and freely available database. The PlantVillage dataset contains a total of 54,306 samples from 14 types of plants. We have considered the images of tomato plants from the PlantVillage dataset with 10 specified classes for our work. The main motivation for using the PlantVillage database is that it contains images with intense changes in the dimensions, color, and location of the diseased areas.
Moreover, samples are suffering from noise, blurring, and light intensity variations. A detailed description of tomato images is given in Fig. 2 while a few samples are shown in Fig. 3.
Figure 2.
Detailed description of the employed dataset.
Figure 3.
Samples of the PlantVillage dataset.
Ground truth generation
To have a robust and accurate training procedure, it is essential to determine the affected portion from the suspected samples correctly. To accomplish this, we have used LabelImg41, an open-source software to develop plant images' annotations to locate the RoIs. The final output containing the coordinate values of affected regions is saved in a CSV file which is later passed along with the input images to the Faster-RCNN framework for model training. Some examples of annotated samples are exhibited in Fig. 4.
Figure 4.
Annotated image samples.
Faster-RCNN
In the introduced framework, we have employed Faster-RCNN, a DL-based method for the automated localization and categorization of several plant leaf abnormalities. The Faster-RCNN approach utilizes the convolve-filters which gives it the ability to evaluate the structure of the suspected sample and extract a reliable set of key points from it. We have selected the Faster-RCNN framework over the RCNN and Fast-RCNN models for tomato plant disease classification as these methods are computationally more complex. Moreover, RCNN and Fast-RCNN models employ handcrafted approaches i.e. EdgeBox51 or selective search52, etc. for feature computation which in turn is unable to learn a discriminative set of image features. Whereas, the Faster-RCNN approach robustly resolves the issues of the RCNN and the Fast-RCNN approaches via introducing a separate module namely the Regional Proposal Network (RPN) for the automated extraction of features from the input image.
In the area of crop leaves disease localization and classification, identifying the ROIs from the suspected samples are suffering from two main challenges: (1) Exactly identifying the affected regions on the leaves since there is a huge color similarity between affected and healthy regions. (2) Correctly identifying the class associated with each identified region.
The selected framework namely Faster-RCNN can better tackle the above-mentioned challenges. As, the RPN module of Faster-RCNN has enabled it to better locate the ROIs and effectively classify the suspected samples as it employs the information of diseased portion size, color, and texture and guarantees a higher recall rate via using a few selected windows.
Custom faster-RCNN
The conventional Faster-RCNN framework employed either VGG-16 or ResNet-101 as the base networks to calculate the reliable keypoints of the suspected sample. However, both these models comprise an extensive number of model parameters which in turn increase the computational cost of the Faster-RCNN. Moreover, such structure of the Faster-RCNN results in the vanishing gradient issue. To tackle the existing problems of the Faster-RCNN model, we have presented a modified version by presenting the ResNet-34 along with the CBAM as the feature extraction module. The presented ResNet-34 approach contains fewer model parameters, giving it a computational advantage over the existing base networks. The detailed demonstration of the trainable parameters used for the proposed method is explained in Table 2. The introduced work was implemented in Python language with Tensorflow and Keras libraries and executed on an Nvidia GTX1070 GPU-based system in Windows 10 environment.
Table 2.
Details of trainable parameters used by the introduced work.
| Network parameters | Value |
|---|---|
| Total epochs | 20 |
| Learning rate | 0.001 |
| Employed batch size | 8 |
| Value of threshold for the matched region | 0.2 |
| Value of threshold for the unmatched areas | 0.5 |
The Faster-RCNN approach comprises four steps: feature extraction module, Region proposal networks (RPN), RoI pooling, and classification to locate and classify several plant diseases.
ResNet-34 based features extractor
A backbone network is usually a CNN model that is responsible for extracting feature maps that present semantic and robust representations of an input image. These feature maps are used for detecting regions of interest and classification in several object detection algorithms. Intuitively, the more robust the computed features of the backbone network are, the higher the detection performance53. In this study, we employed ResNet-3454 along with Convolutional Block Attention Module (CBAM)55 for improved feature extraction in the proposed Faster RCNN model, which ultimately improved the classification performance for plant disease identification tasks. ResNet54 is the latest CNN model that utilizes identity shortcut connections and residual mapping across layers to achieve high accuracy. In typical deeper networks, each layer feeds its output into the next layer and thus can compute more complex features that improve the model's robustness and performance56. However, in these networks, performance abruptly degrades with the increase in the network depth due to gradient vanishing problems during training. To address this issue, the ResNet model introduced skip connections in deep networks that bypass one or more layers and form the basis for residual blocks. The resulting architecture allows reusing the feature maps from the preceding layers, which provides better accuracy and easier training. The detailed structure of the ResNet-34 model is presented in Table 3. The model contains 33 convolutions (Conv) layers in total, which are grouped into 5 Conv stages, each consisting of several residual blocks placed on top of one another. Figure 5 depicts the structure details of the individual residual blocks. A residual block comprises several Conv layers, a ReLU activation method, a batch normalization layer, and a shortcut connection. In residual block, the stacked layers conduct residual mapping by establishing shortcut connections that perform locating mapping (x). Their results are combined with the stacked layers' output residual function F (x). The resultant value of the residual block can be expressed as:
| 1 |
Here, x is the input, F represents the residual function and Z is the output of the residual function.
Table 3.
Architecture details of ResNet-34 original and modified.
| Layer name | Original | Modified |
|---|---|---|
| Conv1 | 7 × 7, 64, 3 × 3 max pool | × 3, 7 × 7 Attention |
| Conv2_x | ||
| Conv3_x | ||
| Conv4_x | ||
| Conv5_x |
Figure 5.

The architectural representation of the basic building block ‘Residual Block’ in ResNet.
We modified the architecture of ResNet-34 by adding CBAM based attention block55 at the initial of the network. The motivation to add the CBAM block is that it improves the representation of features by employing an attention mechanism. The attention block assists the network to focus on the disease-effected locations while suppressing irrelevant background information and improving recognition performance under varying challenging conditions, such s color, light, and intensity variations. The CBAM module refines the CNN-derived features by integrating spatial and channel-wise attention and thus enhances DNN performance. Due to its lightweight architecture, the CBAM block adds a negligible overhead and can be trained end-to-end along with base CNNs. Table 3 presents the architecture of a modified backbone network with an attention block. We further replaced the early 7 × 7 Conv and max-pooling layer with three stacked 3 × 3 Conv layers to avoid downsampling steps in the early convolutional layer57. Furthermore, to decrease computational costs, the channel for newly added Conv layers is set to 64.
Region proposal networks (RPN)
The RPN unit comprises 3 × 3 convolutional layers which create the anchors and bounding boxes from which object proposals are computed.
RoI pooling
The ROI pooling layer employs the result of both the convolution layers and the RPN module to calculate the proposal feature maps which are later passed as input to all fully connected layers. This phase includes employing the result of the preceding two layers by utilizing the feature map and object proposals to compute the proposal feature maps which are later fed to fully connected layers.
Classification
In the last step, the Faster-RCNN module performs localization and classification by generating the bounding boxes to identify the diseased region of the tomato crop and determine the associated category.
The detailed steps of the proposed solution are presented in Algorithm 1.
Results
In this part, we have discussed the description of the employed dataset and explained the obtained results in detail.
Evaluation metrics
The evaluation metrics namely accuracy, mean average precision (mAP), intersection over union (IOU), precision, and recall are used to evaluate the robustness of our work which are defined in Eqs. (2) and (3) and Fig. 6, respectively.
| 2 |
| 3 |
In Eq. (3), AP denotes the average precision of each class and t is the query or test image. T is the total number of test samples.
Figure 6.
Visual demonstration of IOU, precision, and recall.
Localization results
The main characteristic of an automated plant disease system is that it should be capable of localizing the various categories of plant diseases accurately. Therefore, to evaluate the detection performance of the introduced solution, the tomato images from the employed database and show some of the visual results in Fig. 7.
Figure 7.
Localization results.
From the reported results, it can be analyzed that the presented approach is robust to classify the samples of varying classes under the occurrence of several image transformations like under the presence of leaf size, color, and diseased portion position variations.
The recognition ability of the Faster-RCNN model allows it to locate and categorize numerous tomato plant leaf diseases correctly. To numerically demonstrate the localization supremacy of the proposed method, two standard metrics namely mAP and IOU are explored. These measures assist to check the detection and classification enactment of the technique for numerous categories of tomato plant anomalies. More specifically, we obtain the mAP score of 0.981and IOU values of 0.973, respectively. It is quite visible from the results that our presented framework can be consistently used to identify and categorize tomato plant diseases.
Classification results
To design an accurate and robust tomato plant leaf disease classification model, it should be able to differentiate between the various classes of plant abnormalities. For this reason, we have designed an experiment to discuss the classification results of our approach in detail. In this section, we have demonstrated the category-wise identification and classification results of ResNet-34-based Faster-RCNN in the form of precision, recall, F1-score, and an error rate. To evaluate the classification performance of the proposed ResNet-34-based Faster-RCNN on ten classes, we have drawn the boxplots (Figs. 8, 9) to show the obtained precision and recall rates as these plots have the ability to better show the results by exhibiting the minimum, maximum, median values along with the symmetry and skewness of the data. Figures clearly show that our technique can detect and classify the different types of tomato plant leaf diseases.
Figure 8.
Evaluation using Precision.
Figure 9.
Proposed method evaluation using Recall.
Furthermore, the F1-score and the error rate for the entire dataset is shown in Fig. 10. More clearly, we have obtained an average F1-score of 99.42% along with the minimum and maximum error rates of 0% and 0.35%, respectively.
Figure 10.
Evaluation using F1 score and error rate.
It is quite evident from the reported results that our approach has better recall ability and can better generalize to tomato plant diseases of several types.
To further elaborate on the recognition accuracy of our approach, we have analyzed to show the class-wise classification performance of our approach, and the results are shown in Fig. 11. More clearly, for all the ten classes of tomato plant leaf diseases from the PlantVillage dataset, the ResNet-34-based Faster-RCNN approach acquires the average accuracies of 99.96%, 99.97%, 100%, 99.95%, 99.96%, 99.98%, 99.95%, 99.97%, 99.96%, and 99.97% respectively which is showing the robustness of our approach with high recall rate.
Figure 11.
Line plot of accuracy measure.
We further demonstrate the obtained results for each class of tomato plant leaf disease, we have shown the confusion matrix (Fig. 12) as it better explains the results by showing the real and predicted class. Figure 12 clearly shows that the presented custom Faster-RCNN can effectively differentiate all classes of tomato plant diseases. So, from the entire result discussion, it can be concluded that the introduced methodology is more effective to both tomato plant leaf diseases portion identification and categorization.
Figure 12.
Confusion matrix of our method.
Comparison with base approaches
We have analyzed the robustness of our framework against several DL-based models namely GoogleNet, ResNet-101, Xception, VGG-19, and SE-ResNet50 via an experiment. The obtained results are shown in Table 4. From the reported results in Table 4, it can be viewed that our approach outperforms the base techniques. More specifically, the GoogleNet approach shows the minimum results with the average precision, recall, F1-score, and accuracy values of 87.16%, 87.09%, 87.12%, and 87.27% respectively. While the second minimum values are attained by the Xception model with the average precision, recall, F1-score, and accuracy values of 88.25%, 88.14%, 88.19%, and 88.16% respectively. While in comparison, the proposed improved Faster-RCNN approach with ResNet-34 shows the highest performance with the average precision, recall, F1-score, and accuracy values of 99.48%, 99.32%, 99.42%, and 99.97%, respectively. More clearly, in the case of precision metrics, the comparative base models show an average value of 90.50%, which is 99.48% for our method, so we have attained a performance gain of 8.97%. While for the recall metric, the comparative approaches show an average value of 90.53%, while our method shows an average recall value of 99.42%. Therefore, in the case of the recall evaluation metric, we have shown an average performance gain of 8.89%. Moreover, for the F1-Score and accuracy metrics, the base models show the average values of 90.91%, and 90.56% respectively, while the proposed solution shows the average values of 99.42%, and 99.97% respectively. So, for the F1-Score and accuracy metrics, the proposed solutions show performance gains of 8.51%, and 9.41% respectively. From the performed analysis, it can be concluded that our work is more effective for both tomato plant leaf disease detection and classification. The major cause of the efficient classification accuracy of the proposed approach is that it uses the ResNet-34 model with CBAM as the feature extractor module which extracts a more robust set of image keypoints by removing the redundant data and minimizing the network complexity which in turn makes the model prone to model overfitting problems.
Table 4.
Comparative analysis of our approach with the base models.
| Model | Precision (%) | Recall (%) | F1-score (%) | Accuracy (%) |
|---|---|---|---|---|
| GoogleNet | 87.16 | 87.09 | 87.12 | 87.27 |
| ResNet-101 | 89.95 | 90.13 | 90.04 | 90.13 |
| Xception | 88.25 | 88.14 | 88.19 | 88.16 |
| VGG-19 | 90.39 | 90.47 | 92.43 | 90.42 |
| SE-ResNet50 | 96.77 | 96.81 | 96.79 | 96.81 |
| Proposed | 99.48 | 99.38 | 99.42 | 99.97 |
Comparative evaluation with DL-based methods
We performed an analysis to compare the performance of the presented approach with other object detection techniques employed for the plant leaf disease classification. All employed approaches have been evaluated for several situations, i.e., a case where several plant leaves are present in a single suspected sample or to check their performance for different plant classes. The main reason to perform this simulation is to check how accurate our system is compared to other DL-based techniques in locating the diseased leaf portion from the healthy parts of image leaves and differentiating abnormalities of the several plant types.
For this reason, we have selected the approaches namely Fast-RCNN58, Faster-RCNN59. You Only Look Once (YOLO)60, and single-shot detector (SSD)61. To perform the evaluation, we have chosen the mAP evaluation metric, a standard and most widely employed performance measure for object detection techniques by the research community. The results are shown in Table 5, clearly demonstrating that our technique has acquired the highest mAP value of 0.981. The SSD approach exhibits a lower mAP value of 0.830. While the second-lowest performance is shown by the YOLOv3 model with the mAP value of 0.830. As, these both methods are unable to locate the diseased portion of small sizes, therefore their detection accuracy dropped for such scenarios.
Table 5.
Comparative analysis of the presented method DL-based approaches.
| Models | mAP |
|---|---|
| Fast-RCNN | 0.860 |
| Faster-RCNN | 0.884 |
| YOLOv3 | 0.842 |
| SSD | 0.830 |
| Proposed Faster-RCNN (RESNET-34) | 0.981 |
While the Faster-RCNN with the base VGG-16 network shows better results with the mAP value of 0.884, this approach is computationally inefficient due to its extensive number of model parameters. On the other hand, the proposed method, ResNet-34-based Faster-RCNN has better tackled the limitations of the comparative methods as its feature extractor allows it to better compute the image features which allows it to learn the image transformations better. More specifically, the comparative DL-based object detection models give an average mAP value of 0.854, which is 0.981 in our case, so our method gives an average performance gain of 12.7%. Furthermore, the proposed Faster-RCNN contains fewer model parameters compared to other approaches, which also gives it a computational advantage.
Proposed work evaluation with the state-of-the-art methods
To further measure the detection performance of our approach, we have taken the latest approaches30,48,49,62–64 and compared our results with them. We have used three metrics for performance analysis, precision, recall, and accuracy, and comparison evaluations reported in Table 6. It can be seen from Table 6 that our work performs well than the comparative approaches. The reason for the efficient performance of our approach is that the methods30,48,49,62–64 employ very complex network architectures which result in model overtraining. Whereas the proposed solution deploys a shallow network architecture as it uses the ResNet-34 model with CBAM which is more empowered to compute the more distinguishing set of image keypoints which assists in attaining robust performance. Moreover, the lightweight architecture of the model further assists to avoid the occurrence of the model overfitting problem. Such architecture of the proposed approach assists in better tackling the image transformation issues like the variations in the light, color, size, and position of leaves and improves the accuracy of tomato plant disease classification.
Table 6.
Performance analysis with the latest approaches.
Conclusions
The manual detection and classification of the various plant leaf diseases require the expertise of humans to identify the small details from the suspected samples. Furthermore, the intense variations in the size, color, and structure of plants further complicate the categorization procedure. To deal with such challenges, we have proposed a robust approach namely ResNet-34 based Faster-RCNN to detect and classify the diseases of tomato plant leaves. More specifically, we employed the ResNet-34 along with the CBAM as the base backbone of the Faster-RCNN model to compute the deep features from the input samples. Then, the extracted features are used to train the Faster-RCNN framework to locate and categorize the suspected samples into ten related classes. The results are evaluated on the tomato leaf diseased images taken from a challenging dataset, the PlantVillage database. Both the qualitative and quantitative results confirm that the presented solution is robust to plant leaf disease detection which can replace the manual systems. Moreover, the proposed method shows a low-cost solution to tomato leaf disease classification which is robust to several image transformations like the variations in the size, color, and orientation of the leaf diseased portion. Furthermore, the framework can locate the affected plant leaves under the occurrence of blurring, noise, chrominance, and brightness variations. As the existing work is concerned to locate the disease portion from the plant leaves only, therefore, in the future, we plan to extend our approach to apply it to other parts of plants as well.
Acknowledgements
The authors would like to thank the editors and the anonymous reviewers for their valuable comments and suggestions.
Author contributions
Conceptualization, M.N., T.N., A.J., M.M., J.R.; data curation, M.N., T.N., J.R.; funding acquisition J.R., J.K.; investigation, T.N., J.R., J.K., A.H.; methodology, M.N., T.N., A.J., M.M., J.R.; software, M.N., T.N., A.J., M.M., J.R., J.K., A.H.; validation, M.N., T.N., J.R.; writing—original draft, M.N., T.N., J.R.; writing—review and editing, J.R., J.K., A.H. All authors read and approved the final manuscript.
Funding
This work was partly supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2021R1A4A1031509) and by the UK Engineering and Physical Sciences Research Council (EPSRC) - Grants Ref. EP/M026981/1, EP/T021063/1, EP/T024917/1.
Data availability
The data generated to support the findings of this study are available from the corresponding author upon reasonable request.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Junaid Rashid, Email: junaidrashid062@gmail.com.
Jungeun Kim, Email: jekim@kongju.ac.kr.
References
- 1.Bruinsma, J. The resource outlook to 2050: by how much do land, water and crop yields need to increase by 2050. In Expert meeting on how to feed the world in (2009).
- 2.Pantazi XE, Moshou D, Tamouridou AA. Automated leaf disease detection in different crop species through image features analysis and One Class Classifiers. Comput. Electron. Agric. 2019;156:96–104. doi: 10.1016/j.compag.2018.11.005. [DOI] [Google Scholar]
- 3.Wolfenson KDM. Coping with the Food and Agriculture Challenge: Smallholders’ Agenda. Food Agriculture Organisation of the United Nations; 2013. [Google Scholar]
- 4.Chowdhury ME, et al. Automatic and reliable leaf disease detection using deep learning techniques. AgriEngineering. 2021;3(2):294–312. doi: 10.3390/agriengineering3020020. [DOI] [Google Scholar]
- 5.Valenzuela, M. E. M. & Restović, F. Valorization of tomato waste for energy production. In Tomato Chemistry, Industrial Processing and Product Development, 245–258 (2019).
- 6.Elnaggar, S., et al. Current status of bacterial wilt (Ralstonia solanacearum) disease in major tomato (Solanum lycopersicum L.) growing areas in Egypt. Arch. Agric. Environ. Sci. 2018; 3(4): p. 399–406.
- 7.Sardogan, M., Tuncer, A., & Ozen, Y. Plant leaf disease detection and classification based on CNN with LVQ algorithm. In 3rd International Conference on Computer Science and Engineering (UBMK) (IEEE, 2018).
- 8.Sankaran S, et al. A review of advanced techniques for detecting plant diseases. Comput. Electron. Agric. 2010;72(1):1–13. doi: 10.1016/j.compag.2010.02.007. [DOI] [Google Scholar]
- 9.Dinh HX, et al. Molecular genetics of leaf rust resistance in wheat and barley. Theor. Appl. Genet. 2020;133(7):2035–2050. doi: 10.1007/s00122-020-03570-8. [DOI] [PubMed] [Google Scholar]
- 10.Ferentinos KP. Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 2018;145:311–318. doi: 10.1016/j.compag.2018.01.009. [DOI] [Google Scholar]
- 11.Patil S, Chandavale A. A survey on methods of plant disease detection. Int. J. Sci. Res. 2015;4(2):1392–1396. [Google Scholar]
- 12.Gebbers R, Adamchuk VI. Precision agriculture and food security. Science. 2010;327(5967):828–831. doi: 10.1126/science.1183899. [DOI] [PubMed] [Google Scholar]
- 13.Joachims, T. Making large-scale SVM learning practical. Technical report (1998).
- 14.Rokach, L. & Maimon, O. Decision trees, in Data Mining and Knowledge Discovery Handbook. 165–192 (Springer, 2005).
- 15.Birgé L, Massart P. Gaussian model selection. J. Eur. Math. Soc. 2001;3(3):203–268. doi: 10.1007/s100970100031. [DOI] [Google Scholar]
- 16.Rashid J, et al. An augmented artificial intelligence approach for chronic diseases prediction. Front. Public Health. 2022;10:860396. doi: 10.3389/fpubh.2022.860396. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Batool, S., et al. A random forest students’ performance prediction (RFSPP) model based on students’ demographic features. In 2021 Mohammad Ali Jinnah University International Conference on Computing (MAJICC) (IEEE, 2021).
- 18.Zaka-ud-din W, et al. Classification of disease in tomato plants’ leaf using image segmentation and SVM. Tech. J. Univ. Eng. Technol. 2018;23(2):81–88. [Google Scholar]
- 19.Rashid J, et al. An efficient topic modeling approach for text mining and information retrieval through K-means clustering. Mehran Univ. Res. J. Eng. Technol. 2020;39(1):213–222. doi: 10.22581/muet1982.2001.20. [DOI] [Google Scholar]
- 20.Bello-Cerezo R, et al. Comparative evaluation of hand-crafted image descriptors vs. off-the-shelf CNN-based features for colour texture classification under ideal and realistic conditions. Appl. Sci. 2019;9(4):738. doi: 10.3390/app9040738. [DOI] [Google Scholar]
- 21.Rashwan SA, Elteir MK. Plant leaf disease detection using deep learning on mobile devices. Int. J. Comput. Vis. Robot. 2022;12(2):156–176. doi: 10.1504/IJCVR.2022.121151. [DOI] [Google Scholar]
- 22.Roska T, Chua LO. The CNN universal machine: an analogic array computer. IEEE Trans. Circuits Syst. II Analog Digital Signal Process. 1993;40(3):163–173. doi: 10.1109/82.222815. [DOI] [Google Scholar]
- 23.Zaremba, W., Sutskever, I., & Vinyals, O. Recurrent neural network regularization. arXiv preprint (2014).
- 24.Salakhutdinov, R., & Hinton, G. Deep boltzmann machines, in Artificial Intelligence and Statistics (PMLR, 2009).
- 25.Gewali, U. B., Monteiro, S. T., & Saber, E. Machine learning based hyperspectral image analysis: a survey. arXiv preprint. https://arxiv.org/abs/1802.08701 (2018).
- 26.Szegedy, C., et al. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2015).
- 27.Yuan, Z.-W., & Zhang, J. Feature extraction and image retrieval based on AlexNet. In Eighth International Conference on Digital Image Processing (ICDIP) (International Society for Optics and Photonics, 2016).
- 28.Vedaldi A, Zisserman A. Vgg Convolutional Neural Networks Practical 66. University of Oxford; 2016. [Google Scholar]
- 29.Thenmozhi K, Reddy US. Crop pest classification based on deep convolutional neural network and transfer learning. Comput. Electron. Agric. 2019;164:104906. doi: 10.1016/j.compag.2019.104906. [DOI] [Google Scholar]
- 30.Maeda-Gutiérrez V, et al. Comparison of convolutional neural network architectures for classification of tomato plant diseases. Appl. Sci. 2020;10(4):1245. doi: 10.3390/app10041245. [DOI] [Google Scholar]
- 31.Paul, A., et al. A review on agricultural advancement based on computer vision and machine learning, in Emerging Technology in Modelling and Graphics. 567–581 (Springer, 2020).
- 32.Le, V. N. T., et al. A novel method for detecting morphologically similar crops and weeds based on the combination of contour masks and filtered Local Binary Pattern operators. GigaScience 9(3), giaa017 (2020). [DOI] [PMC free article] [PubMed]
- 33.Ahmad W, Shah S, Irtaza A. Plants disease phenotyping using quinary patterns as texture descriptor. KSII Trans. Internet Inf. Syst. 2020;14(8):3312–3327. [Google Scholar]
- 34.Sun Y, et al. SLIC_SVM based leaf diseases saliency map extraction of tea plant. Comput. Electron. Agric. 2019;157:102–109. doi: 10.1016/j.compag.2018.12.042. [DOI] [Google Scholar]
- 35.Ramesh, S., et al. Plant disease detection using machine learning. In International Conference on Design Innovations for 3Cs Compute Communicate Control (ICDI3C) (IEEE, 2018).
- 36.Kuricheti, G., & Supriya, P. Computer vision based turmeric leaf disease detection and classification: a step to smart agriculture. In 3rd International Conference on Trends in Electronics and Informatics (ICOEI) (IEEE, 2019).
- 37.Kaur NJ, Education M. Plant leaf disease detection using ensemble classification and feature extraction. Turk. J. Comput. Math. Educ. 2021;12(11):2339–2352. [Google Scholar]
- 38.Shrivastava VK, Pradhan MK. Rice plant disease classification using color features: a machine learning paradigm. J. Plant Pathol. 2021;103(1):17–26. doi: 10.1007/s42161-020-00683-3. [DOI] [Google Scholar]
- 39.Walter, M. J.g.g.X.S., Is this the end? Machine learning and 2 other threats to radiologys future. p. l3 (2016).
- 40.Argüeso D, et al. Few-shot learning approach for plant disease classification using images taken in the field. Comput. Electron. Agric. 2020;175:105542. doi: 10.1016/j.compag.2020.105542. [DOI] [Google Scholar]
- 41.Agarwal M, et al. ToLeD: tomato leaf disease detection using convolution neural network. Procedia Comput. Sci. 2020;167:293–301. doi: 10.1016/j.procs.2020.03.225. [DOI] [Google Scholar]
- 42.Richey, B., et al. Real-time detection of maize crop disease via a deep learning-based smartphone app. In Real-Time Image Processing and Deep Learning 2020 (International Society for Optics and Photonics, 2020).
- 43.Batool, A., et al. Classification and identification of tomato leaf disease using deep neural network. In International Conference on Engineering and Emerging Technologies (ICEET) (IEEE, 2020).
- 44.Karthik R, et al. Attention embedded residual CNN for disease detection in tomato leaves. Appl. Soft Comput. 2020;86:105933. doi: 10.1016/j.asoc.2019.105933. [DOI] [Google Scholar]
- 45.Dwivedi R, et al. Grape disease detection network based on multi-task learning and attention features. IEEE Sens. J. 2021;21:17573–17580. doi: 10.1109/JSEN.2021.3064060. [DOI] [Google Scholar]
- 46.Akshai, K., & Anitha, J. Plant disease classification using deep learning. In 3rd International Conference on Signal Processing and Communication (ICPSC) (IEEE, 2021).
- 47.Saleem MH, et al. Image-based plant disease identification by deep learning meta-architectures. Plants. 2020;9(11):1451. doi: 10.3390/plants9111451. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Zhao S, et al. Tomato leaf disease diagnosis based on improved convolution neural network by attention module. Agriculture. 2021;11(7):651. doi: 10.3390/agriculture11070651. [DOI] [Google Scholar]
- 49.Bhujel A, et al. A lightweight attention-based convolutional neural networks for tomato leaf disease classification. Agriculture. 2022;12(2):228. doi: 10.3390/agriculture12020228. [DOI] [Google Scholar]
- 50.Hughes, D., & Salathé, M. An open access repository of images on plant health to enable the development of mobile disease diagnostics. arXiv preprint. https://arxiv.org/abs/1511.08060 (2015).
- 51.Uijlings JR, et al. Selective search for object recognition. Int. J. Comput. Vis. 2013;104(2):154–171. doi: 10.1007/s11263-013-0620-5. [DOI] [Google Scholar]
- 52.Dollár P, Zitnick CL. Fast edge detection using structured forests. IEEE Trans. Pattern Anal. Mach. Intell. 2014;37(8):1558–1570. doi: 10.1109/TPAMI.2014.2377715. [DOI] [PubMed] [Google Scholar]
- 53.Amjoud, A. B. & Amrouch, M. Convolutional neural networks backbones for object detection. In International Conference on Image and Signal Processing (Springer, 2020).
- 54.He, K., et al. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016).
- 55.Woo, S., et al. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV) (2018).
- 56.Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv preprint. https://arxiv.org/abs/1409.1556 (2014).
- 57.Zhu, R., et al. ScratchDet: training single-shot object detectors from scratch. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2019).
- 58.Girshick, R. Fast R-Cnn. In Proceedings of the IEEE International Conference on Computer Vision (2015).
- 59.Ren S, et al. Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016;39(6):1137–1149. doi: 10.1109/TPAMI.2016.2577031. [DOI] [PubMed] [Google Scholar]
- 60.Redmon, J. & Farhadi, A. Yolov3: an incremental improvement. arXiv preprint. https://arxiv.org/abs/1804.02767 (2018).
- 61.Liu, W., et al. Ssd: Single shot multibox detector. In European Conference on Computer Vision (Springer, 2016).
- 62.Agarwal M, Gupta SK, Biswas K. Development of Efficient CNN model for Tomato crop disease identification. Sustain. Comput. Inform. Syst. 2020;28:100407. [Google Scholar]
- 63.Tm, P., et al. Tomato leaf disease detection using convolutional neural networks. In Eleventh International Conference on Contemporary Computing (IC3) (IEEE, 2018).
- 64.Kaur, M. & Bhatia, R. Development of an improved tomato leaf disease detection and classification method. In IEEE Conference on Information and Communication Technology (IEEE, 2019).
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data generated to support the findings of this study are available from the corresponding author upon reasonable request.











