Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2019 Jan 14;19(2):313. doi: 10.3390/s19020313

Development of a Recognition System for Spraying Areas from Unmanned Aerial Vehicles Using a Machine Learning Approach

Pengbo Gao 1, Yan Zhang 1, Linhuan Zhang 1, Ryozo Noguchi 2, Tofael Ahamed 2,*
PMCID: PMC6359728  PMID: 30646586

Abstract

Unmanned aerial vehicle (UAV)-based spraying systems have recently become important for the precision application of pesticides, using machine learning approaches. Therefore, the objective of this research was to develop a machine learning system that has the advantages of high computational speed and good accuracy for recognizing spray and non-spray areas for UAV-based sprayers. A machine learning system was developed by using the mutual subspace method (MSM) for images collected from a UAV. Two target lands: agricultural croplands and orchard areas, were considered in building two classifiers for distinguishing spray and non-spray areas. The field experiments were conducted in target areas to train and test the system by using a commercial UAV (DJI Phantom 3 Pro) with an onboard 4K camera. The images were collected from low (5 m) and high (15 m) altitudes for croplands and orchards, respectively. The recognition system was divided into offline and online systems. In the offline recognition system, 74.4% accuracy was obtained for the classifiers in recognizing spray and non-spray areas for croplands. In the case of orchards, the average classifier recognition accuracy of spray and non-spray areas was 77%. On the other hand, the online recognition system performance had an average accuracy of 65.1% for croplands, and 75.1% for orchards. The computational time for the online recognition system was minimal, with an average of 0.0031 s for classifier recognition. The developed machine learning system had an average recognition accuracy of 70%, which can be implemented in an autonomous UAV spray system for recognizing spray and non-spray areas for real-time applications.

Keywords: precision agriculture, recognition system, image classifiers, machine learning system, mutual subspace method

1. Introduction

With the development of unmanned aerial vehicle (UAV) technologies, the use of UAVs has rapidly expanded to different applications such as aerial photography to monitor vegetation, survey mapping, and scouting with wireless networking [1]. UAVs have the potential for use in agricultural applications, and they are ideal for precision agriculture, compared to aerial mapping and satellite remote sensing. The use of UAVs is not only more efficient, but also more cost-effective than areal or high-resolution commercial satellite datasets [1,2,3,4,5]. They can help monitor crops in real-time and provide high-resolution images of the field and canopy, for crop growth and production. High-resolution and machine vision images are used for the identification of weeds and non-weed areas, using ground-based conventional sprayers [6,7]. In recent advancements, sprayers have been attached to UAV systems to deliver sprays in the field. However, as the payload of a UAV with a sprayer makes it heavier, it becomes difficult to fly in the field while carrying large quantities of liquid chemicals. The process of spraying agricultural crops with liquids needs to be very efficient, to avoid spraying non-crop areas. Similarly, the orchard spray system needs to fly at a high altitude, to spray chemicals on the top of the canopy. High payloads of chemicals in the tank also cause problems. Large tank sizes require more power, and generate more safety concerns while flying. It is very important to recognize the spray areas above orchards and non-orchards, to ensure the precise application of spray chemicals. For the autonomy of UAV-based spraying systems, the ability to recognize crop and orchard areas is significantly important. Most autonomous and artificial intelligence systems need to be trained on data, prior to application. Training provides confidence in operations. Machine learning systems have the potential for training and testing on datasets to add artificial intelligence for different agricultural operations. In the spraying system, a machine learning system is required to discern spraying spots and non-spraying spots in operational environments of UAVs, prior to implementing an autonomous spraying system. Ground-based vehicles can function as image processing systems, with advantages for housing onboard sprayers [8,9]. However, ground vehicles have local mapping systems without predetermined field coverage. UAVs have the advantage of identifying the field coverage in advance, with good trained datasets and machine learning systems.

Most studies using UAV have employed limited dataset collection and machine learning systems. Some research has reported that aerial applications result in only 50% of the targets being sprayed from altitudes of less than 1 m [10]. To the best of our knowledge, UAV-based sprayers were introduced to the market, and largely implemented in mountains and cropping areas, to enable spraying with precision. Most commercial UAVs with sprayers are operated under regulations in many countries. As the technology is tending toward the development of autonomous systems, it is likely that UAV spray systems will have a high potential for autonomous spraying applications. To increase the flight time, UAV manufacturers have improved the endurance of systems by increasing the battery capacity, and reducing the total weight of the UAVs. In this regard, the application efficiency of chemicals to the spray area from a high altitude must be improved. The height of the operation greatly influences how orchard areas can be covered in a minimum amount of time. A similar consideration in selecting the height of the operation pertains to croplands.

Unmanned aerial systems (UASs) have mounted sensors that offer an extraordinary opportunity for bridging the existing gap between field observations and traditional air and space-borne remote sensing, by providing high spatial detail over relatively large areas in a cost-effective way. UAV systems also have enhanced temporal retrieval for natural and agricultural ecosystem monitoring, in order to identify future directions, applications, developments, and challenges [3]. Small UAVs (approximately 1.5 kg) are capable of measuring the range to a water surface, using a global navigation satellite system (GNSS) receiver with radar, sonar, and an in-house-developed camera-based laser distance sensor (CLDS) for determining the water level (orthometric height) [4]. The remote detection of water stress in a citrus orchard was also investigated by using leaf-level measurements of chlorophyll fluorescence, and photochemical reflectance index (PRI) data, seasonal time-series of crown temperature and PRI, and high-resolution airborne imagery [5]. These studies have significant impacts on the application of UAV in agriculture and hydrology. Furthermore, machine learning systems have the opportunities for decision-making of remote sensing applications while using UAV-based sprayers.

How to generate quantitative remote sensing products by using rotating-wing and fixed-wing UAVs equipped with commercial off-the-shell (COTS) thermal and narrowband multispectral imaging sensors was also evaluated for vegetation monitoring. Radiometric calibration, atmospheric correction, and photogrammetric methods were employed, which are required to obtain accurate remote sensing products that are useful for vegetation monitoring [11]. The growing research community comprises tech-enthusiastic hydrologists that aim to design and develop their own sensing systems and to adopt a multi-disciplinary perspective to tackling complex observations, often using low-cost equipment that is intended for other applications to build innovative sensors for measurements [12]. A previously reported UAV system was integrated with optical sensing to allow for the quantitative characterization of surface flow phenomena, to yield accurate surface flow maps of sub-meter water bodies [13]. Vegetation monitoring and surface flow phenomena help in further confirmation of UAV applications for the recognition of classifiers. These studies significantly increase the scope of using UAV applications in the research community.

UAVs equipped with inexpensive thermal and narrowband multispectral imaging sensors have been used for agricultural applications, and have yielded comparable estimations [14]. Relying on UAV-based remote sensing and imaging techniques, high-throughput field phenotyping (HTFP) was conducted by using thermal imaging for the field phenomics of poplar and other tree species, for accelerating forest tree genetic improvement against abiotic stress [15]. HFTP and genetic improvement against abiotic stress also further increase applications of machine learning systems. Several machine learning systems have been introduced in ground-based sprayers using deep learning neural networks and Bayesian classifiers [16,17,18]. Most machine learning systems have high complexity data training and large time requirements for real-time applications. In our previous research, we found that the kernel mutual subspace method (KMSM) has a high potential for recognize features and actions of tracking, with accuracies of greater than 80% in real-time [19]. Furthermore, the KMSM, along with the Hankel matrix were used for action recognition of the machinery operator with a processing time of 0.07 s [19]. The mutual subspace method (MSM) has been used for the recognition of human faces and objects [20,21,22]. MSM has the advantage of object recognition from image sets or video sequences, due to receiving a variability of patterns, to achieve higher performances compared to other methods [22]. Therefore, the mutual subspace method has a highly promising capability in machine learning systems for recognizing features. In on-board spraying applications using UAVs, the recognition of features with minimum time and high accuracy can be performed using MSM.

Therefore, the objective of this research is to develop a machine learning system for recognizing the features of spraying and non-spraying areas for applying UAV-based sprayers in agricultural croplands and orchards. It is expected that MSM machine learning systems can be employed, offering the advantages of low computational complexity and good accuracy in feature recognition systems for real-time applications.

2. Materials and Methods

2.1. Mutual Subspace Method (MSM)

The MSM was introduced to the field of pattern recognition, a well-known method for object recognition based on image sets [23]. MSM is an extension of the subspace method (SM) that classifies a set of input pattern vectors into several classes, based on multiple canonical angles between the input subspace and the class subspaces (Figure 1). The input subspace was generated from a set of input patterns as a class [23,24]. The SM has a high performance in pattern recognition and was developed independently as CLAFIC (the class-SELFIC method: the original version of the subspace model) [25] and the multiple similarity method [26]. It classifies an input pattern vector into several classes, based on the minimum distance or angle between the input pattern vector and each class subspace, where a class subspace corresponds to the distribution of pattern vectors of the class in high-dimensional vector space [27].

Figure 1.

Figure 1

Subspace method (SM).

We considered that the input vector P and m class subspaces belong to a k–dimensional vector space; the similarity is defined by the length or the minimum angle between the input vector P and the i-th class subspace, where the length of P is often normalized to 1.0. The angle-based similarity can be derived as follows:

cos2θ=i=1d(P·φi)2P2 (1)

where d is the dimension of the class subspace, and φi is the i-th k–dimensional orthogonal normal vector (PCA). First, the conventional PCA operates by diagonalizing the covariance matrix Cfrom k feature vectors xj (a = 1, 2, …, k) in an n-dimensional feature space, Rn:

C=1kj=1k(xj·xjT) (2)

It gives an eigen decomposition of the covariance matrix by PCA to obtain the principal components νi(i=1, 2,,k) of the distribution:

λν=Cν (3)

However, we assume that all of data here were calculated from the data centroid. This principal component describes the direction of the largest data variation under a linear approximation [23]. The above characteristic equation can be transformed as follows:

λx=[1kj=1k(xj·xjT)]ν (4)
=1kj=1k(xj·xjT)ν=1kj=1k(xj·ν)xj (5)

Because ν is in {x1, xk }, we obtain:

λ(xa·ν)=xa·Cν (6)

where xa is a feature vector.

The MSM has been used to compare small variations in the training data and recognition target data, and this results in a powerful recognition technique when the data distribution can be linearly approximated, which occurs when multiple data can be used as recognition target image inputs. In the subspace method, a subspace that has d-dimensional vectors is selected according to a criterion such as the cumulative contribution rate from the eigenvectors, which are obtained by using PCA on the entered images [22,26,28]. Then, the similarity between the subspaces is defined according to the angle θ between the eigenvectors P={μi} (registered as a dictionary) and the eigenvectors Q={νj} (obtained from the input data) (Figure 2).

Figure 2.

Figure 2

Comparison between two sets of images using the mutual subspace method (MSM).

According to Equation (1), the angle θ between the subspaces is given as the maximum eigenvalue [21,22]:

cosθ=maxμiP maxνiQµTν (7)

where μiT μi=νjT νj=1, μiT µj=νiT νj=0, ij, 0<i,jd, and d is the dimensionality of the subspace used for recognition.

2.2. Research Design for Classifiers and MSM

The classifiers are required to be established before the MSM application. The MSM research approach involves two steps: offline and online recognition systems. The offline recognition system was used to validate the model and the accuracy of the recognition of the classifiers (Figure 3). The online recognition system was proposed, to determine the computational times required to enable the real-time system. In offline recognition, videos must be captured by using the UAV, and converted through a JPG converter. For offline recognition, selected images were taken from different datasets of crops and orchards for training and testing of the classifiers. For online recognition, a new video stream was captured. From the stream video, one frame was chosen out of 20 frames from a new video stream. Considering the restricted computational time required by a real-time system, RGB images were converted to grayscale. While testing using the online recognition system, a sliding window was used to obtain four consecutive frames, and PCA was applied, using the subspace method. In the subspace method, multiple images were required, and we noted that four frames were optimal for use in the subspace method.

Figure 3.

Figure 3

The research framework establishing the classifiers and the MSM.

In the following sections, details of field experiments for training and testing different datasets and the offline and online recognition systems are described.

2.3. Field Experiment for Training and Testing with Datasets

To implement the MSM for feature recognition, different crops and orchards are required for training the subspace patterns and verifying the recognition accuracy. While selecting the datasets for training, image acquisition at close range was preferable for agricultural croplands. On the other hand, for orchards, a high altitude allowed the canopy to be covered in a minimum time. Generally, close-range spraying can effectively reduce the drift and waste of chemicals. However, the UAV sprayer payload and the battery operational time are major concerns in enabling autonomous spraying. In this study, two working patterns were defined, depending on the flying height. The corresponding work areas are described as follows: for cropland (i.e., carrot, cabbage, and onions), the plant height was less than 5 m, and image acquisition was performed by using a UAV from a height of 5 m. In the case of orchards or plantations (i.e., chestnut, persimmons, and tall trees), we considered the height of the orchards to be less than 15 meters, and thus, the acquisition of images was conducted from a height of 15 m from the ground (Table 1). Two classifier datasets were collected for cropland spray area recognition: one dataset for spray areas (carrot, cabbage, onions) and another dataset for non-spray areas (inner farm roads, ridges, bare soil). Similarly, two classifier datasets (spray and non-spray areas) were also collected for orchards: one dataset for orchard areas (chestnuts and persimmon), and another dataset for trees, which included structured areas (farm houses, green house structure, farm buildings). The classifier datasets were captured by using a commercial UAV (DJI Phantom 3 Pro) with an onboard 4K camera with 1/2.3″ CMOS and FOV 94° 20 mm f/2.8 lens. The 4K videos were collected and converted to images by using a JPG converter at the preprocessing stage. The images were collected in the morning from 10 a.m. to 12 p.m., to ensure uniform lighting while the UAV flew over the croplands and orchards. Days with clear skies were generally chosen for collecting the videos by flying the UAV. The classifiers were segmented from the videos according to the flight heights for croplands and orchards (Table 1). Three field experiments were conducted with the UAV in three randomly selected zones; a rural farm with a combination of croplands and orchards (L1), a farm with different croplands with orchards (L2), and a research farm with croplands and orchards (L3) (Figure 4a-l). MATLAB 2015a® (MathWorks, MA, USA) was used to develop the user interface and training and testing datasets for offline and online recognition systems.

Table 1.

Training and testing with datasets classified into two categories for offline and online recognition systems.

Targets Data Sets Training Image Numbers Testing Image Numbers
Spray Nonspray Offline
(Spray + Nonspray)
Online
(Spray + Nonspray)
Offline
(Spray + Nonspray)
Online
Carrot 120 120 First half (60 + 60) All (120 + 120) Last half (60 + 60) New video
(89)
Cabbage 198 198 First half (99 + 99) All (198 + 198) Last half (99 + 99) New video (298)
Onion 107 107 First half (53 + 53) All (107 + 107) Last half (54 + 54) New video (204)
Chestnut 97 97 First half (48 + 48) All
(97 + 97)
Last half (49 + 49) New video (180)
Persimmon 94 94 First half (47 + 47) All
(94 + 94)
Last half (47 + 47) New video (210)
Trees and Structures 118 118 First half (59 + 59) All (118 + 118) Last half (59 + 59) New video (141)

Figure 4.

Figure 4

Figure 4

Figure 4

(a–l) Training and testing datasets for building the classifiers for recognizing spray areas and non-spray areas.

2.4. Offline Recognition System

The offline recognition system consisted of learning and recognition phases. The learning phase was commenced by collecting training image datasets of each class m{1,,M}, and inputting them into the system. For offline experiments of each land type, we used one of the videos, with the first half for training, and the last half for testing. The recognition phase was confirmed to begin, once the learning phases of the classifiers using scene sequences were completed (Figure 5). Then, PCA was applied, to establish the linear subspace as a reference subspace for each class. The training phase was completed in three stages. First, all of the collected testing images of Ij{1,,J} were input into the system, and each I had frames of {f1,,fn}. Second, PCA was applied to establish the linear subspace for testing the subspace for each class Ij. Finally, the canonical angles between the current testing subspace and each reference subspace were calculated. The current image was assigned to the class with whom it shared the smallest canonical angles, which indicated that it had the highest similarity when referenced to the training datasets. In an offline experiment setting, the UAV was flown 5 m above the cropland. The first half of the images for training (99 images, spray, and 99 images, non-spray) and the last half of images for testing (99 images, spray, and 99 images, non-spray) were selected for cabbage fields (Table 1). The first half of the images for training (53 images, spray, and 53 images, non-spray) and the last half of the images for testing (54 images, spray, and 54 images, non-spray) were selected for onion fields. Similarly, the first half of the images for training (60 images, spray, and 60 images, non-spray) and the last half of the images (60 images, spray, and 60 images, non-spray) were selected for testing carrot fields. A height of 15 m was chosen for flying over orchard areas to collect the first half of the images for training (48 images, spray, and 48 images, non-spray) and the last half of the images for testing (49 images, spray, and 49 images, non-spray) for chestnut trees. Again, the first half of the images for training (47 images, spray, and 47 images, non-spray) and the last half of the images for testing (47 images, spray, and 47 images, non-spray) were used in the case of the persimmon fields. Finally, the first half of the images for training (59 images, spray, and 59 images, non-spray) and the last half of the images were used for testing (59 images, spray, and 59 images, non-spray) for trees and structures. The accuracy analysis of the offline recognition system was compared with the true positive and true negative values (Table 2). For further confirmation, the extended datasets were considered, to check the recognition accuracy of classifiers by using MSM.

Figure 5.

Figure 5

Image sets in classifier recognition in the learning and recognition phases for MSM applications.

Table 2.

Accuracy analysis for the offline recognition system.

True condition (offline recognition)
Spray Nonspray ∑Total
Predicted condition
(tested by recognition phase)
Spray True Positive False Positive Total Positive
Nonspray False Negative True Negative Total Negative
Accuracy Accuracy=True Positive+True NegativeTotal

2.5. Online Recognition System

The subspace patterns were trained during the offline recognition process. These patterns were used for the online recognition development of classifiers. A sliding window was used to select four images that were converted to four vectors through resizing and reshaping. The grayscale images were resized to 8 by 8 and reshaped to one column vector using MATLAB®. A test subspace was generated using PCA for creating a matrix from the vectors. The online recognition progress was completed in the following stages. First, each video from each target crop or orchard was preprocessed, and one image was extracted from every 20 frames. Among the extracted images, there were several frames captured that did not belong to either class during takeoff and landing, or that included other plants during entry and exit. Such images were marked as noise images and removed to improve recognition accuracy. In the experiment, two datasets were collected for each target land. For the online experiment, we used all of the frames (removed noise) from one of the videos as training, and we used another video for testing (the video was not directly read; rather, the video was extracted to image frames, and the noise was removed). In the second step, we classified the set of sequential images, using the MSM classifier. Finally, the spray areas were recognized, based on the training datasets (Figure 6). In the datasets, 198 images (spray) and 198 images (non-spray) were collected from a 5 m height for training, and a reference subspace was built for use in the online experiment for cabbage. In the case of testing, a new video was taken, where one frame was selected out of 20 frames. There were a total of 298 frames that were used for testing for cabbage. Similarly, 107 images (spray) and 107 images (non-spray) were selected for training in online experiments. The new video stream was used with a total of 204 images for onion. In the case of carrots, 120 images (spray) and 120 images (non-spray) were used for training, and a new video stream with 89 images was used for testing the datasets. For orchard categories from a height of 15 m, two classifiers were trained using 97 images (spray) and 97 images (non-spray); 94 images (spray) and 94 images (non-spray); 118 images (spray) and 118 images (non-spray) for chestnut, persimmon, and trees, respectively. For testing the datasets of each target, a new video stream was taken, with a total of 180 images being extracted for chestnut, 210 images for persimmon, and 141 images for trees.

Figure 6.

Figure 6

Online recognition system for the classification of spraying, based on MSM classifiers.

3. Results

3.1. Offline Recognition Performance

In the offline recognition system, the accuracy was 80.5% in the cropland classifiers for spray and non-spray area recognition in the first experimental areas (L1). In the case of orchards, the spray and non-spray area recognition was 75% (Table 3). In the second experimental area (L2), the recognition accuracy was 70.4% and 86.1% for croplands and orchards, respectively. Finally, mixed crop and orchard areas (L3) were chosen for offline recognition by classifiers. The recognition accuracy was 72.3% and 70% for croplands and orchards, respectively. The overall accuracy was 74.3% (croplands) and 77% (orchards) for the L1, L2, and L3 locations, which had a combination of croplands and orchards (Table 3). Wide crop canopy or orchards had the advantage of higher recognition by classifiers. The high accuracy of the recognition system was obtained by using the MSM for training and testing the datasets from the three different types of experimental fields.

Table 3.

Offline classifier recognition and accuracy analysis.

True condition (offline recognition)
Location
(Croplands, Orchards)
Work patterns Cropland Orchard
Classifiers Spray Nonspray Spray Nonspray
Predicted condition
(Tested by the recognition phase)
L1 Spray 74 21 35 9
Nonspray 16 79 13 31
Accuracy 80.5% 75%
L2 Spray 38 11 41 2
Nonspray 18 31 10 33
Accuracy 70.4% 86.1%
L3 Spray 56 0 37 18
Nonspray 31 25 15 40
Accuracy 72.3% 70%

(L1: a farm with a combination of croplands and orchards, L2: a farm with different croplands with orchards, L3: a research farm with croplands and orchards)

For further confirmation, frame numbers were increased for the training and testing of datasets, whether there were significant differences in the recognition accuracy of the classifiers. Extended datasets confirmed that the accuracy of MSM method did not change much, even if the frames were increased to double for the testing and training of datasets in the offline recognition system (Table 4).

Table 4.

Extended datasets for the training and testing of classifiers, using an offline recognition system.

Croplands and Orchards Data Sets Training Image Numbers Testing Image Numbers Accuracy
Spray Nonspray Offline Offline
Carrot 256 256 First half (128 + 128) Last half (128 + 128) 73.79%
Cabbage 440 440 First half (220 + 220) Last half (220 + 220) 81.25%
Onion 210 210 First half (105 + 105) Last half (105 + 105) 66.32%
Chestnut 224 224 First half (112 + 112) Last half (112 + 112) 77.31%
Persimmon 248 248 First half (124 + 124) Last half (124 + 124) 70.94%
Trees and Structures 216 216 First half (108 + 108) Last half (108 + 108) 64.58%

3.2. Online Recognition Performance

The developed user interface had the advantage of online information, which included the current cropped image, the tested image sets using a sliding window, the predicted category, the recognition rate (the correct classifications were known during the test), the computational time, and the similarity plot. For the cropland classifiers, the UAV was flown at a 5 m height, and the recognition rate was observed to be 65.5% for the L1 experimental areas. The computational time was only 0.0031 s for classifier recognition (Figure 7a–c). The flying height was 15 m for orchard classification, and recognition was observed at 69.1%, with a computation time of 0.0031 s for each classifier. In the second experimental flying areas (L2), the recognition accuracy of the classifiers for the noted spray and non-spray areas was 60.8% and 82.2% for croplands and orchards, respectively. The computational time was only 0.0031 s for recognition by the classifiers, and orchard classifier recognition also required only 0.0031 s for each classifier (Figure 8a–c). In the third experimental location (L3), the online recognition rate by the classifiers reached 69% in 0.0048 s for each classifier, and 71.7% in 0.0031 s for each classifier, in croplands and orchards, respectively. The online recognition system had an average accuracy of 65.1% and 75.1% for croplands and orchards, respectively, with a recognition time of 0.0031 s (Table 5).

Figure 7.

Figure 7

Figure 7

(a–c) Online recognition performance of a classifier of croplands from a height of 5 m.

Figure 8.

Figure 8

Figure 8

(a–c) Online recognition performance of a classifier of orchards from a height of 15 m.

Table 5.

Online classifier recognition and accuracy analysis

Croplands and Orchards Flying Height
(m)
Accuracy
(%)
Recognition Time of Classifier
(s)
Carrot 5 65.51 0.0031
Cabbage 5 60.88 0.0048
Onion 5 69.00 0.0031
Chestnut 15 69.10 0.0031
Persimmon 15 82.21 0.0031
Trees and Structures 15 74.10 0.0031

4. Discussion

The field experiments were conducted in different types of fields, to increase the dataset variety for the selection of spray and non-spray areas inside the croplands and orchards. The offline recognition system shows the effectiveness of MSM for training and testing of the datasets for croplands and orchards. The classifiers were used for croplands and orchards, and they were limited to being trained and tested on datasets that were acquired in the late fall season. MSM has the flexibility of increasing the number of classifiers, which may increase the computational time requirement. As UAV spraying is performed at higher speeds, we tend to focus on minimizing the computation time, to reduce the computational burden for decision-making, to recognize the spray and non-spray areas in croplands and orchards. UAVs operating at a high speed with limited battery life and a small payload of liquid chemicals demand high computational speeds and fast operation with good recognition accuracy. With this consideration, the online recognition system provided some advantages, although its accuracy was not as high as that of the offline recognition system. The system needs further training data to increase its accuracy, especially for the identification of croplands less than 5 m high and orchard areas from 15 m high. In the online experiment setting, similar environments resulted in increased recognition, while adding different categories of orchards reduced recognition. It was very challenging to test the datasets from a UAV fast operating speed at a high altitude. Classifiers were trained and tested on datasets acquired from three different locations to confirm the recognition accuracy. However, complex canopy systems were not present in the features. This MSM system had limitations in recognizing classifiers in complex canopies of crops or orchards. We could not collect images of complex canopy crops, and we assume that in such canopy systems, upward and downward image acquisitions are required to identify the spray and non-spray areas under different lighting conditions. Lighting is a key point that needs to be carefully considered, especially for its interception through the canopy. It would be ideal to train the UAV features of spray and non-spray areas on a large field, to obtain higher accuracies in precision applications ranging from normal to complex canopies of crops. Further studies are required to deal with such complexities of canopies, very large datasets in different lighting conditions, and the processing of images to remove noise by using extended Kalman filters, in onboard UAV systems.

5. Conclusions

A machine learning system was developed by using MSM for images collected by a UAV in different types of farm fields and orchards. The machine learning system was developed to train and test two classifiers, one for agricultural croplands and one for orchard areas, on different datasets, to distinguish spray and non-spray areas for the future development of autonomous spraying systems. Images were collected from low (5 m) and high altitudes (15 m). The accuracy of the offline recognition system was found to be 74.4% and 77% for low- and high-altitude systems, respectively. On the other hand, the online recognition system performance had an average accuracy of 65.1% and 75.1% for low-altitude and high-altitude image acquisition systems, respectively. The computation time for the online recognition systems was observed to have a minimum of 0.0031 s (on average) for reporting classifier recognition. The developed machine learning system for recognizing the classifiers of spray and non-spray areas can be implemented in the autonomous UAV spray system in real-time. In our future experiments, we will improve the training and testing system by incorporating an artificial neural network (ANN) and deep learning to develop a UAV-based autonomous spraying unit for croplands and orchards.

Author Contributions

P.G. is the first author and PhD candidate. He designed the research, conducted field experiments, collected data, analyzed data and prepared the manuscript. Y.Z. and L.Z. contributed in analyzing data. R.N. advised during manuscript preparation. T.A. is the corresponding author contributed in designing research, idea development and manuscript editing.

Funding

This research was supported by the University of Tsukuba and received no external funding. The APC was funded by the University of Tsukuba, Japan.

Conflicts of Interest

The authors declare no conflict of interest.

References

  • 1.Zhang Y., Chen D., Wang S., Tian L. A promising trend for field information collection: An air-ground multi-sensor monitoring system. Inf. Process. Agric. 2018;5:224–233. doi: 10.1016/j.inpa.2018.02.002. [DOI] [Google Scholar]
  • 2.Peña Barragán J.M., Kelly M., Castro A.I.D., López Granados F. Object-based approach for crop row characterization in UAV images for site-specific weed management; Proceedings of the 4th Geographic Object-Based Image Analysis (GEOBIA); Rio de Janeiro, Brazil. 7–12 May 2012; p. 426. [Google Scholar]
  • 3.Salvatore M., Matthew F.M., Pauline E.M., Richard L., Victor P.M., Giorgos M., Eyal B.D., David H., Lyndon E., Giuseppe C., et al. On the use of unmanned aerial systems for environmental monitoring. Remote Sens. 2018;10:641. [Google Scholar]
  • 4.Bandini F., Jakobsen J., Olesen D., Reyna-Gutierrez J.A., Bauer-Gottwein P. Measuring water level in rivers and lakes from lightweight unmanned aerial vehicles. J. Hydrol. 2017;548:237–250. doi: 10.1016/j.jhydrol.2017.02.038. [DOI] [Google Scholar]
  • 5.Zarco-Tejada P.J., González-Dugo V., Berni J.A. Fluorescence, temperature and narrow-band indices acquired from a UAV platform for water stress detection using a micro-hyperspectral imager and a thermal camera. Remote Sens. Environ. 2012;117:322–337. doi: 10.1016/j.rse.2011.10.007. [DOI] [Google Scholar]
  • 6.Hung C., Xu Z., Sukkarieh S. Feature learning based approach for weed classification using high resolution aerial images from a digital camera mounted on a UAV. Remote Sens. 2014;6:12037–12054. doi: 10.3390/rs61212037. [DOI] [Google Scholar]
  • 7.Rebetez J., Satizábal H.F., Mota M., Noll D., Büchi L., Wendling M., Cannelle B., Pérez-Uribe A., Burgos S. Augmenting a convolutional neural network with local histograms—A case study in crop classification from high-resolution UAV imagery; Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning; Bruges, Belgium. 27–29 April 2016; pp. 27–29. [Google Scholar]
  • 8.Peteinatos G., Weis M., Andújar D., Rueda Ayala V., Gerhards R. Potential use of ground-based sensor technologies for weed detection. Pest Manag. Sci. 2014;70:190–199. doi: 10.1002/ps.3677. [DOI] [PubMed] [Google Scholar]
  • 9.Lee D.H., Lee K.S., Cho Y.J., Lee J.Y., Chung S.-O. Development of autonomous sprayer considering tracking performance on geometrical complexity of ground in greenhouse. J. Biosyst. Eng. 2012;37:287–295. doi: 10.5307/JBE.2012.37.5.287. [DOI] [Google Scholar]
  • 10.Pimentel D., Burgess M. Small amounts of pesticides reaching target insects. Environ. Dev. Sustain. 2011;14:1–2. doi: 10.1007/s10668-011-9325-5. [DOI] [Google Scholar]
  • 11.Berni J., Zarco-Tejada P., Suárez L., González-Dugo V., Fereres E. Remote sensing of vegetation from UAV platforms using lightweight multispectral and thermal imaging sensors. Remote Sens. 2009;38:722–738. doi: 10.1109/TGRS.2008.2010457. [DOI] [Google Scholar]
  • 12.Tauro F., Selker J., van de Giesen N., Abrate T., Uijlenhoet R., Porfiri M., Manfreda S., Caylor K., Moramarco T., Benveniste J. Measurements and observations in the XXI century (MOXXI): Innovation and multi-disciplinarity to sense the hydrological cycle. Hydrol. Sci. J. 2018;63:169–196. doi: 10.1080/02626667.2017.1420191. [DOI] [Google Scholar]
  • 13.Tauro F., Porfiri M., Grimaldi S. Surface flow measurements from drones. J. Hydrol. 2016;540:240–245. doi: 10.1016/j.jhydrol.2016.06.012. [DOI] [Google Scholar]
  • 14.Berni J.A., Zarco-Tejada P.J., Suárez Barranco M.D., Fereres Castiel E. Thermal and narrow-band multispectral remote sensing for vegetation monitoring from an unmanned aerial vehicle. Inst. Electr. Electron. Eng. 2009;47:722–738. [Google Scholar]
  • 15.Ludovisi R., Tauro F., Salvati R., Khoury S., Mugnozza Scarascia G., Harfouche A. UAV-based thermal imaging for high-throughput field phenotyping of black poplar response to drought. Front. Plant Sci. 2017;8:1681. doi: 10.3389/fpls.2017.01681. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Carrio A., Sampedro C., Rodriguez-Ramos A., Campoy P. A review of deep learning methods and applications for unmanned aerial vehicles. J. Sens. 2017 doi: 10.1155/2017/3296874. [DOI] [Google Scholar]
  • 17.Majidi B., Bab-Hadiashar A. Real time aerial natural image interpretation for autonomous ranger drone navigation; Proceedings of the IEEE Digital Image Computing: Techniques and Applications (DICTA’05); Queensland, Australia. 6–8 December 2005. [Google Scholar]
  • 18.Tellaeche A., Burgos-Artizzu X.P., Pajares G., Ribeiro A. A vision-based method for weeds identification through the Bayesian decision theory. Pattern Recognit. 2008;41:521–530. doi: 10.1016/j.patcog.2007.07.007. [DOI] [Google Scholar]
  • 19.Yan Z., Pengbo G., Tofael A. Development of a rescue system for agricultural machinery operators using machine vision. Biosyst. Eng. 2018;169:149–164. [Google Scholar]
  • 20.Sakano H., Mukawa N., Nakamura T. Kernel mutual subspace method and its application for object recognition. Electron. Commun. Jpn. (Part II Electron.) 2005;88:45–53. doi: 10.1002/ecjb.20190. [DOI] [Google Scholar]
  • 21.Fukui K., Yamaguchi O. Face recognition using multi-viewpoint patterns for robot vision; Proceedings of the Eleventh International Symposium on Robotics Research, ISRR; Siena, Italy. 19–22 October 2003; Berlin/Heidelberg, Germany: Springer; 2005. pp. 192–201. [Google Scholar]
  • 22.Fukui K., Yamaguchi O. The kernel orthogonal mutual subspace method and its application to 3D object recognition; Proceedings of the Asian Conference on Computer Vision; Tokyo, Japan. 18–22 November 2007; Berlin/Heidelberg, Germany: Springer; 2007. pp. 467–476. [Google Scholar]
  • 23.Maeda K., Watanabe S. Pattern matching method with local structure. Trans. IEICE. 1985;68:345–352. [Google Scholar]
  • 24.Watanabe S., Pakvasa N. Subspace method of pattern recognition; Proceedings of the 1st IJCPR; Washington, DC, USA. Novermber 1973; pp. 25–32. [Google Scholar]
  • 25.Iijima T., Genchi H., Mori K.-I. Learning Systems and Intelligent Robots. Springer; Boston, MA, USA: 1974. A theory of character recognition by pattern matching method; pp. 437–450. [Google Scholar]
  • 26.Yamaguchi O., Fukui K., Maeda K.-I. Face recognition using temporal image sequence; Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition; Nara, Japan. 14–16 April 1998; p. 318. [Google Scholar]
  • 27.Fukui K., Yamaguchi O. Computer Vision. Springer; Berlin/Heidelberg, Germany: 2014. Subspace methods; pp. 777–781. [Google Scholar]
  • 28.Scholkopf B., Smola A., Uuller K.-R. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 1998;10:1299–1319. doi: 10.1162/089976698300017467. [DOI] [Google Scholar]

Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES