Abstract
Mobile Mapping Technology (MMT) is one of the most important 3D spatial data acquisition technologies. The state-of-the-art mobile mapping systems, equipped with laser scanners and named Mobile LiDAR Scanning (MLS) systems, have been widely used in a variety of areas, especially in road mapping and road inventory. With the commercialization of Advanced Driving Assistance Systems (ADASs) and self-driving technology, there will be a great demand for lane-level detailed 3D maps, and MLS is the most promising technology to generate such lane-level detailed 3D maps. Road markings and road edges are necessary information in creating such lane-level detailed 3D maps. This paper proposes a scan line based method to extract road markings from mobile LiDAR point clouds in three steps: (1) preprocessing; (2) road points extraction; (3) road markings extraction and refinement. In preprocessing step, the isolated LiDAR points in the air are removed from the LiDAR point clouds and the point clouds are organized into scan lines. In the road points extraction step, seed road points are first extracted by Height Difference (HD) between trajectory data and road surface, then full road points are extracted from the point clouds by moving least squares line fitting. In the road markings extraction and refinement step, the intensity values of road points in a scan line are first smoothed by a dynamic window median filter to suppress intensity noises, then road markings are extracted by Edge Detection and Edge Constraint (EDEC) method, and the Fake Road Marking Points (FRMPs) are eliminated from the detected road markings by segment and dimensionality feature-based refinement. The performance of the proposed method is evaluated by three data samples and the experiment results indicate that road points are well extracted from MLS data and road markings are well extracted from road points by the applied method. A quantitative study shows that the proposed method achieves an average completeness, correctness, and F-measure of 0.96, 0.93, and 0.94, respectively. The time complexity analysis shows that the scan line based road markings extraction method proposed in this paper provides a promising alternative for offline road markings extraction from MLS data.
Keywords: mobile LiDAR scanning, point clouds, scan line, road points extraction, road markings extraction
1. Introduction
Mobile Mapping Technology (MMT) emerged during the nineties in the twentieth century inspired by the availability of GPS technology for civilian uses. The world’s first mobile mapping system, named as GPSVan, which integrated a Code-Only GPS receiver, two CCD cameras, two video cameras and multiple dead reckoning sensors on a van, was integrated by Ohio State University in 1991 [1,2]. Later, The University of Calgary implemented the VISAT system including a dual-frequency carrier differential receiver and Inertial Measurement Unit (IMU) instead of the Code-Only GPS receiver and dead reckoning sensors, CCD cameras were also used in the VISAT system [3,4]. Similar systems were developed and reported in KiSS [5], GPSVision [6] and GI-EYE™ [7]. Since 2000, the main mapping sensors used in mobile mapping systems have been switched from cameras to laser range finders or laser scanners [8]. The mobile mapping systems equipped with mobile laser scanners are named Mobile LiDAR Scanning (MLS) systems, and most of the newly developed systems belong to this type.
MLS has been used in a variety of applications. The most obvious applications are road mapping and road inventory [9,10,11,12,13], in which road information such as curbstones, basic structures of road, road markings and traffic signs are extracted from MLS data. In [14,15], mobile LiDAR systems are used in tunnel cross-section and road cross-section measurements. In [16,17,18], the usage of mobile LiDAR systems in fluvial studies are reported. In [19], the railway infrastructure is automatically recognized from MLS data. In [20,21,22], mobile LiDAR systems are used for large scale plot mapping and individual tree crown reconstruction in forest research. In [23,24], MLS data are used to detect street environment changes and street-side vehicles. Geo-spatial information is extracted from MLS data for crowd control planning in [25].
Among the applications of MLS mentioned above, the most important and most promising applications are road mapping and road inventory, especially with the commercialization of Advanced Driving Assistance Systems (ADASs) and self-driving technology. In ADAS and self-driving, lane-level detailed 3D maps are needed for route planning and navigation in which the traditional 2D digital maps such as Google Maps are no longer sufficient. The most promising technology to generate such large scale lane-level detailed 3D maps is mobile LiDAR technology for its efficiency, cost-effectiveness, high accuracy and high density of the acquired point clouds.
To generate lane-level detailed 3D maps using mobile LiDAR point clouds, road markings and road edges should be extracted from the raw LiDAR point clouds. Road markings are generally made of high reflectivity material, so the intensity values of road marking points are theoretically higher than that of asphalt road surface points. Thus road markings can be distinguished from road points according to intensity information. However, the intensity values recorded by laser scanners are affected by target surface characteristics, acquisition geometry, instrumental effects and environmental effects [26]. In MLS, instrumental parameters, including the transmitted energy, the intensity bit depth and scaling, the amplifier for low reflective surfaces, the automation gain control, the brightness reducer for near distances and the aperture size, can be considered as constant, and the environmental parameters, including atmospheric transmittance and wetness, can be considered as constant in a scanning project, so the main parameters affecting the intensity values are target surface characteristics and acquisition geometry. As pointed out in [27], the incidence angle and scanning distance affect the intensity values of terrestrial laser scanning greatly. Thus the intensity values of road points vary not only because of different target characteristics (i.e., asphalted road surface and road markings), but also because of different scanning incidence angle and scanning distance, making it difficult to separate road marking points from road points.
In order to extract road markings more efficiently and simplify the road marking points extraction algorithm, road points (including asphalted road surface points and road marking points) are first extracted from the raw LiDAR point clouds and then road markings are further extracted from the road points. In [28], road surface is segmented from the feature image generated by the raw 3D LiDAR points. In [29], road edges are extracted by combination of two modified versions of the parametric active contour model and road points can be extracted by the obtained road edges. In [30], road points are extracted by curbstone information on both sides of the road. In [15,31], road points are segmented from points in each scanner cycle by PCA analysis.
There are two typical types of methods to extract road markings from mobile LiDAR point clouds. The first type of methods convert the point clouds into georeferenced images and image processing methods are then used to extract road markings from the georeferenced images. In [11], 2D Georeferenced Feature (GRF) images are interpolated from the 3D road points by modified Inverse Distance Weight (IDW) interpolation, then Weighted Neighboring Difference Histogram is used to segment the GRF images and Multi-Scale Tensor Voting (MSTV) is used to extract road markings. In [30], geo-referenced intensity image is generated first by extended IDW method, and then road markings are recognized by a point-density-dependent multi-threshold segmentation of the intensity image and refined by morphological closing operations. In [31], road points are rasterized into images using a nearest neighbor algorithm to assign intensity data to a regular matrix, and then zebra crossings are detected by Standard Hough Transform from edges of the binary image. In [32], intensity curve fitting is used to reduce the variance of the intensity value, then intensity raster image is generated and road markings are extracted from the raster image by image processing method. In [33], candidate lane markings are localized in 3D by adaptive thresholding in scan lines, then 2D images are interpolated using the 3D candidate lane marking points and all candidate road markings are detected by Hough Transform clustering. Finally, false detections are eliminated by trajectory constraints and geometry checks. The second type of methods extract road markings from the 3D point clouds directly, which is much less reported until now. In [34], a method that extracts road marking points from 3D road points directly is reported. In this method, road points are divided into multi-segments, and road marking points are extracted from each segment by multi-thresholding method and refined by spatial density based filtering.
Extracting road markings by converting 3D point clouds into 2D georeferenced images simplifies the algorithms and many mature image processing algorithms can be used, but this method may lead to loss of high precision of the point clouds, and incompleteness and incorrectness of some objects. Extracting road marking points from 3D point clouds directly has seldom been reported. This paper proposes a scan line-based method to extract road marking points from 3D mobile LiDAR point clouds directly. The workflow of the method is illustrated in Figure 1. In this method, the acquired point clouds are first preprocessed to remove the isolated LiDAR points in the air and organized into scan lines, then the road points are extracted by moving least squares line fitting using the seed road points extracted by Height Difference (HD) between trajectory data and road surface. Finally, road marking points are extracted from scan line points by Edge Detection and Edge Constraint (EDEC) method after intensity smoothing and Fake Road Marking Points (FRMPs) are eliminated by segment and dimensionality feature based refinement.
This paper is structured as follows: Section 2 introduces our preprocessing methods including the method to remove isolated LiDAR points in the air from the raw LiDAR data and the approach to organize LiDAR point clouds into scan lines. In Section 3, the method to extract road points is addressed in detail and Section 4 focuses on the algorithms of road marking points extraction from road points and FRMPs elimination. Experiments, results and discussion are presented in Section 5. Finally, Conclusions are presented in Section 6.
2. Preprocessing
2.1. Isolated LiDAR Points Removal
The isolated LiDAR points in the air occurring in the scanned point clouds affect the algorithms used to extract object information from point clouds. Actions should be taken to remove or reduce the isolated LiDAR points in the air before any subsequent operations. There are many causes for the isolated LiDAR points in the air, and the dust in the air is a case in point. No matter what the causes are, the intensity values of the isolated LiDAR points in the air are low. Figure 2 shows the intensity histogram of some isolated LiDAR points manually labeled from a dataset, which indicates that the intensity values of almost all isolated LiDAR points in the air are less than 10. This may be because that the isolated LiDAR points in the air are reflected by objects with very small diameter present in the air. Due to the small diameter, a very small part of the laser energy is reflected, leading to the observed small intensity values. Another characteristic of the isolated LiDAR points in the air is that their scanning distance (the distance from object to laser scanner) is obviously shorter than that of the previous scanned point and the next scanned point, as illustrated in Figure 3.
According to the two characteristics of the isolated LiDAR points in the air identified above, a point p with scanning distance and intensity value , is classified as an isolated LiDAR point in the air based on Equation (1). In the equation, Sp and Sn correspond to the scanning distance of the previous point and the next point, and is the intensity threshold:
(1) |
Some LiDAR points, especially tree points and vegetation points, may be mistakenly classified as isolated LiDAR points based on Equation (1), but this does not matter since we are only interested in road points and road points are seldom classified as isolated LiDAR points based on Equation (1).
2.2. Scan Line Separation
The original mobile LiDAR points are ordered sequentially by timestamp, and the point clouds can either be regarded as a whole or a large amount of points. In the principle of mobile LiDAR scanning, objects are scanned scan line by scan line with the movement of the platform. Scan lines can be recovered from the point clouds and subsequent processing can be done scan line by scan line. There are many advantages of processing mobile LiDAR points scan line by scan line: (1) it is easier to manage the small number of points in a scan line by simpler data structure than huge amount of points in a whole; (2) it is more time effective to process mobile LiDAR point clouds in scan lines compared with processing in a whole, especially with the increase of points number; (3) the scan lines can be processed in parallel.
The scanning angle of a point p can be acquired from the raw LiDAR data or computed from its coordinates in Scanner Owned Coordinate System (SOCS). The scanning angles of points in each scan line increase consecutively from the minimum to the maximum in the Field Of View (FOV). For a mobile LiDAR with FOV of 360 degrees, the scanning angles of points in a scan line range from 0 to 360. The whole mobile LiDAR points can be separated into different scan lines according to the scanning angle. Figure 4 illustrates a small part of point clouds separated into scan lines. In order to show the scan lines more clearly, one scan line is selected in every 50 scan lines.
3. Road Points Extraction
The Position and Orientation System (POS) is rigidly fixed on the platform of a mobile LiDAR system. The points under the trajectory data are certainly road points. The Height Difference (HD) between trajectory and road surface (Figure 5) can be estimated based on the trajectory data and scanned mobile LiDAR point clouds. This HD can be used to help extract seed road points from scan line data effectively. After seed road points are extracted, full road points are extracted by moving least squares line fitting.
3.1. HD Estimation
Time data, position data(x, y, z) and attitude data (roll, pitch, heading) are included in every record of trajectory data; meanwhile, time data, coordinate data(x, y, z) and intensity data are included in every record of point clouds. In order to estimate HD, the scanned points right below the trajectory data should be found. Given a trajectory record with time and coordinate , the points scanned within and can be extracted from the originally scanned point clouds, where δt is a small time value. Then the points horizontally neighboring the trajectory coordinate within a small distance are extracted. If N points are extracted, denote as and , HD can be estimated according to Equation (2):
(2) |
HD can be estimated for every scan line, but we usually estimate just one HD for each block of MLS data which is robust and more time efficient.
3.2. Seed Road Points Extraction from Scan Line
Seed road points can be extracted by HD. The trajectory data are outputted with constant time interval. Given a point p with time t and coordinate in a scan line, there may be no trajectory record at exact time t. The trajectory data of time t should be interpolated from the trajectory records before t and after t. Assume that the interpolated trajectory coordinate is , if the z value of p conforms to Equation (3), where dhd is HD estimated by Equation (2) and ρhd is a small threshold, p is labeled as a seed road point:
(3) |
Some LiDAR points reflected by vegetation or other non-road objects may also have z values conforming to Equation (3) and are mistakenly labeled as seed road points. To identify the final seed road points, they are clustered into groups by their indexes. The points with consecutive indexes are divided into the same group, and the group with most point number is identified as the final seed road points.
3.3. Full Road Points Extraction from Scan Line
The whole road points in a scan line cannot be represented by a straight line, but a small part or a local part of the road surface can be fitted as a straight line. The distance from a point to the line fitted can be used to determine whether the point is road point or not, as shown in Figure 6. In order to identify whether the red point p in the right side of the extracted road points (colored in green) is a road point or not, a line can be fitted by the already identified road points near point p. If the distance from point p to the fitted line is smaller than threshold ρd, the point is determined as road point, otherwise not. The line fitting will continue until two adjacent points are both determined as non-road points.
In order to fit a line conforming to the local part of the road surface, moving least squares [35] line fitting is efficient. In moving least squares line fitting, the contribution of a point to the fitted line is determined by the weight. By assigning smaller weight to the points far away and greater weight to the points near the local part, the fitted line will conform to the local part. Besides, in order to simplify the line fitting, the points of the scan line are transformed to the 2D Scan Line Coordinate System (SLCS) whose origin is set at the middle of the seed road points and X axis parallel to the scan line direction, as illustrated in Figure 7. Equation (4) can be used as the line equation in SLCS.
If the already extracted road points are denoted as ri(xi, yi) in scan line coordinate system and the point nearest to p is r1, as illustrated in Figure 6. The line conforming to the local of the road at p can be fitted by minimizing Equation (5):
(4) |
(5) |
In Equation (5), xi and yi are the 2-D coordinate in SLCS and wi is the weight and calculated by the exponential weight function:
(6) |
where di is the distance between point ri and r1. In this way, the bigger the distance is, the smaller the weight is. Point r1, which is nearest to p, has a weight of 1 and other points have weights smaller than 1. h is the spacing parameter which can control within what distance the points has effect on the line to be fitted. Since e−5 equals to 0.0067, which is very small and almost have no effect on the fitted line, h can be determined by Equation (7) with the maximum distance D carrying effect:
(7) |
4. Road Markings Extraction and Refinement
4.1. Intensity Data Smoothing by Dynamic Window Median Filter
The intensity values of road points in a scan line can be classified into two types: (1) the intensity values of road marking points; (2) the intensity values of asphalt road surface points. The intensity values of asphalt road surface points in a scan line (except the edges between road marking points and asphalt road points) should theoretically be the same or change smoothly, but in fact, the intensity values of adjacent asphalt road points in a scan line may vary greatly (Figure 8a). In order to smooth the intensity data of points in a scan line, median filter is used, which can not only smooth the intensity data of asphalt road surface points but also can preserve the intensity value of road marking edge point. Figure 9 shows four situations in filtering process by median filter with window size of 7. In Figure 9a,b, the points inside the window are all asphalt road points or all road marking points. In Figure 9a, the intensity value of the point at the center of the window is normal and will be kept. In Figure 9b, the intensity value of the point at the center of the window is obviously greater than that of others in the window and will be replaced by the median value of intensity values inside the window. In Figure 9c,d, the center point of the window is at the edge of road marking and the intensity value of the center point will be preserved after filtering since the window size is odd.
In order to preserve the intensity value of road marking edge points successfully, the window size should be smaller than the point number of road markings in the scan line. Because of different scanning distances, the distribution of mobile LiDAR points is non-uniform, and the point number of road markings varies in different scanning distances even when the width of the road markings are the same, so the window size of the median filter should be dynamically changed according to the point resolution. According to the standard for road markings [36], the minimum width of road markings in China is 0.2 m; thus in this paper, the window size of the median filter is dynamically determined by the minimum width and the point resolution. Figure 8b illustrates the intensity values after filtering by dynamic window median filter, with the intensity values of asphalt road surface points effectively smoothed and the intensity value of road marking edge point successfully preserved.
4.2. Road Markings Extraction by EDEC
After smoothing by dynamic window median filter, the intensity values of points in a scan line are considered to change smoothly except at the edges between asphalt road surface and road markings. There are two types of edges between asphalt road surface points and road marking points as illustrated in Figure 10: (1) the steep edge; (2) the smooth edge. At the steep edge, the intensity value suddenly increases from low to high or otherwise decrease from high to low (Figure 10a). At the smooth edge, the intensity value changes smoothly from low to high or conversely from high to low (Figure 10b). Theoretically, all edges between road markings and asphalt road surface should be steep edges, but in fact, smooth edges also exist.
The road markings with steep edges can be extracted easily by thresholding, but those with smooth edges are hard to extract because the intensity values of road marking points are close to that of asphalt road surface points nearby. Road marking points are extracted from road points by detecting the edge points between road markings and asphalt road surface in a scan line in this paper. The edge points are detected by intensity gradient of the points in scan line. In order to detect smooth edge points successfully, the gradient value of ith point in scan line is estimated by:
(8) |
In Equation (8), Gi is the gradient of ith point, and Ii and Ii−k are the intensity values of ith and (i − k)th point. Since the intensity values change smoothly after filtering, the absolute value of Gi equals to zero or one at non-edge points and the absolute value of Gi is greater than 1 at edge points. In order to detect edge points more conveniently, the gradients are further normalized to:
(9) |
There are two edge points between a road marking and asphalt road surface, and the two edge points correspond to a pair of non-zero normalized gradient values with one positive and one negative. Figure 11 shows the normalized gradient values of a scan line, and the pairs of non-zero values inside the green rectangles represent the edge points of road markings. The pairs of non-zero values with one positive and one negative are detected from the normalized intensity gradient values, and the points between the two edge points are extracted as road marking points. This method is called Edge Detection and Edges Constraint method (EDEC). The extracted road marking points between a pair of edge points belong to a road marking segment.
4.3. Segment and Dimensionality Feature Based Refinement
Besides the road marking points, some asphalt road surface points are mistakenly extracted as road marking points by EDEC method. We call these mistakenly extracted points Fake Road Marking Points (FRMPs). The FRMPs are eliminated by segment based refinement and dimensionality feature based refinement. The segment based refinement is based on the fact that road markings are scanned in multiple consecutive scan lines while the FRMPs are scanned in a few consecutive scan lines or even only in one scan line. If a road marking extracted by EDEC is scanned by M consecutive scan lines and M is smaller than the minimum scan line threshold ρs, the points consist of the road marking are classified as FRMPs and eliminated. ρs is estimated by Equation (10) according to the minimum width of road markings along the platform moving direction (Wmin) and the distance between two scan lines (Dr).
(10) |
Another type of FRMPs is illustrated in Figure 12.
Such FRMPs can be eliminated by dimensionality feature [37] because the neighborhood of road marking points make up a 2D plane and the neighborhood of FRMPs make up a 1D line. The dimensionality feature linearity Lλ is defined as follows:
(11) |
where , and are eigenvalues with estimated from the covariance matrix constructed for point by involving its neighborhood . If is greater than 0.9, point is regarded as FRMPs and eliminated.
5. Results and Discussion
5.1. The Mobile LiDAR System and Mobile LiDAR Dataset
To evaluate the performance of the scan line based road markings extraction method proposed in this paper, a dataset is collected by a mobile LiDAR system operated by us (Figure 13). In this system, the SPAN-SE system with UIMU-LCI provided by Novatel Inc. in Canada, the VUX-1 laser scanner provided by RIEGL Laser Measurement Systems GmbH in Austria and a multi-camera panorama camera are integrated on a SUV car. In order to fuse the data in post-processing, all sensors are synchronized to GPS time. Besides, all sensors are configured through software run on the embedded computer and data sensed by the sensors are saved in the embedded computer. The boresight and misalignment between laser scanner and IMU are carefully calibrated before data collection.
The mobile LiDAR point clouds data of the Jincheng highway in Beijing (China) were collected by the mobile LiDAR system shown in Figure 13. The speed of mobile LiDAR system in data acquisition was about 40 kilometers per hour and the distance between two scan lines is about 0.06 m. The total data coverage is about 57 kilometers. To improve the accuracy and reliability of the mobile LiDAR point clouds, two GNSS base stations with data acquisition rate of 1 HZ were used during the data acquisition. The GNSS base stations were mounted along the highway and the baseline from mobile LiDAR system to the nearer GNSS base station was always shorter than 15 km. The GNSS data of base stations, the GNSS data and the IMU data of mobile LiDAR system were post-processed by Inertial Explorer software from Novatel to get optimized time, position and attitude data. The mobile point clouds were obtained by direct geo-referencing of the LiDAR sensed data using the optimized time, position and attitude data. Mobile point clouds scanned from both directions of the highway may suffer from inconsistency due to the error in position and attitude data [38,39,40]. The method in [39] was used to reduce the inconsistency of the point clouds scanned from both directions of the highway in our data.
For convenience, three typical data samples (data sample A, data sample B and data sample C as illustrated in Figure 14) are selected as test data from the whole dataset. Data sample A is about 100 m long and contains about 2.8 million points. Arrow shape and rectangle road markings are included in data sample A. Data sample B is about 70 m long and contains about 2.2 million points. Zebra crossings and other rectangle road markings are included in data sample B. Data sample C is about 100 m long and contains about 2.8 million points. Arrow shape road markings, rectangle road markings and exit signs are included in data sample C.
In the proposed road markings extraction method, the height difference threshold ρhd in seed road points extraction, the distance threshold ρd in full road points extraction and k in EDEC should be determined. In Section 5.2, a small block of mobile LiDAR dada is chose to analyze the parameter sensitivity and determine the empirical value of these parameters. The length of the chosen mobile LiDAR data is about 140 m and the number of points is about 4.8 million. Rectangle road markings and arrow shape road markings are included in the data. Besides, a car is scanned in the data and thus occlusion is existing in the chosen data (Figure 15). The chosen mobile LiDAR data for parameter sensitivity analysis is not included in the sample data used for quantitative evaluation in Section 5.3.
5.2. Parameter Sensitivity Analysis
The height difference threshold ρhd will affect the point number of seed road points extracted. A smaller ρhd will result in fewer seed road points and bigger ρhd will result in more seed road points. We tested a set of parameter configurations on seed road points extraction and the testing results are shown in Figure 15. As seen from the seed road points extraction results, all configurations extract seed road points quite well, in which the bigger ρhd is, the more seed road points are extracted. In order to ensure that enough seed road points are extracted and avoid non-road points being mistakenly extracted as seed road points, we set ρhd to 0.03 m in our experiments.
The distance threshold ρd in moving least squares line fitting for full road points extraction from scan lines will affect the road points extraction result. We tested a set of parameter configurations on full road points extraction to select an empirical ρd value and the results of two scan lines are shown in Figure 16. From the results we can see that road points are not fully extracted when ρd is set to 0.01 m and 0.02 m. When ρd is up to 0.03 m, all road points are correctly extracted. When ρd is increased to 0.04 m, no more points are extracted at these two scan lines. In order to ensure that all road points are extracted and avoid non-road points being mistakenly extracted as road points, ρd is set to 0.03 m in our experiments.
Ideally, k with value 2 is enough to extract road marking points in EDEC, but some road marking segments may be missed because of the existence of smooth edges (some road marking segments are missed in #1, #2, #3 and #4 in Figure 17). With bigger k value, fewer road marking segments will be missed (none road marking segment is missed when k is set to 3, 4, and 5 in b–d), but more FRMPs appear in the result (it can be seen from #5, #6, #7, #8 and #9, #10, #11, #12 in Figure 17 where more FRMPs appear with the increase of k value). To extract road marking points more completely and to minimize the number of FRMPs, k is set to 3 in the experiments.
5.3. Road Points Extraction and Road Markings Extraction
The three data samples mentioned in Section 5.1 (data sample A, data sample B and data sample C) were used to evaluate the performance of the algorithms proposed in this paper. The parameters ρhd, ρd, k are analyzed in Section 5.2 and the configuration of these parameters in the experiments is based on the analysis. The minimum width of road markings along the platform moving direction is the width of stop line, which is 0.2 m in China [36], and the distance between two scan lines of the test data is about 0.06 meters, so ρs is set to 3 in the experiments according to Equation (10). The parameter configuration of the experiments is illustrated in Table 1.
Table 1.
ρhd | ρd | k | ρs |
---|---|---|---|
0.03 | 0.03 | 3 | 3 |
The data samples were first preprocessed to remove the isolated LiDAR points in the air and organized into scan lines, then seed road points were extracted from the scan line by HD and full road points were extracted by moving least squares line fitting. The seed road points extraction and full road points extraction results are illustrated in Figure 18 and Figure 19. Finally, the intensity values were filtered by dynamic window median filter and road marking points were extracted by EDEC. FRMPs are eliminated by segment and dimensionality feature based refinement. The road marking points extracted by EDEC before and after segment and dimensionality based refinement are illustrated in Figure 20 and Figure 22 respectively. To compare our method with other methods, the road markings in the three data samples were also extracted using Yongtao’s method [34]. In Yongtao’s method, the road points were divided into 4 blocks. In [34], dataset was collected by a Riegl VMX 450 mobile LiDAR system. Since we use a different mobile LiDAR system and the scanning parameters can hardly be the same, the ρSD in our experiments by Yongtao’s method was set to 11, which performs better in our dataset than the value recommended in [34]. The road markings extracted by Yongtao’s method and our method are illustrated in Figure 21 and Figure 22 respectively. Besides, the road markings in the three data samples were also manually classified from road points as reference data and illustrated in Figure 23. Quantitative evaluation of the extraction by our method and Yongtao’s method were conducted and the results are illustrated in Table 2. The computing time of our method for the three data samples are recorded to analyze the time complexity of our method and the results are listed in Table 3. As seen from the road points extraction results in Figure 18 and Figure 19, the seed road points are correctly extracted in every scan line in all three datasets, and the full road points are correctly extracted from three data samples except a part of one scan line missing in the road points extraction result of data sample B (#1 in Figure 18 and #4 in Figure 19). In Figure 19, there is no points in #2 and #5 because of moving car occlusion in data scanning and two cylindrical object placed on the edge of the road, respectively.
Table 2.
Data Sample | Data Sample A | Data Sample B | Data Sample C | Average | ||||
---|---|---|---|---|---|---|---|---|
Method | Yongtao’s | Our | Yongtao’s | Our | Yongtao’s | Our | Yongtao’s | Our |
cpt | 0.82 | 0.97 | 0.87 | 0.96 | 0.87 | 0.95 | 0.85 | 0.96 |
crt | 0.70 | 0.92 | 0.95 | 0.93 | 0.96 | 0.93 | 0.87 | 0.93 |
F | 0.76 | 0.94 | 0.91 | 0.94 | 0.91 | 0.94 | 0.86 | 0.94 |
Table 3.
Data Sample | Road Points Extraction | Road Markings Extraction and Refinement | Total |
---|---|---|---|
Data sample A | 48.81 | 6.94 | 55.75 |
Data sample B | 39.01 | 5.44 | 44.45 |
Data sample C | 33.82 | 7.60 | 41.42 |
As seen from Figure 20, all road markings appeared in Figure 19 can be identified in Figure 20. Besides the road markings appear in Figure 19, many FRMPs are also mistakenly extracted as road marking points in Figure 20. There is no points in #6 in Figure 20 as a result of car occlusion in data acquisition.
Compared with the manually labeled reference data, the road markings extracted by Yongtao’s method miss road marking points in #7, #8, #9, #11, #12 and #14 in Figure 21 and many FRMPs exist in #10 and #13 in Figure 21. #7 and #9 are missed because the points are divided into blocks in Yongtao’s method and the intensity of these points are much lower than that of other road markings belong to the same block, especially #7 which corresponds to #3 in Figure 19. We can clearly identify that the intensity values of points in #3 in Figure 19 are lower than the nearby road markings. The right side of data sample A and data sample B are emergency lanes where vehicles are not allowed to run on except during an emergency. Similarly, vehicles are not allowed to run on exit lines on the right side of data sample C, so the road surface on the right side is less worn, leading to high intensity values of the points scanned on these areas. Due to the high intensity values of points scanned on the right side, many points are mistakenly extracted as road marking points (#10 and #13 in Figure 21) and some of the road markings belong to the same block are missed (#11, #12, #14 in Figure 21). Our method only misses a small area in #15 in Figure 22 which corresponds to #3 in Figure 19. Even after segment and dimensionality refinement, some FRMPs still exist in our result (#16, #17, #18, #19, #20 and #21 in Figure 22).
The method similar to [30] was used to evaluate the performance of the proposed algorithms quantitatively. Instead of rasterizing the extracted points into 2D georeferenced images directly, we first extract the actual road marking points from the extracted points manually. The actual road marking points and the manually classified reference data were rasterized into 2D georeferenced images using the rasterization method in [30]. The completeness (cpt), correctness (crt) and F-measure are defined as follows:
(12) |
(13) |
(14) |
where Cpixel and Rfpixel are the pixel numbers in rasterized georeferenced images of actual road markings extracted and manually classified reference data, Cpoint and Epoint are the point numbers of actual road markings extracted and all road markings extracted. The results are illustrated in Table 2. It can be seen from Table 2 that the average cpt, crt and F-measure of Yongtao’s method are 0.85, 0.87 and 0.86 respectively and the average cpt, crt and F-measure of our method are 0.96, 0.93 and 0.94 respectively for these three data samples. The high cpt value of our method indicates that our method can extract road markings with little omission and the high crt value indicates that little FRMPs are remained in the extracted road marking points.
The computing time for road points extraction, road markings extraction and refinement are recorded to analyze the time complexity of our method. The proposed methods to extract road points and road marking points are implemented using C++ and running on an Intel(R) Core(TM) i7-3770 computer. Table 3 lists the computing time of the three data sample in seconds. It can be seen from the table that road points extraction is time consuming and the total time for each data sample is more than 40 seconds. The method proposed can be used to post-process mobile LiDAR point clouds which is not so critical for time complexity.
6. Conclusions
Extracting road markings from mobile LiDAR point clouds correctly and completely is still challenging. This paper proposed a scan line-based road markings extraction method from mobile LiDAR point clouds. The proposed method was applied to three data samples acquired by a mobile LiDAR system integrated by us. The experimental results indicated that the road points extraction method can effectively and accurately extract road points from the raw LiDAR point clouds, and the road markings extraction method can effectively extract road markings from the road points. The comparative study showed that the method proposed in this paper missed less road marking points, and less FRMPs are remained in the result. The quantitative study indicated that the method achieved average completeness, correctness, and F-measure of 0.96, 0.93, and 0.94 respectively, and outperformed Yongtao’s method.
Acknowledgments
This research is supported by Surveying and Mapping Public Service Special Research Project of China (Grant No. 201512008).
Author Contributions
Hua Liu and Zan Li conceived, designed the method and performed the experiment; Hua Liu and Junxiang Tan wrote the paper; Hong Xie revised the paper; Li Yan and Changjun Chen guided and supervised the overall process.
Conflicts of Interest
The authors declare no conflict of interest.
References
- 1.Novak K. The Ohio State University Highway Mapping System: The Stereo Vision System Component; Proceedings of the 47th Annual Meeting of the Institute of Navigation; Williamsburg, VA, USA. 10–12 June 1991; pp. 121–124. [Google Scholar]
- 2.Goad C.C. The Ohio State University Highway Mapping System: The Positioning Component; Proceedings of the 47th Annual Meeting of The Institute of Navigation; Williamsburg, VA, USA. 10–12 June 1991; pp. 117–120. [Google Scholar]
- 3.Schwarz K.P., Martell H.E., El-Sheimy N., Li R., Chapman M.A., Cosandier D. VIASAT—A Mobile Highway Survey System of High Accuracy; Proceedings of the Vehicle Navigation and Information Systems Conference; Ottawa, ON, Canada. 12–15 October 1993; pp. 476–481. [Google Scholar]
- 4.El-Sheimy N., Schwarz K.P. Kinematic Positioning in Three Dimensions Using CCD Technology; Proceedings of the Institute of Electrical and Electronics Engineers Vehicle Navigation and Information Systems Conference; Ottawa, ON, Canada. 12–15 October 1993; pp. 472–475. [Google Scholar]
- 5.Hock C., Caspary W., Heister H., Klemm J., Sternberg H. Architecture and Design of the Kinematic Survey System KiSS; Proceedings of the 3rd International Workshop on High Precision Navigation; Stuttgart, Germany. April 1995; pp. 569–576. [Google Scholar]
- 6.He G. Design and Application of the GPSVision Mobile Mapping System. Int. Arch. Photogram. Remote Sens. Spat. Inf. Sci. 2002;34:163–168. [Google Scholar]
- 7.Brown A. High Accuracy Targeting Using a GPS-Aided Inertial Measurement Unit; Proceedings of the ION 54th Annual Meeting; Denver, CO, USA. 1–3 June 1998. [Google Scholar]
- 8.Puente I., González-Jorge H., Martínez-Sánchez J., Arias P. Review of mobile mapping and surveying technologies. Measurement. 2013;46:2127–2145. doi: 10.1016/j.measurement.2013.03.006. [DOI] [Google Scholar]
- 9.Zhou L., Vosselman G. Mapping curbstones in airborne and mobile laser scanning data. Int. J. Appl. Earth Obs. Geoinform. 2012;18:293–304. doi: 10.1016/j.jag.2012.01.024. [DOI] [Google Scholar]
- 10.Pu S., Rutzinger M., Vosselman G., Oude Elberink S. Recognizing basic structures from mobile laser scanning data for road inventory studies. ISPRS J. Photogramm. 2011;66:S28–S39. doi: 10.1016/j.isprsjprs.2011.08.006. [DOI] [Google Scholar]
- 11.Guan H., Li J., Yu Y., Ji Z., Wang C. Using Mobile LiDAR Data for Rapidly Updating Road Markings. IEEE Trans. Intell. Transp. Syst. 2015;16:2457–2466. doi: 10.1109/TITS.2015.2409192. [DOI] [Google Scholar]
- 12.Soilán M., Riveiro B., Martínez-Sánchez J., Arias P. Traffic sign detection in MLS acquired point clouds for geometric and image-based semantic inventory. ISPRS J. Photogramm. 2016;114:92–101. doi: 10.1016/j.isprsjprs.2016.01.019. [DOI] [Google Scholar]
- 13.Riveiro B., Diaz-Vilarino L., Conde-Carnero B., Soilan M., Arias P. Automatic Segmentation and Shape-Based Classification of Retro-Reflective Traffic Signs from Mobile LiDAR Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015;9:295–303. doi: 10.1109/JSTARS.2015.2461680. [DOI] [Google Scholar]
- 14.Ishikawa K., Takiguchi J., Amano Y., Hashizume T., Fujishima T. Tunnel Cross-Section Measurement System Using a Mobile Mapping System. J. Robot. Mechatron. 2009;21:193–199. [Google Scholar]
- 15.Holgado-Barco A., Gonzalez-Aguilera D., Arias-Sanchez P., Martinez-Sanchez J. An automated approach to vertical road characterisation using mobile LiDAR systems: Longitudinal profiles and cross-sections. ISPRS J. Photogramm. 2014;96:28–37. doi: 10.1016/j.isprsjprs.2014.06.017. [DOI] [Google Scholar]
- 16.Alho P., Vaaja M., Kukko A., Kasvi E., Kurkela M., Hyyppä J., Hyyppä H., Kaartinen H. Mobile laser scanning in fluvial geomorphology: Mapping and change detection of point bars. Z. Geomorphol. Suppl. Issues. 2011;55:31–50. doi: 10.1127/0372-8854/2011/0055S2-0044. [DOI] [Google Scholar]
- 17.Vaaja M., Hyyppä J., Kukko A., Kaartinen H., Hyyppä H., Alho P. Mapping Topography Changes and Elevation Accuracies Using a Mobile Laser Scanner. Remote Sens. 2011;3:587–600. doi: 10.3390/rs3030587. [DOI] [Google Scholar]
- 18.Hohenthal J., Alho P., Hyyppa J., Hyyppa H. Laser scanning applications in fluvial studies. Prog. Phys. Geogr. 2011;35:782–809. doi: 10.1177/0309133311414605. [DOI] [Google Scholar]
- 19.Arastounia M. Automated Recognition of Railroad Infrastructure in Rural Areas from LIDAR Data. Remote Sens. 2015;7:14916–14938. doi: 10.3390/rs71114916. [DOI] [Google Scholar]
- 20.Liang X., Hyyppa J., Kukko A., Kaartinen H., Jaakkola A., Yu X. The use of a mobile laser scanning system for mapping large forest plots. IEEE Geosci. Remote Sens. Lett. 2014;11:1504–1508. doi: 10.1109/LGRS.2013.2297418. [DOI] [Google Scholar]
- 21.Lin Y., Hyyppä J. Multiecho-recording mobile laser scanning for enhancing individual tree crown reconstruction. IEEE Trans. Geosci. Remote Sens. 2012;50:4323–4332. doi: 10.1109/TGRS.2012.2194503. [DOI] [Google Scholar]
- 22.Rutzinger M., Pratihast A.K., Oude Elberink S.J., Vosselman G. Tree modelling from mobile laser scanning data-sets. Photogramm. Record. 2011;26:361–372. doi: 10.1111/j.1477-9730.2011.00635.x. [DOI] [Google Scholar]
- 23.Xiao W., Vallet B., Schindler K., Paparoditis N. Street-side vehicle detection, classification and change detection using mobile laser scanning data. ISPRS J. Photogramm. 2016;114:166–178. doi: 10.1016/j.isprsjprs.2016.02.007. [DOI] [Google Scholar]
- 24.Xiao W., Vallet B., Brédif M., Paparoditis N. Street environment change detection from mobile laser scanning point clouds. ISPRS J. Photogramm. 2015;107:38–49. doi: 10.1016/j.isprsjprs.2015.04.011. [DOI] [Google Scholar]
- 25.Leslar M., Leslar M. Extraction of geo-spatial information from LiDAR-based mobile mapping system for crowd control planning; Proceedings of the 2009 IEEE Toronto International Conference on Science and Technology for Humanity (TIC-STH); Toronto, ON, Canada. 26–27 September 2009; pp. 468–472. [Google Scholar]
- 26.Kashani A., Olsen M., Parrish C., Wilson N. A Review of LIDAR Radiometric Processing: From Ad Hoc Intensity Correction to Rigorous Radiometric Calibration. Sensors. 2015;15:28099–28128. doi: 10.3390/s151128099. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Tan K., Cheng X. Intensity data correction based on incidence angle and distance for terrestrial laser scanner. J. Appl. Remote Sens. 2015;9:94094. doi: 10.1117/1.JRS.9.094094. [DOI] [Google Scholar]
- 28.Yang B., Fang L., Li Q., Li J. Automated Extraction of Road Markings from Mobile Lidar Point Clouds. Photogramm. Eng. Remote Sens. 2012;78:331–338. doi: 10.14358/PERS.78.4.331. [DOI] [Google Scholar]
- 29.Kumar P., McElhinney C.P., Lewis P., McCarthy T. An automated algorithm for extracting road edges from terrestrial mobile LiDAR data. ISPRS J. Photogramm. 2013;85:44–55. doi: 10.1016/j.isprsjprs.2013.08.003. [DOI] [Google Scholar]
- 30.Guan H., Li J., Yu Y., Wang C., Chapman M., Yang B. Using mobile laser scanning data for automated extraction of road markings. ISPRS J. Photogramm. 2014;87:93–107. doi: 10.1016/j.isprsjprs.2013.11.005. [DOI] [Google Scholar]
- 31.Riveiro B., González-Jorge H., Martínez-Sánchez J., Díaz-Vilariño L., Arias P. Automatic detection of zebra crossings from mobile LiDAR data. Opt. Laser Technol. 2015;70:63–70. doi: 10.1016/j.optlastec.2015.01.011. [DOI] [Google Scholar]
- 32.Jaakkola A., Hyyppä J., Hyyppä H., Kukko A. Retrieval Algorithms for Road Surface Modelling Using Laser-Based Mobile Mapping. Sensors. 2008;8:5238–5249. doi: 10.3390/s8095238. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Chen X., Stroila M., Wang R., Kohlmeyer B., Alwar N., Bach J. Next Generation Map Making: Geo-Referenced Ground-Level LIDAR Point Clouds for Automatic Retro-Reflective Road Feature Extraction; Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems; Seattle, DC, USA. 4–6 November 2009. [Google Scholar]
- 34.Yu Y., Li J., Guan H., Fukai J., Wang C. Learning Hierarchical Features for Automated Extraction of Road Markings From 3-D Mobile LiDAR Point Clouds. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015;8:709–726. doi: 10.1109/JSTARS.2014.2347276. [DOI] [Google Scholar]
- 35.Levin D. The approximation power of moving least-squares. Math. Comput. Am. Math. Soc. 1998;67:1517–1531. doi: 10.1090/S0025-5718-98-00974-0. [DOI] [Google Scholar]
- 36.GB51038–2015. China Planning Press; Beijing, China: 2015. Code for Layout of Urban Road Traffic Signs and Markings; p. 290. [Google Scholar]
- 37.Weinmann M., Jutzi B., Hinz S., Mallet C. Semantic point cloud interpretation based on optimal neighborhoods, relevant features and efficient classifiers. ISPRS J. Photogramm. 2015;105:286–304. doi: 10.1016/j.isprsjprs.2015.01.016. [DOI] [Google Scholar]
- 38.Han J., Chen C., Lo C. Time-Variant Registration of Point Clouds Acquired by a Mobile Mapping System. IEEE Geosci. Remote Sens. Lett. 2014;11:196–199. doi: 10.1109/LGRS.2013.2252417. [DOI] [Google Scholar]
- 39.Takai S., Date H., Kanai S., Niina Y., Oda K., Ikeda T. Accurate registration of MMS point clouds of urban areas using trajectory. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2013;II-5/W2:277–282. doi: 10.5194/isprsannals-II-5-W2-277-2013. [DOI] [Google Scholar]
- 40.Yu F., Xiao J.X., Funkhouser T. Semantic Alignment of LiDAR Data at City Scale; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Boston, MA, USA. 8–10 June 2015; pp. 1722–1731. [Google Scholar]