Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2021 Sep 29;21(19):6501. doi: 10.3390/s21196501

Calibration of Planar Reflectors Reshaping LiDAR’s Field of View

Michał Pełka 1, Janusz Będkowski 2,*
Editor: Mengdao Xing
PMCID: PMC8512862  PMID: 34640828

Abstract

This paper describes the calibration method for calculating parameters (position and orientation) of planar reflectors reshaping LiDAR’s (light detection and ranging) field of view. The calibration method is based on the reflection equation used in the ICP (Iterative Closest Point) optimization. A novel calibration process as the multi-view data registration scheme is proposed; therefore, the poses of the measurement instrument and parameters of planar reflectors are calculated simultaneously. The final metric measurement is more accurate compared with parameters retrieved from the mechanical design. Therefore, it is evident that the calibration process is required for affordable solutions where the mechanical design can differ from the inaccurate assembly. It is shown that the accuracy is less than 20 cm for almost all measurements preserving long-range capabilities. The experiment is performed based on Livox Mid-40 LiDAR augmented with six planar reflectors. The ground-truth data were collected using Z + F IMAGER 5010 3D Terrestrial Laser Scanner. The calibration method is independent of mechanical design and does not require any fiducial markers on the mirrors. This work fulfils the gap between rotating and Solid-State LiDARs since the field of view can be reshaped by planar reflectors, and the proposed method can preserve the metric accuracy. Thus, such discussion concludes the findings. We prepared an open-source project and provided all the necessary data for reproducing the experiments. That includes: Complete open-source code, the mechanical design of reflector assembly and the dataset which was used in this paper.

Keywords: LiDAR, ICP, mapping, calibration, reshape field of view, solid state LiDAR

1. Introduction and Related Work

LiDAR (light detection and ranging) in general is the method for determining ranges (variable distance) between an object and a laser by measuring the time it takes for reflected light to return to the receiver. This measurement instrument is mainly used for the 3D digitization of the urban environment, cultural heritage and archaeology, underground environment, environmental monitoring, forestry and agriculture [1]. The purpose of the study was to verify an experimentally proposed novel calibration process capable of preserving LiDAR’s accuracy and range after reshaping the FOV (Field Of View) with planar reflectors. Base on our best knowledge such work has not yet been elaborated in the literature. We use Livox Mid-40 (manufacturer Livox, Hong Kong, China, https://www.livoxtech.com/): The robotic LiDAR sensor based on incommensurable scanning that allows straightforward mass production and adoption in autonomous robots [2]. There are many types of LiDAR applications such as Terrestrial Laser Scanning, which uses a highly accurate measurement instrument that works in the so-called stop-scan fashion looking from a mobile robotics point of view [3]. This means that the robot stops at certain goals to acquire highly accurate 3D measurements. Another family of applications is mobile mapping systems composed of highly accurate planar LiDARs, or 3D multi-beam lasers such as Velodyne. An overview of scanning and reconstruction methods is discussed in Lehtola et al. [4]. Recently, the development of Solid-State LiDARs shows the potential reduction of such devices. Typical data produced by LiDAR are measurements of distance and reflectance; thus white reflective objects such as lines or signs can be identified easily. An important aspect of the multi beam LiDAR is the intrinsic calibration [5,6,7]. Such calibration is crucial in mobile mapping applications where the measurement instrument is heavily occupied and works assuming diverse environmental conditions. The gap between rotating and Solid-State LiDARs is evident since the field of view of Solid-State LiDARs is still limited. Review [8] shows an extensive overview of MEMS (Micro Electro Mechanical Systems) scanning mirrors specifically for applications in LiDAR systems. Such technology will improve Solid-State LiDARs but it requires a sophisticated laboratory and the time from the design to the delivery is rather a barrier in fast prototyping. For this reason, the investigation of reshaping the field of view of the LiDAR with planar reflectors is the main topic of the paper. Many other researchers proposed different methods incorporating the rotation of the LiDAR [9,10]. In such cases, the rotation of the LiDAR requires additional mechanics and electronics. For such approaches the synchronization of the LiDAR data with GPS (Global Positioning System) and odometry is crucial. This leads our research to reduce the number of additional mechanical and electronic parts for reshaping LiDAR’s field of view. Therefore, the overall design is simplified and the final prototype is more affordable compared with competitive solutions.

Reshaping the field of view of the sensors is an interesting research topic since it allows to customize a specific application for certain needs. For example, authors Endres et al. [11] introduced the combination of an RGB-D camera with two planar mirrors to split the field of view. It covers both the front and rear views of a mobile robot. They describe how to estimate the extrinsic calibration parameters of the modified sensor using a standard parametrization. For solving the graph SLAM (simultaneous localization and mapping) optimization problem the g2o framework [12] is used. More information concerning the graph SLAM can be found in [13,14,15,16]. An alternative approach is discussed in Akay et al. [17] where a proposed solution employs mirrors to introduce virtual RGB-D cameras into the system. The proposed system does not have any space limitations, data bandwidth constraints or synchronization problems and it is more affordable since it does not require extra cameras. The authors developed formulations for the simultaneous calibration of real and virtual RGB and RGB-D cameras and provided methods for 3D reconstruction from these cameras. It is worth mentioning that RGB-D mapping and 3D reconstruction have been intensively researched topics in recent years. An interesting comparative study of registration methods for RGB-D video of static scenes is discussed in Morell-Gimenez et al. [18].

The authors Aalerud et al. [19] proposed a method for the simulation and design of a radially reshaped field of view. This work shows great potential for an increased number of usable measurements. Unfortunately, the authors did not address the problem of decreased accuracy related to the not ideal geometrical placement of the reflectors. Moreover, it is difficult to find relevant discussion in the literature [20]; therefore, we address this problem by providing an end-to-end framework for the calibration of the geometric placement of the reflectors. Adding reflectors can improve the mobile robotic perception [21] by augmenting planar 2D measurements; thus, Unmanned Aerial Vehicles can easily navigate in indoor environments. Mirrors can be used for extending the LiDAR field of view in self-driving vehicles [22]. Generally, it is easy to imagine plenty of applications requiring changing the LiDAR field of view, e.g., security applications, autonomous driving, etc. For this reason, we propose a general framework that can be used for other researchers in developing a specific LiDAR field of view configuration.

This topic is relevant looking at recent developments in mobile robotics such as SLAM and search and rescue applications where reconstructing 3D scenes is one of the goals of the autonomous machine and autonomous cars that will use a predefined map for localization purposes [23,24]. Obviously, it is evident that autonomous cars can collect data and contribute to global map updates; thus, we cope with a large-scale problem that has recently been addressed by many researchers. The term SLAM [25,26] corresponds to the so-called “chicken and egg dilemma”—what was first: The chicken or the egg? Therefore we should have a proper map representation that is compatible with observations derived from sensors to localize within this map, and we need accurate localization to build the map. In the literature [12], the SLAM problem is divided into so-called front-end and back-end. The front-end is typically responsible for initial trajectory generation and the back-end is responsible for so-called loop closing. Loop is the situation when the measurement instrument visits the same location as visited some time ago, assuming a continuous trajectory between these two time intervals. The back-end is typically solved using a so-called graph-based technique that minimizes the error between the observed and desired relative pose between trajectory/trajectory nodes. Due to the large scale aspect of SLAM, it can be mentioned that recent research shows an interest in cloud-based calculations of back-end [25,27,28].

The objective and at the same time the rationale addressing the research problem is lowering the cost of LiDAR sensors. This is related to reshaping the field of view, which can be done with planar reflectors. Such an approach requires additional calibration due to the fact of assembly and manufacturing imperfections. Moreover, some applications such as autonomous mobile robots equipped with such sensors will deal with the problem of self-calibration during daily operations. For this reason, we propose a novel calibration process and an open-source approach. It fulfils the gap in affordable LiDAR sensors with a reshaped field of view. Based on our best knowledge such an approach has not yet been elaborated in the literature; many studies even used additional planar reflectors for LiDARs. Our future work will be investigating the usage of presented LiDAR sensors for mobile robot localization; therefore, this research contributes to affordable mobile robotics.

2. Materials and Methods

2.1. Mechanical Design

The Livox Mid-40 LiDAR has an originally conical field of view with the apex in the optical centre and an apex angle equal to 38.4 degrees. Relevant parameters of this LiDAR [29] are shown in Table 1. This table shows also Z + F IMAGER 5010 [30] parameters as a reference ground-truth source. Figure 1 shows a prototype used in experiments and simulations of the reshaped field of view. The field of view after modification consists of six segments that have properties:

  • The field of view (vertical) spreads from −12° to 9.4° (Figure 2).

  • The field of view (horizontal) has six segments orientated radially with an angle up to 18.7° (Figure 3).

Table 1.

Livox Mid-40 [29] and Z + F IMAGER 5010 [30] parameters. (*) Measured in an environment of 25 °C with a target (80% reflectivity) 20 m away. The result may vary under different test conditions.

Parameter Livox Mid-40 Z + F IMAGER 5010
Laser Wavelength 905 nm 1500 nm
Laser Safety Class 1 (IEC60825-1) Class 1
Detection Range (@ 100 klx) 90 m @ 10% reflectivity 187.3 m
130 m @ 20% reflectivity
260 m @ 80% reflectivity
FOV 38.4° Circular 320° Vertical, 360° Horizontal
Range Precision (1σ @ 20 m) 2 cm (*) 1 mm

Figure 1.

Figure 1

Prototype of LiDAR with planar reflectors.

Figure 2.

Figure 2

Reshaped field of view; plot of vertical situation.

Figure 3.

Figure 3

Reshaped field of view; plot of horizontal situation.

Due to the limited precision of the mirror assembly (the main contributor is the thickness of the top PMMA (PolyMethyl MethAcrylate) surface of the mirror, the usable angular range is limited. Therefore, measurement outliers are evident when a given ray is assigned to the wrong mirror. The reflector assembly was designed using the Autodesk Fusion 360 software [31]. It is CAD (Computer Aided Design) software that helps in the mechanical design process. It allows producing 3D models for 3D printing and necessary documentation for further manufacturing. The designed structure consists of the main chassis (which resembles a hexagonal pyramid) with a set of mirrors screwed into the pyramid’s faces. The screws’ holes are placed outside the useful surface of the mirror (marked as 1 in Figure 4). The main chassis was 3D printed using a consumer-grade FDM (Fused Deposition Modeling) printer with PLA (PolyLactic Acid) filament. The mirrors’ shape is obtained with a CNC plotter from the PMMA sheet with the reflective bottom surface (marked as 4 in Figure 4). The original laser scanner has a front plate attached. The front plate (marked as 3 in Figure 4) has a rectangular hole for lasers’ scanner optics and multiple holes for assembly—ϕ 3 mm for mounting laser scanner and ϕ 6 mm for three pillars. The hexagonal pyramid is attached to pillars.

Figure 4.

Figure 4

Three-dimensional CAD model of assembly. 1—hexagonal pyramid; 2—Livox LiDAR; 3—face plate; 4—one of six mirrors.

2.2. Calibration Method

Due to the imperfect assembly of the planar reflectors, the metric measurement is poor. To solve this issue the calibration method is designed and implemented as an open-source project [32]. It is composed of observation equations implemented with Ceres solver [33]. It is based on the commonly used Iterative Closest Point procedure [34,35,36]. The calibration method is independent of mechanical design and does not require any fiducial markers on the mirrors like in Chen et al. [22].

2.2.1. Data Acquisition

Due to six planar reflectors, it is advised to use stop-scan fashion in different locations assuming additional rotation of the measurement instrument around its axis (Figure 5). For this purpose we used a precise rotating table; therefore, we collected 36 static measurements for each measurement station by rotating the table by 10 degrees. Each static measurement consists of three seconds of the recorded Livox Mid-40 data. In such a scheme of data acquisition, there is a sufficient number of overlapping data of different mirrors. Additionally, the entire scene was scanned using Terrestrial Laser Scanning measurement instrument Z + F IMAGER 5010 (manufacturer: Z + F, 88239 Wangen, Germany, https://www.zofre.de/) (range uncertainty of 1 mm [30]). This ground-truth data is an order of magnitude more accurate than the experimental prototype.

Figure 5.

Figure 5

Experimental prototype mounted onto precise rotating table during data acquisition procedure. A—axis for changing rotation angle.

2.2.2. Iterative Closest Point

Local point Pl(xl,yl,zl,1) is represented in Euclidean space as the point in local reference frame. The matrix [R, T] is the transformation of local point Pl into point Pg(xg,yg,zg) in global reference frame; thus

Pg=[R,T]Pl (1)
ΨRj,Tj(Rj,Tj,xil,j,yil,j,zil,j)=(xig,j,yig,j,zig,j)=r11jr12jr13jt1jr21jr22jr23jt2jr31jr32jr33jt3jxil,jyil,jzil,j1 (2)
xδyδzδresiduals=000target valuesΨRj,Tjβj(Rj,Tj,xl,j,yl,j,zl,j)ΨRk,Tkβk(Rk,Tk,xl,k,yl,k,zl,k)model function (3)

Equation (2) transforms the ith local point (xil,j,yil,j,zil,j,1) in jth [R, T] into global reference system (xig,j,yig,j,zig,j); therefore, it can be used for building point to point observation Equation (3), incorporating two poses [R1,T1] and [R2,T2], where xδyδzδ are residuals, 000 are target values and ΨRj,Tj(Rj,Tj,xl,j,yl,j,zl,j)ΨRk,Tk(Rk,Tk,xl,k,yl,k,zl,k) is the model function. The Iterative Closest Point optimization problem for point to point observations is defined as Equation (4), where there are C pairs of points contributing to the optimization process.

minRj,Tj,Rk,Tki=1C000ΨRj,Tj(Rj,Tj,xil,j,yil,j,zil,j)ΨRk,Tk(Rk,Tk,xil,k,yil,k,zil,k)2 (4)

The Iterative Closest Point approach converges to the optimal solution in an iterative fashion.

2.2.3. Line with Plane Intersection

Each LiDAR beam intersects with the reflective plane. The following plane equation is considered:

ax+by+cz+d=0 (5)

where

abc=1 (6)

Vpl=(a,b,c) is the unit vector orthogonal to the plane and d is the distance from the origin to the plane. It satisfies the following condition with the intersection point in 3D space:

abcdxyz1=0 (7)

The main assumption is that all LiDAR beams starts at origin (0,0,0). Assuming the LiDAR beam has unique representation as beam origin bo=(0,0,0) and beam direction bd=(bxd,byd,bzd), bd=1 the beam with reflective plane intersection equation is given (8).

Pint=bxdbydbzddVpl·bd (8)

where (·) is the dot product.

2.2.4. Reflection Observation Equation

The direction rd of the LiDAR beam bd after reflecting with the plane with the normal vector Vpl is given by Equation (9) and the supportive plot is given in Figure 6.

rd=2(bd·Vpl)Vplbd (9)
Figure 6.

Figure 6

The direction rd of the LiDAR beam bd after reflecting with the plane with normal vector Vpl.

With the measurement point Pl(xl,yl,zl) expressed in the local coordinate system of the measurement instrument, lp=Pl and distance from origin to intersection lint=Pint, the final reflected measurement point Pr is given in Equation (10).

Pr=(Pint+rd(lplint)) (10)

Finally, the measurement point expressed in the global coordinate system is given as (11).

Pg=[R,T]Pr=Ψr(R,T,a,b,c,d,Pl) (11)

2.2.5. Ground-Truth Data Observation Equation

For ground-truth data as a point cloud composed of points Pgt(xgt,ygt,zgt) expressed in global reference systems, the ground-truth data observation equation is given by Equation (12).

xδyδzδresiduals=000target valuesΨRj,Tj,am,bm,cm,dmβj,mr(Rj,Tj,am,bm,cm,dm,Pl,j)xgt,kygt,kzgt,kmodel function (12)

The optimization process will not modify Pgt,k. The expected result will converge to obtain minimal distances (target values) between measurement points Pl transformed into a global reference system via Ψr and the corresponding ground-truth Pgt.

2.2.6. Calibration Algorithm

The angle information from a precise rotating table is used as the additional constraints. The rotation axis of the rotating table and the main optical axis of the calibrated system need to be as close together as possible. Ideally, it should be identical; however, a small discrepancy is acceptable. This geometrical discrepancy is represented by a homogenous transformation [Rlt,Tlt]SE(3). That homogenous transformation transforms the Livox LiDAR’s local coordinate system to the local coordinate system of the rotating table’s plate. Ideally, it should be the identity, but due to a lack of coaxiality of rotation and optical axis, it represents some unknown, small displacement. Homogenous transformation [Rlt,Tlt] is treated as an unknown and stationary parameter during calibration. In other words, it is the same for all measurements—Livox LiDAR is rigidly assembled to the rotating table’s plate during calibration. The second homogenous transformation is [Rtp,Ttp]SE(3), which represents a rotation angle of the rotating table. It is a rotation around the ‘X’ axis with a given, known angle. This homogenous transformation [Rtp,Ttp] is treated as a known parameter during the calibration process. In our experiment, it is an angle that was set on the rotating table. The third transformation is [Rpg,Tpg]SE(3). It transforms a stationary measurement station’s local coordinate system to the global coordinate system. It is treated as an unknown parameter during the calibration process. The reflected point (Pr) in Equation (11) is transformed into the global coordinate system (Pg) with a chain of transformations:

Pg=[Rpg,Tpg][Rtp,Ttp][Rlt,Tlt]Pr (13)

Finally, Equation (11), taking into account the chain of SE(3) transforms, has the form:

Pg=[Rpg,Tpg][Rtp,Ttp][Rlt,Tlt]Pr=Ψg(Rpg,Tpg,Rtp,Ttp,Rlt,Tlt,a,b,c,d,Pl) (14)

Each of those homogenous transformations belongs to a special Euclidean group SE(3). SE(3) is a Lie group. That means it is possible to represent the given transformation in SE(3) as se(3) algebra. That is a common technique used in the optimization of non-linear problems that contains parameters that belong to SO(3) or SE(3) groups. Further, necessary information can be found in Sola et al. [37]. Properties of Lie algebra se(3) and Lie group SE(3) that are utilized are listed below:

  • the point in tangent space to the manifold is a minimal representation (six degrees of freedom).

  • the point in tangent space does not have any constraints and can be always exponentially mapped to a valid SE(3).

  • every valid SE(3) member can be mapped back to an exact point in tangent space (logarithmic mapping). It is computed using a closed-form solution (Euler–Rodrigues formula).

  • every optimized SE(3) transformation contributes six parameters to the optimization problem.

Optimization of all SE(3) parameters is done in the tangent domain. To manipulate an SE(3) and se(3) Sophus C++ library was used [38]. For modelling and optimization of the calibration problem, Ceres solver was used [33]. A mirror is represented in the optimization problem as a plane given with Equation (5). In the calibration process, multiple nearest neighbourhood searches are performed. For this problem the well-known [39] Kd-tree algorithm was utilized. It allows effectively to decompose a point cloud and find the nearest point to a query point.

The calibration algorithm for single measurement station without ground-truth data (the simplest scenario):

  • with current calibration point clouds and Kd-trees are built in the global coordinate system.

  • using Kd-tree, it searches for pairs of the nearest neighbourhood points that were reflected by a different mirror.

  • every found nearest neighborhood point pair (Pkg,Pjg) creates an observation equation rjkR3. The point Pkg was observed with the measurement Pkl (taken in the instrument’s local coordinate system), reflected by the mirror φ and the laser scanner was at rotation ϕ. Mirror φ is represented by its plane parameters via: aφ,bφ,cφ,dφ. Current rotation of the rotating table is represented with homogenous transformation [Rlϕt,Tlϕt]SE(3). Finally, the residual for the point pair (Pkg,Pjg) is given by Equation (15).
    rjk=PjgΨg(Rlϕt,Tlϕt,Rlt,Tlt,aφ,bφ,cφ,dφ,Pkl) (15)
  • Every found pair contributes a new residual. The number of those equations creates an optimization problem. The equation is (15) and is differentiated automatically against all optimized parameters, which are: Rlt,Tlt,aφ,bφ,cφ,dφ

  • The optimization problem is solved using the Levenberg–Marquardt algorithm until convergence using the Ceres solver.

  • New, the found parameters are applied correctly according to its parametrization and the whole cycle is repeated.

The calibration algorithm for multiple measurement stations uses the stations’ poses [Rpg,Tpg]. In this scenario, if the nearest neighbourhood consists of points from the same measurement station, it will contribute an observation where parameters corresponding to its pose are treated as constant ([Rpg,Tpg]). Otherwise, if the nearest neighbourhood exists between points that were captured from different measurement stations, it will contribute also to these poses an optimization. Finally, if there is ground-truth data the observation Equation (12) can be added. This calibration scenario results in calibration parameters and poses updates; thus, it can be used for 3D map reconstruction.

2.2.7. Calibration Accuracy Evaluation

The main assumption of the calibration process is the fact that ground-truth data are used only for validation purposes. The evaluation is based on quantitative and qualitative measures of the discrepancy between ground-truth and LiDAR data (Livox Mid-40 with calibrated planar reflectors). The qualitative result is shown in Figure 7 as multiple views of the same features in the scene. The expected outcome of the validation of the calibrated sensor is the consistent view of the cross-section. Moreover, Figure 8 shows the degradation of the identification of the corner shapes. Thus, the negative impact of planar reflectors on LiDAR’s performance is evaluated as the discrepancy between measurements obtained with Livox Mid-40 and the same LiDAR with planar reflectors. Quantitative evaluation is based on measuring error distribution. The error is expressed as the distance of the measured point to its projection onto approximated planar features in ground-truth data. Error distribution is shown in the form of histograms (Figure 9 and Figure 10). Moreover, the quantitative evaluations were performed for planar target detection (Figure 11, Table 2). The goal was to verify the accuracy (measured as mean distance to the target plane) and precision (measured as the standard deviation of distance to the target plane).

Figure 7.

Figure 7

Qualitative result of the calibration impact on improved 3D data accuracy.

Figure 8.

Figure 8

Degradation of LiDAR’s perception with planar reflector. Black—Ground-truth; Green—Livox Mid-40 LiDAR data; Red—Livox Mid-40 LiDAR data with planar reflector. The degradation of sharp corners is evident. The planar target is detected with precision (1σ) 1.8 cm (without reflector) and 2.2 cm (with reflector).

Figure 9.

Figure 9

Result of calibration algorithm for single measurement station without ground-truth data (simplest scenario).

Figure 10.

Figure 10

Result of the calibration algorithm for multiple measurement stations with histogram of errors.

Figure 11.

Figure 11

Location of planar targets that are evaluated in Table 2.

Table 2.

The evaluation of accuracy and precision of planar targets detection using Livox Mid-40 with planar reflector. Locations of planar targets are shown in Figure 11.

Plane Number Accuracy [cm]
(Mean Distance to the Plane)
Precision [cm]
(Standard Deviation of Distance)
Distance to Sensor [m]
1 2.95 1.16 18.3
2 0.32 2.01 20.2
3 0.22 1.26 15.5
4 2.07 2.73 35.7
5 3.26 3.05 60.3
6 2.01 1.20 24.5

3. Results

To validate the proposed calibration method the system was tested in a known indoor environment. The environment was mapped with the geodetic measurement instrument Z + F Imager 5010 with accuracy much better than Livox Mid-40. These data are considered as ground truth. Two calibration scenarios were investigated: (1) Single measurement station without ground-truth data; (2) multiple measurement stations. We decided to not add ground-truth data to the optimization process to simulate operational conditions. Ground-truth data are used for qualitative and quantitative evaluation. The qualitative result of the calibration impact on improved 3D data accuracy is shown in Figure 7.

Figure 9 shows the results of the calibration algorithm for a single measurement station without ground-truth data (simplest scenario). Figure 10 shows the results for multiple measurement stations. We observed that calibration of planar reflectors can be done based on a single measurement station. Better calibration results can be reached for multiple measurement stations, but it is a more complicated procedure requiring more time for data acquisition. The better calibration outcome can be seen comparing histograms Figure 9a and Figure 10b. Calibration found with multiple measurement stations has unimodal distribution of error and the most common error value is smaller. Calibration with multiple measurement stations is more robust and performs significantly better. Calibration with single measurement stations is much easier to perform, although it is susceptible to convergence to a sub-optimal solution. Both distributions of the errors (Figure 9a and Figure 10b) are Gaussian with large positive skewness. Long-tail is present in distributions. Both are caused by outliers. Deviation of the optimized parameters before and after calibration is shown in Table 3. For example, the angular update of the first planar reflector is 0.65°. This is an important insight into the calibration impact. Unfortunately, planar reflectors generate outlier points and some artifacts affecting point cloud quality. This is shown in Figure 8. The accuracy and the precision were evaluated based on six planar ground-truth targets shown in Figure 11. The results confirm minor degradation of the accuracy and the precision compared with reported parameters by the Livox vendor (Table 1 [29]).

Table 3.

The optimized parameters (planar reflectors’ coefficients); u: Unitless quantity; m: Meter; [a,b,c]: Unit vector of planar reflector; d: Distance of planar reflector to the origin.

Parameter Initial Calibrated
[a1,b1,c1,d1] [0.793 u, −0.304 u, −0.527 u, −0.075 m] [0.799 u, −0.311 u, −0.519 u, −0.044 m]
[a2,b2,c2,d2] [−0.793 u, 0.609 u, −0.000 u, 0.075 m] [−0.787u,0.614u, −0.012u,0.049m]
[a3,b3,c3,d3] [0.793 u, −0.304 u, 0.527 u, −0.075 m] [0.791 u, −0.301 u, 0.531 u, −0.049 m]
[a4,b4,c4,d4] [0.793 u, 0.304 u, 0.527 u, −0.075 m] [0.789 u, 0.311 u, 0.528 u, −0.049 m]
[a5,b5,c5,d5] [−0.793 u, −0.609 u, −0.000 u, 0.075 m] [−0.798 u, −0.605 u, 0.008 u, 0.045 m]
[a6,b6,c6,d6] [0.793 u, 0.304 u, −0.527u, −0.075 m] [0.793 u, 0.291 u, −0.534 u, −0.047 m]

For this reason, in future work, we will focus on filtering algorithms.

4. Discussion

The calibration process of planar reflectors reshaping LiDAR’s field of view enables preserving accuracy and precision of the LiDAR to some extend. Based on our best knowledge such an investigation has not yet been discussed in the literature. We demonstrated by the experiment that the calibration process is required for affordable solutions where the mechanical design can differ from the inaccurate assembly. We use a state-of-the-art LiDAR Livox Mid-40 sensor as an object of investigation of the reshaping field of view with planar reflectors. Ground-truth data were collected with a precise Z + F IMAGER 5010 terrestrial laser scanning system. The calibration process incorporates the state-of-the-art reflection observation equation integrated with Iterative Closest Point optimization. We observed that adding planar reflectors slightly degrades the precision of the measurements; in particular, detected corners are slightly curved. This observation yields the conclusion that such a sensor will be rather useful for localization purposes than environmental mapping. Moreover, future autonomous mobile robots will require such calibration if their LiDAR FOV will be reshaped by planar reflectors.

The impact of planar reflectors on another popular LIDAR (Velodyne VLP-16) was investigated in detail in the paper [19]. In their work, the authors investigated power loss and LiDAR’s scanning pattern after FOV reshaping. They attempted to reshape the original omnidirectional FOV to a narrow FOV maximizing the angular resolution of the system. They performed simulations and examined the built prototype. They concluded that FOV reshaping using a planar reflector is feasible and yields a reduction in the range of 3.9% (when the incident angle to the target is preserved). In our work, we built a similar setup using alike technology of prototyping, but we investigated the unaddressed issue of planar reflector calibration. We also introduced a slightly different LiDAR to that used in the discussed paper. The difference between Velodyne and Livox Mid-40 LiDARs can be found in the paper [2].

Our work contributes to omnidirectional perception based on 3D LiDAR sensor similarly to [40]. The authors developed a 3D laser scanner based on a SICK LMS 200 LiDAR, which consists of the LMS 200 facing upwards into a rotating mirror driven by a stepper motor. It is worth mentioning the work [41] ona 3D imaging LiDAR based on the high-speed 2D laser scanner and the work [42] on the automatic calibration of spinning actuated LiDAR internal parameters. These approaches relate also to omnidirectional perception derived from spinning LiDARs. Such an approach is not affordable; moreover, introducing other moving parts into the design decreases the robustness of the entire system. A recent survey [43] of low-cost 3D laser scanning technology discusses other approaches; unfortunately, the calibration of the planar reflectors reshaping LiDAR’s field of view is not elaborated. For this reason, we are focused on this aspect in this paper.

5. Conclusions

This paper describes the calibration process of planar reflectors reshaping LiDAR’s field of view. The calibration method is based on the reflection equation used in the Iterative Closest Point optimization. The final metric measurement is more accurate compared with parameters retrieved from the mechanical design. The experiment is performed based on Livox Mid-40 LiDAR augmented with six planar reflectors. The ground-truth data were collected using Z + F IMAGER 5010 3D Terrestrial Laser Scanner. We show two scenarios: (1) Single measurement station without ground-truth data, (2) multiple measurement stations. It is documented by an experiment that the scenario of multiple measurement stations gives better qualitative and quantitative results compared with a single measurement station. However, the procedure requires greater effort in data acquisition. Moreover, this multi-view data registration scheme enables optimizing poses and parameters of planar reflectors simultaneously. It can be used for 3D map reconstruction. The calibration method is independent of mechanical design and does not require any fiducial markers on the mirrors. This work fulfills the gap between rotating and Solid-State LiDARs since the field of view can be reshaped by planar reflectors, and the proposed method can preserve the metric accuracy. There are many disadvantages of additional planar reflectors such as reducing the number of useful data and the negative impact on the sensor’s range. We observed a decreased range of the Livox Mid-40 LiDAR, even down to 80 per cent. The important finding is the degradation of LiDAR’s perception with a planar reflector. The degradation of sharp corners is evident. In contrast, an interesting fact is that the precision of the detection of the planar shape is rather similar; thus, an example planar target was detected with precision (1σ) 1.8 cm (without reflector) and 2.2 cm (with reflector). These quantitative measures confirm the nominal range precision reported by the Livox manufacturer. This work can be applied for multi LiDAR sensor calibration and other practical scenarios, e.g., 3D digitization of the urban environment, cultural heritage and archaeology, underground environment, environmental monitoring, forestry and agriculture assuming the usage of an affordable LiDAR with the reshaped field of view with planar reflectors. In future work, we will focus on localization aspects using the proposed 3D LiDAR prototype since it is an affordable, wide field of view and a long-range solution does not exist in the market and is not discussed in the literature. We prepared an open-source project and provided all the necessary data, including software, CAD design and sample captures for reproducing the experiments.

Abbreviations

The following abbreviations are used in this manuscript:

CAD Computer Aided Design
FDM Fused Deposition Modeling
GPS Global Positioning System
FOV Global Positioning System
ICP Iterative Closest Point
LiDAR Light Detection and Ranging
MEMS Micro Electro Mechanical Systems
PLA PolyLactic Acid
PMMA PolyMethyl MethAcrylate
SLAM Simultaneous Localization and Mapping
ROI Region of iterest

Author Contributions

Conceptualization, J.B. and M.P., methodology, J.B. and M.P.; software, M.P.; validation, J.B. and M.P.; formal analysis, J.B. and M.P.; investigation, J.B. and M.P.; resources, J.B. and M.P.; data curation, J.B. and M.P.; writing—original draft preparation, J.B. and M.P.; writing—review and editing, J.B. and M.P.; visualization, J.B. and M.P.; supervision, J.B.; project administration, J.B.; funding acquisition, J.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/michalpelka/catoptric_livox [32].

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Stefano F.D., Chiappini S., Gorreja A., Balestra M., Pierdicca R. Mobile 3D scan LiDAR: A literature review. Geomat. Nat. Hazards Risk. 2021;12:2387–2429. doi: 10.1080/19475705.2021.1964617. [DOI] [Google Scholar]
  • 2.Liu Z., Zhang F., Hong X. Low-cost Retina-like Robotic Lidars Based on Incommensurable Scanning. IEEE/ASME Trans. Mechatron. 2021 doi: 10.1109/TMECH.2021.3058173. [DOI] [Google Scholar]
  • 3.Lin Y., Hyyppä J., Kukko A. Stop-and-Go Mode: Sensor Manipulation as Essential as Sensor Development in Terrestrial Laser Scanning. Sensors. 2013;13:8140–8154. doi: 10.3390/s130708140. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Lehtola V.V., Nikoohemat S., Nüchter A. Handbook of Big Geospatial Data. Springer International Publishing; Cham, Switzerland: 2021. Indoor 3D: Overview on Scanning and Reconstruction Methods; pp. 55–97. [Google Scholar]
  • 5.Chen C.Y., Chien H.J., Huang P.S., Hong W.B., Chen C.F. Proceedings of the 27th Conference on Image and Vision Computing New Zealand, Dunedin, New Zealand, 26–28 November 2012. Association for Computing Machinery; New York, NY, USA: 2012. Intrinsic Parameters Calibration for Multi-Beam LiDAR Using the Levenberg–Marquardt Algorithm; pp. 19–24. [DOI] [Google Scholar]
  • 6.Bergelt R., Khan O., Hardt W. Improving the Intrinsic Calibration of a Velodyne LiDAR Sensor; Proceedings of the IEEE SENSORS; Glasgow, UK. 29 October–1 November 2017; [DOI] [Google Scholar]
  • 7.Levinson J., Thrun S. Unsupervised Calibration for Multi-beam Lasers. In: Khatib O., Kumar V., Sukhatme G., editors. Experimental Robotics: The 12th International Symposium on Experimental Robotics. Springer; Berlin/Heidelberg, Germany: 2014. pp. 179–193. [DOI] [Google Scholar]
  • 8.Wang D., Watkins C., Xie H. MEMS Mirrors for LiDAR: A Review. Micromachines. 2020;11:456. doi: 10.3390/mi11050456. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Zevering J., Bredenbeck A., Arzberger F., Borrmann D., Nüchter A. L.U.N.A.—A Laser-Mapping Unidirectional Navigation Actuator; Proceedings of the 17th International Symposium of Experimental Robotics (ISER ’21); La Valletta, Malta. 9–12 November 2021; pp. 85–94. [Google Scholar]
  • 10.Borrmann D., Jörissen S., Nüchter A. RADLER—A RADial LasER scanning device; Proceedings of the 16th International Symposium of Experimental Robotics (ISER ’18); Buenos Aires, Argentina. 5–8 November 2018; pp. 655–664. [Google Scholar]
  • 11.Endres F., Sprunk C., Kümmerle R., Burgard W. A catadioptric extension for RGB-D cameras; Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems; Chicago, IL, USA. 14–18 September 2014; pp. 466–471. [DOI] [Google Scholar]
  • 12.Kümmerle R., Grisetti G., Strasdat H., Konolige K., Burgard W. G2o: A general framework for graph optimization; Proceedings of the 2011 IEEE International Conference on Robotics and Automation; Shanghai, China. 9–13 May 2011; pp. 3607–3613. [DOI] [Google Scholar]
  • 13.Thrun S., Burgard W., Fox D. Probabilistic Robotics. MIT Press; Cambridge, MA, USA: 2005. [Google Scholar]
  • 14.Sünderhauf N., Protzel P. Switchable constraints for robust pose graph SLAM; Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems; Vilamoura-Algarve, Portugal. 7–12 October 2012; pp. 1879–1884. [DOI] [Google Scholar]
  • 15.Olson E., Agarwal P. Inference on networks of mixtures for robust robot mapping. Int. J. Robot. Res. 2013;32:826–840. doi: 10.1177/0278364913479413. [DOI] [Google Scholar]
  • 16.Agarwal P., Tipaldi G.D., Spinello L., Stachniss C., Burgard W. Robust map optimization using dynamic covariance scaling; Proceedings of the 2013 IEEE International Conference on Robotics and Automation; Karlsruhe, Germany. 6–10 May 2013; pp. 62–69. [Google Scholar]
  • 17.Akay A., Akgul Y.S. 3D reconstruction with mirrors and RGB-D cameras; Proceedings of the 2014 International Conference on Computer Vision Theory and Applications (VISAPP); Lisbon, Portugal. 5–8 January 2014; pp. 325–334. [Google Scholar]
  • 18.Morell-Gimenez V., Saval-Calvo M., Azorin-Lopez J., Garcia-Rodriguez J., Cazorla M., Orts-Escolano S., Fuster-Guillo A. A Comparative Study of Registration Methods for RGB-D Video of Static Scenes. Sensors. 2014;14:8547–8576. doi: 10.3390/s140508547. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Aalerud A., Dybedal J., Subedi D. Reshaping Field of View and Resolution with Segmented Reflectors: Bridging the Gap between Rotating and Solid-State LiDARs. Sensors. 2020;20:3388. doi: 10.3390/s20123388. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Dong H., Anderson S., Barfoot T.D. Two-axis scanning lidar geometric calibration using intensity imagery and distortion mapping; Proceedings of the 2013 IEEE International Conference on Robotics and Automation; Karlsruhe, Germany. 6–10 May 2013; pp. 3672–3678. [DOI] [Google Scholar]
  • 21.Abiko S., Sakamoto Y., Hasegawa T., Yuta S., Shimaji N. Development of constant altitude flight system using two dimensional laser range finder with mirrors; Proceedings of the 2017 IEEE International Conference on Advanced Intelligent Mechatronics (AIM); Munich, Germany. 3–7 July 2017; pp. 833–838. [DOI] [Google Scholar]
  • 22.Chen M., Pitzer B., Droz P.y., Grossman W. Mirrors to Extend Sensor Field of View in Self-Driving Vehicles. Application 20200341118. U.S. Patent. 2020 October 29;
  • 23.Im J.H., Im S.H., Jee G.I. Extended Line Map-Based Precise Vehicle Localization Using 3D LIDAR. Sensors. 2018;18:3179. doi: 10.3390/s18103179. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Badue C., Guidolini R., Carneiro R.V., Azevedo P., Cardoso V.B., Forechi A., Jesus L., Berriel R., Paixão T.M., Mutz F., et al. Self-driving cars: A survey. Expert Syst. Appl. 2021;165:113816. doi: 10.1016/j.eswa.2020.113816. [DOI] [Google Scholar]
  • 25.Leonard J., Durrant-Whyte H. Simultaneous map building and localization for an autonomous mobile robot; Proceedings of the IROS’91:IEEE/RSJ International Workshop on Intelligent Robots and Systems’91; Osaka, Japan. 3–5 November 1991; Nov 3–5, pp. 1442–1447. [Google Scholar]
  • 26.Skrzypczyński P. Simultaneous localization and mapping: A feature-based probabilistic approach. Int. J. Appl. Math. Comput. Sci. 2009;19:575–588. doi: 10.2478/v10006-009-0045-z. [DOI] [Google Scholar]
  • 27.Ayush K., Agarwal N.K. Real time visual SLAM using cloud computing; Proceedings of the 2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT); Tiruchengode, India. 4–6 July 2013; pp. 1–7. [DOI] [Google Scholar]
  • 28.Liu Y., Zhang H.H.C. A Novel RGB-D SLAM Algorithm Based on Cloud Robotics. Sensors. 2019;19:5288. doi: 10.3390/s19235288. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Livox-Mid-40. 2021. [(accessed on 23 September 2021)]. Available online: https://www.livoxtech.com/mid-40-and-mid-100/specs.
  • 30.z-f-imager-5010. 2021. [(accessed on 23 September 2021)]. Available online: https://geo-matching.com/terrestrial-laser-scanners/z-f-imager-5010.
  • 31.360, A.F. [(accessed on 23 September 2021)]. Available online: https://www.autodesk.com/products/fusion-360.
  • 32.Pelka M. Catoprtric Calibration Tool. 2021. [(accessed on 23 September 2021)]. Available online: https://github.com/michalpelka/catoptric_livox.
  • 33.Agarwal S., Mierle K. Ceres Solver. [(accessed on 23 September 2021)]. Available online: http://ceres-solver.org.
  • 34.Besl P., McKay N.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992;14:239–256. doi: 10.1109/34.121791. [DOI] [Google Scholar]
  • 35.Zhang J., Yao Y., Deng B. Fast and Robust Iterative Closet Point. IEEE Trans. Pattern Anal. Mach. Intell. 2021 doi: 10.1109/TPAMI.2021.3054619. [DOI] [PubMed] [Google Scholar]
  • 36.Segal A., Hähnel D., Thrun S. Generalized-ICP. In: Trinkle J., Matsuoka Y., Castellanos J.A., editors. Robotics: Science and Systems. The MIT Press; Cambridge, MA, USA: 2009. [Google Scholar]
  • 37.Solà J., Deray J., Atchuthan D. A micro Lie theory for state estimation in robotics. arXiv. 20201812.01537 [Google Scholar]
  • 38.Strasdat H., Lovegrove S. C++ implementation of Lie Groups Using Eigen. [(accessed on 23 September 2021)]. Available online: https://github.com/strasdat/Sophus.
  • 39.Rusu R.B., Cousins S. 3D is here: Point Cloud Library (PCL); Proceedings of the IEEE International Conference on Robotics and Automation (ICRA); Shanghai, China. 9–13 May 2011. [Google Scholar]
  • 40.Ryde J., Hu H. Mobile Robot 3D Perception and Mapping without Odometry Using Multi-Resolution Occupancy Lists; Proceedings of the 2007 International Conference on Mechatronics and Automation; Harbin, China. 5–8 August 2007; pp. 331–336. [Google Scholar]
  • 41.Hu C., Huang Z., Qin S., Wang X. A new 3D imaging lidar based on the high-speed 2D laser scanner. In: Shimura T., Xu G., Tao L., Zheng J., editors. Optoelectronic Imaging and Multimedia Technology II. Volume 8558. International Society for Optics and Photonics, SPIE; Bellingham, WA, USA: 2012. pp. 27–34. [Google Scholar]
  • 42.Alismail H., Browning B. Automatic Calibration of Spinning Actuated Lidar Internal Parameters. J. Field Robot. 2015;32:723–747. doi: 10.1002/rob.21543. [DOI] [Google Scholar]
  • 43.Bi S., Yuan C., Liu C., Cheng J., Wang W., Cai Y. A Survey of Low-Cost 3D Laser Scanning Technology. Appl. Sci. 2021;11:3938. doi: 10.3390/app11093938. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://github.com/michalpelka/catoptric_livox [32].


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES