Abstract
To optimize the structure and workflow of the 3D measurement robot system, reduce the dependence on specific calibration targets or high-precision calibration objects, and improve the versatility of the system’s self-calibration, this paper proposes a robot hand–eye calibration algorithm based on irregular targets. By collecting the 3D structural information of an object in space at different positions, a random sampling consistency evaluation based on the fast point feature histogram (FPFH) is adopted, and the iterative closest point (ICP) registration algorithm with the introduction of a probability model and covariance optimization is combined to iteratively solve the spatial relationship between point clouds, and the hand–eye calibration equation group is constructed through spatial relationship analysis to solve the camera’s hand–eye matrix. In the experiment, we use arbitrary objects as targets to execute the hand–eye calibration algorithm and verify the effectiveness of the method.
Keywords: hand–eye calibration, 3-dimention measurement, RANSAC, robotic
1. Introduction
Robot-based 3D measurement systems have been widely applied in quality inspection [1,2,3,4,5] and reverse engineering [6,7]. Among these systems, hand–eye calibration is an essential initialization step, where its accuracy significantly impacts the final measurement precision [8].
Current 3D camera calibration methods primarily rely on specific targets, including 2D patterns with distinctive features and specially designed 3D targets [9]. For 2D targets, notable methods include Tsai’s calibration approach based on radial alignment constraints using planar targets with known 3D structures [10], Zhang’s flexible camera calibration technique using planar patterns [11], as well as target designs based on Zhang’s method such as ARToolkit [12], ARTag [13], and AprilTag [14] calibration boards. Regarding 3D targets, designs include standard spherical targets [15] and those with specific geometric features [16,17,18,19].
Although target-based calibration methods demonstrate excellent accuracy and stability, their performance depends heavily on the manufacturing precision of the targets themselves, requiring specific designs and preparations while being sensitive to environmental variations. Consequently, researchers have recently explored target-less calibration approaches. Reference [20] proposed a method using arbitrary free-form surfaces, estimating hand–eye parameters through feature matching between measurement data and design models. Reference [21] introduced a scene-feature-based approach that computes camera intrinsics and hand–eye relationships by extracting ORB features from consecutive scene images. However, this method depends on the features contained in depth images and imposes strict requirements on capture range and motion sequence continuity.
To enhance the robustness of robot hand–eye calibration and reduce dependence on high-precision targets and controlled environments, this paper presents a novel calibration method using unfixed 3D objects. Our approach employs arbitrary objects of appropriate size as 3D targets, utilizing a two-step improved ICP algorithm with RANSAC-based feature recognition. By performing local registration of the target from different viewpoints, we establish relative relationships between target positions in various camera frames. Through the comprehensive modeling of the hand–eye system and the introduction of a virtual camera coordinate origin, we transform target relationships into inter-camera transformations, constructing the parameters required by traditional calibration methods to solve for high-accuracy hand–eye matrices.
This research implements robot hand–eye calibration by establishing coordinate transformations of non-fixed 3D targets through registration based on RANSAC feature recognition and improved ICP algorithms. Section 2.3 introduces the calibration methodology and core algorithms. Section 2.4 introduces the multi-step registration algorithm. Section 3 presents experiments using arbitrary targets, accuracy validation tests, and detailed error analysis. Section 4 summarizes the paper, discusses the limitations of the current study, and outlines potential future research directions.
2. Methods
2.1. Hand–Eye Robot System
Based on the position of the camera used to acquire calibration information—whether it is mounted on the robot’s end-effector or external to the robot—robot hand–eye systems can be classified into two categories as follows: Eye-In-Hand and Eye-to-Hand. The system designed in this paper includes both a structured light camera mounted on the robot’s end-effector to capture target information and a binocular tracking camera to obtain the pose of the robot’s end-effector. Therefore, this system does not fall into either of these two typical categories. As shown in Figure 1, the hand–eye system used in this paper consists of the following components:
Fixed binocular tracking camera (W): Used to track the robot’s end-effector;
Tracking target (T): Mounted on the robot’s end-effector, tracked by the binocular camera;
Structured light 3D scanner (C): Mounted on the robot’s end-effector, used to capture 3D information of the target;
Arbitrary 3D target (B): Used as the calibration object.
Figure 1.
Measurement system.
This system integrates both internal and external sensing capabilities, enabling robust and flexible calibration without relying on a fixed target or a single camera configuration.
In this system, the coordinate systems are defined as follows: : The fixed arbitrary 3D target coordinate system. : The tracking target coordinate system attached to the robot’s end-effector. : The structured light 3D camera coordinate system. : The fixed world coordinate system of the tracking camera.
The transformation matrices are defined as follows: : The homogeneous transformation matrix from the tracking target coordinate system to the world coordinate system , obtained by the tracking camera. : The homogeneous transformation matrix from the tracking target coordinate system to the structured light camera coordinate system that needs to be calibrated. : The homogeneous transformation matrix from the arbitrary 3D target coordinate system to the structured light camera coordinate system .
When the measurement system is operational, a point measured in the structured light camera coordinate system is transformed into a point in the world coordinate system using the following transformation:
| (1) |
Here, represents the pose of the rigid body composed of the structured light camera and its attached tracking target in the world coordinate system at the time of measurement. By unifying the measurement results from the structured light camera at different positions and orientations, the final complete measurement result is obtained.
This approach ensures that the measurements from the structured light camera are accurately mapped to the world coordinate system, enabling precise and consistent calibration and reconstruction.
2.2. Hand–Eye Calibration Algorithm
As shown in Figure 2, the raw data for the calibration algorithm proposed in this paper includes measurement data of non-specific 3D targets and the corresponding tracking data. The computation of the hand–eye matrix is completed in three steps as follows: registration, equation construction, and solution.
Figure 2.
Calibration algorithm flow.
2.3. Construction of Calibration Equation
As shown in Figure 3, the robot is used to drive the structured light camera to capture the point set of the arbitrary 3D target in the structured light camera coordinate system . Simultaneously, the tracking camera obtains the pose of the tracking target t in the world coordinate system for each acquisition.
Figure 3.
Multi-pose calibration measurement.
At positions i and j, the homogeneous transformation matrices between the world coordinate system and the structured light camera coordinate system are denoted as and , respectively. These matrices satisfy the following relationships:
| (2) |
The homogeneous transformation matrix between the structured light camera coordinate systems at the two positions, denoted as , satisfies the following:
| (3) |
By combining Equations (2) and (3), we obtain the following:
| (4) |
Here, represents the registration result of the point cloud of the stereo calibration target at the two positions, and it satisfies the following:
| (5) |
Substituting Equation (5) into Equation (4), we obtain the following:
| (6) |
| (7) |
In Equation (7), let , , and . For a pair of 3D calibration target measurements and their corresponding tracked target poses, the structure of the robot hand–eye matrix can be formulated as the equation .
As shown in Figure 4, this process is extended to perform 3D scanning of the stereo calibration target at multiple poses, obtaining multiple sets of point cloud data in the structured light camera coordinate system, as well as the corresponding tracked target poses . Sequentially, the absolute poses of the tracked target in the world coordinate system, and , are used to obtain pairwise relative pose homogeneous matrices . Sequentially, adjacent point clouds are pairwise calibrated, resulting in homogeneous matrices from n sets of data. After obtaining multiple matrix equations, quaternion decomposition and the least squares method are used to solve the system of matrix equations.
Figure 4.
Multi-pose measurement transformation matrix relationship.
In addition to the direct derivation of the transformation matrix equation system mentioned earlier, this calibration process can also utilize existing data and the spatial relationships contained within to convert parameters into a homogeneous transformation matrix of the 3D calibration target in the structured light camera coordinate system, similar to the calibration method described previously. For any point on the 3D calibration target measured by the structured light camera at positions i and j, the transformation relationship between the point clouds after registration satisfies the following:
| (8) |
In the world coordinate system, any point p on the 3D calibration target satisfies the following in the structured light camera coordinate system:
| (9) |
Combining these equations, we obtain the following:
| (10) |
Thus, the following holds:
| (11) |
Here, represents the poses of the camera in the world coordinate system. The transformation of the structured light camera’s pose in space is the inverse of the point cloud transformation, expressed as follows:
| (12) |
After indirectly obtaining the position relationship of the camera in the world coordinate system through point cloud registration, variables can be constructed to satisfy the data input requirements of traditional hand–eye calibration, and the corresponding algorithm can be invoked. The inputs are the homogeneous matrix from the robot flange to the world coordinate system and the homogeneous matrix from the target to the camera. The output is the hand–eye matrix.
In the system constructed in this work, the homogeneous matrix of the tracked target in the world coordinate system of the tracking camera is equivalent to the homogeneous matrix from the robot flange to the world coordinate system. The virtual camera coordinate system is defined as the coordinate system of the first stereo calibration target, and its matrix is the identity matrix I. For the k-th stereo calibration target, the homogeneous matrix relative to the virtual camera coordinate system is the registration result of the stereo calibration target point cloud, which can also be expressed as the product of the sequential registration results as follows:
| (13) |
By constructing the virtual camera coordinate system using the relative relationships of the stereo calibration targets obtained through registration, the transformation of homogeneous matrices can be achieved, making it compatible with existing hand–eye matrix calculation functions. This enables the implementation of the robot hand–eye calibration calculation based on arbitrarily shaped 3D calibration targets as designed in this work.
2.4. Multi-Step Registration Algorithm
The hand–eye calibration method proposed in this paper that is based on arbitrary 3D targets relies on the scanning and calibration of the target’s point cloud. It requires precise calibration calculations for point clouds with significant initial pose differences and partially non-overlapping regions to accurately determine the spatial relationship of the target’s measured area. This, in turn, enables the determination of the spatial relationship of the camera.
As shown in Figure 5, to meet the requirements of the calibration algorithm, this paper designs a registration method that includes the following two steps: coarse registration based on local geometric feature calculation and matching, and fine registration based on an improved ICP algorithm, achieving the preliminary processing of the collected data.
Figure 5.
Multi-step registration algorithm.
The RANSAC (Random Sample Consensus) algorithm is used to calculate the local feature distribution of point clouds based on a FPFH (Fast Point Feature Histogram) [22], enabling the registration of the source point cloud with the target point cloud. First, the voxel size for voxel filtering is set as the dimension for voxel downsampling, along with the radius for normal vector estimation and the search radius for FPFH feature extraction. For the downsampled voxel grid point cloud, the voxel size is used as the basic unit for the KD-tree search radius, and the nearest neighbors within each point’s neighborhood are searched to estimate the normal vectors.
The estimated normal vectors of each point are used to compute the SPFH (Simplified Point Feature Histogram) for that point. The voxel size from the voxel downsampling is set as the basic unit for the neighborhood point search radius. For each point, it forms a point pair and their estimated normal vectors , a coordinate system is constructed, satisfying:For a point pair and their estimated normal vectors , and a coordinate system is constructed, satisfying the following:
| (14) |
In the coordinate system, the pairwise features , , and for the point pair are calculated as follows:
| (15) |
where d represents the Euclidean distance between the point pair as follows:
| (16) |
Through the above calculations, for each point and its neighboring points , the n point pairs form the SPFH for each point, which can also be expressed as follows:
| (17) |
where , , and are histograms based on the values of , , and .
For each point, the SPFH is weighted and summed, incorporating the information from neighboring points into the final feature parameters of the point. For point , its consists of the following two parts: the first part is its own , and the second part is the weighted average of the k neighboring points using the inverse distance as the weight. Thus, the can be expressed as follows:
| (18) |
Through the above calculations, the FPFH of the source point cloud and the target point cloud are obtained. Randomly select N feature point pairs from the source and target point clouds where the FPFH matches. First, we use SVD (Singular Value Decomposition) to estimate the rotation part of the rigid transformation matrix for these point pairs. Centroid normalization is performed on these points as follows:
| (19) |
For the centroid-normalized point sets and , the covariance matrix H is calculated as follows:
| (20) |
where ⨂ denotes the outer product.
The covariance matrix H is decomposed using SVD as follows:
| (21) |
From this, the rotation transformation R for the point sets and is obtained as follows:
| (22) |
Based on the rotation transformation R, the origin displacement of the point sets and is calculated to obtain the translation transformation t as follows:
| (23) |
Combining R and t yields the homogeneous transformation matrix T, as follows:
| (24) |
Based on the initially calculated homogeneous transformation matrix T, the distance difference between the feature-matched corresponding points is computed. If the difference is below the threshold, the global feature registration is completed. If the difference exceeds the threshold, the above steps are repeated, and the final homogeneous transformation matrix T is obtained by cumulative multiplication and combination.
For the point cloud after coarse registration based on RANSAC and FPFH features, the optimized Iterative Closest Point (ICP) algorithm [23] is executed to further improve registration accuracy, achieving results that meet the precision requirements of the calibration process.
Using the registration result from RANSAC as the initial transformation, for each point in the source point cloud and in the target point cloud, the Mahalanobis distance is calculated as the error metric as follows:
| (25) |
where and are the covariance matrices of the source and target points within their neighborhoods. For any point in the point cloud, its covariance is the outer product of the coordinate differences between and its neighboring points , expressed as follows:
| (26) |
where n is the neighborhood size.
For the registration step, the target transformation matrix T is as follows:
| (27) |
The Mahalanobis distance-based error between the target point set and the source point set is expressed as follows:
| (28) |
Expanding this gives the following:
| (29) |
The Gauss–Newton method is used to perform a first-order Taylor expansion of the Mahalanobis distance error for iterative optimization. First, the transformation matrix T is parameterized as a vector x as follows:
| (30) |
where is the Lie algebra representation of .
For each point pair , the corresponding residual is as follows:
| (31) |
The weighted sum of squared residuals is expressed as follows:
| (32) |
The residual is Taylor-expanded to approximate the nonlinear error as follows:
| (33) |
where is the parameter increment, and is the Jacobian matrix of the error with respect to parameters x. The rotation derivative and translation derivative are expressed as follows:
| (34) |
| (35) |
where denotes the cross-product matrix of the vector. Through Taylor expansion, the objective function can be approximated as a least squares problem as follows:
| (36) |
| (37) |
Ignoring higher-order terms, it is approximated as follows:
| (38) |
To minimize E, the derivative with respect to is taken and set to zero, yielding the Gauss–Newton equation, expressed as follows:
| (39) |
where is the Gauss–Newton Hessian matrix, and is the gradient vector.
After obtaining the linear system of equations for , an approximate solution is computed iteratively until is below the threshold . The cumulative corresponds to , yielding the final transformation matrix T.
3. Experiment
3.1. Experimental Platform and Evaluation Criteria
3.1.1. Experimental Platform
We build the experimental hardware system as shown in Figure 6. The experiment uses the JAKA Zu7 collaborative robot to drive the structured light camera and the tracking target, with a repeatability of mm. The PhoXi 3D Scanner S structured light camera is used for 3D data scanning, with a scanning range of 384–520 mm, point spacing of 0.174 mm, and calibration accuracy of 0.050 mm. The AITS D-Series Smart tracking camera is used to track the target on the robotic arm to obtain the end-effector pose, with a tracking distance of 0.5–2.0 m and tracking accuracy of 0.15–0.30 mm. Arbitrarily selected objects are used as the calibration targets for hand–eye calibration experiments. After completing system calibration, a ceramic standard sphere with a diameter of mm is used for calibration experiments. The experiments include 3D reconstruction error experiments and distance measurement error experiments.
Figure 6.
Experimental hardware system. (a) Experimental system, (b) six-axis robotic arm, (c) structured light camera, (d) ceramic standard ball, and (e) tracking camera.
3.1.2. Evaluation Criteria
The 3D reconstruction error refers to controlling the robot to measure the point cloud of a standard sphere placed at a fixed position from different locations. The sphere center position in the structured light camera coordinate system is obtained through fitting. Using the hand–eye matrix from the previous calibration experiment and the transformation matrix of the structured light camera in the world coordinate system obtained from the tracking camera, the sphere center positions scanned from different locations are unified into the world coordinate system of the tracking camera. The position error of the fixed standard sphere center unified into the world coordinate system under different structured light camera positions is the metric for evaluating the 3D reconstruction error of the robot measurement system.
The distance measurement error refers to controlling the robot system to remain stationary and using a slide rail with a movement accuracy of 0.1 mm to measure the point cloud of the standard sphere before and after movement. After sphere center identification and coordinate system transformation, the error between the distance of the sphere center positions in the world coordinate system and the actual measured distance is the distance measurement error.
3.2. Experimental Procedures
According to the calibration process described in the methodology, the proposed calibration method requires experimental data consisting of a set of 3D point cloud scans of an object and a corresponding set of tracking poses. As shown in Figure 7, the robot-controlled scanning camera is used to scan a toolbox as the 3D target at different positions, resulting in 16 sets of corresponding point clouds and tracking poses.
Figure 7.
Calibration data collection.
According to the registration method proposed in this study, the number of iterations in the initial stage of global registration significantly influences the registration accuracy. As illustrated in Figure 8, during the registration of the first two sets of target point clouds, the impact of the number of RANSAC iterations on the RMSE is demonstrated. It can be observed from Figure 8 that a relatively small number of iterations is sufficient for achieving stable registration performance, with the RMSE reaching its optimal and stable value after approximately 8 iterations.
Figure 8.
Influence of RANSAC iteration count on the RMSE of global registration.
As shown in Figure 9, following the registration method described in Section 2.2, 15 sets of registrations are performed on the 16 sets of point clouds, resulting in 15 transformation matrices that represent the relative poses of point clouds 2–16 with respect to point cloud 1.
Figure 9.
Target pose registration.
Because the calibration method proposed in this paper involves the registration of the same object under significant pose variations, the standard ICP registration algorithm is prone to becoming stuck in local optima and failing to achieve correct registration due to large initial pose differences and small overlapping areas. However, since this paper introduces a global registration algorithm and a covariance-optimized ICP algorithm, it can effectively achieve point cloud registration in such scenarios. Figure 10 shows the registration of the first four pairs of point clouds in this experiment using the standard ICP method, and most of the point clouds failed to achieve correct registration.
Figure 10.
Pointcloud registration by standard ICP (First Four Sets).
To compare the computation time of the standard ICP method and the proposed registration method under this dataset, the experiment was conducted on a laptop equipped with an AMD Ryzen 7 4800U processor. This processor is based on the Zen 2 architecture, featuring eight physical cores and 16 threads, a base clock frequency of 1.8 GHz, a maximum boost clock frequency of up to 4.2 GHz, and 8 MB of L3 cache. As shown in Figure 11, due to the introduction of global registration and covariance optimization, the proposed method can consistently maintain the computation time for each step at approximately 4 s. In contrast, standard ICP is prone to becoming trapped in local optima, which increases computation time. The proposed registration method reduces computation time by 81.87.
Figure 11.
Comparison of registration time.
The spatial poses contained in this set of data are illustrated in Figure 12. Simultaneously, during each scan, the tracking camera obtains the transformation matrices of the tracking target on the structured light camera in the world coordinate system.
Figure 12.
Spatial distribution of pose data.
The measured values are input into the hand–eye calibration algorithm proposed in Section 2, yielding the transformation matrix of the tracking target relative to the structured light camera coordinate system, i.e., the hand–eye matrix , with the following value:
3.3. Three-Dimensional Reconstruction Error
This paper uses the three-standard-sphere calibration method as the control group. The robot hand–eye matrix is obtained by following the calibration process proposed in Section 2.2, using 16 sets of coordinate systems composed of three standard spheres. Figure 13 shows the visualization results of the 3D reconstruction error validation for the ceramic standard sphere using the calibration results obtained in the previous chapter. The mean error of the 16 sets of sphere center coordinates is 1.53 mm, with a standard deviation of 0.42. In the control experiment under the same hardware system, as shown in Figure 14, the 3D reconstruction results of the ceramic standard sphere using the proposed method are compared with those obtained using the three-sphere-center coordinate system method. The mean error of the 16 sets of sphere center coordinates for the latter is 3.79 mm, with a standard deviation of 0.99. The experimental results demonstrate that the proposed method, compared to the three-point coordinate system method, achieves a 59.62% reduction in mean error and a 57.54% reduction in standard deviation, while requiring lower precision for the calibration object and simpler execution steps.
Figure 13.
Three-Dimensional reconstruction error.
Figure 14.
Three-Dimensional reconstruction error distribution.
A comparison of the 3D reconstruction position errors between the method proposed in this paper, the three-standard-sphere coordinate system method, and other research methods is shown in Table 1.
Table 1.
Mean and distribution of 3D reconstruction errors for different methods.
| Method | Camera Type | Target Type | Mean Error (mm) | Standard Deviation |
|---|---|---|---|---|
| Three-Ball System Method | Structured Light | Three Standard Spheres | (X: 2.77, Y: 1.49, Z: 1.47) 3.79 | 0.99 |
| Zhe’s Method [24] | Line Laser | 2D Calibration Board | (X: 1.334, Y: 0.511, Z: 0.925) 1.855 | - |
| Murali’s Method [25] | Laser Profiler | Arbitrary 3D Object | (X: 0.701, Y: 0.443, Z: 0.366) 0.906 | - |
| Peter’s Method [26] | Structured Light | Arbitrary 3D Object | 1.77 | - |
| Proposed Method | Structured Light | Arbitrary 3D Object | (X: 1.10, Y: 0.60, Z: 0.66) 1.53 | 0.42 |
The 3D reconstruction data demonstrates that the proposed method in this paper achieves the smallest mean error in the calibration of structured light cameras for arbitrary 3D objects. Compared to the experimental results obtained using laser scanner devices, the error is significantly larger. While the evaluation metric used in this paper is a 3D reconstruction error, recent learning-based methods such as EasyHeC [27] and EasyHeC++ [28] report their performance in terms of translation error during hand–eye calibration. According to the original publications, EasyHeC(2023) achieves a mean translation error of 2.06 mm, and EasyHeC++(2024) improves this to 1.35 mm, both under specific experimental setups involving fiducial targets and high-performance industrial cameras. In comparison, the 3D reconstruction error reported in this paper is 1.53 mm, which inherently incorporates the calibration error as one of its components. This indicates that the proposed method achieves an overall performance comparable to or better than EasyHeC and EasyHeC++ in practical applications, despite using low-cost arbitrary 3D targets and a more accessible experimental setup. The results highlight the flexibility and applicability of the proposed method in scenarios where specialized hardware and fiducial targets are unavailable.
3.4. Distance Measurement Error
As shown in Figure 15, for a set of linear sliders, the point clouds of the spherical surface obtained at positions of −10.0 mm, 0.0 mm, and 10.0 mm have sphere center distances of 10.47 mm and 10.31 mm, respectively, with an average distance measurement error of 0.39 mm. The sources of error mainly include the calibration error of the structured light camera (0.05 mm), the tracking error of the tracking camera (0.15 mm), potential positional errors in the sphere center fitting calculation, and cumulative errors that may arise during the computational process.
Figure 15.
Distance measurement error.
4. Conclusions
In robotic measurement systems, hand–eye calibration is an essential task. For the calibration of 3D cameras, it is often necessary to perform complex intrinsic camera calibration, followed by calibration using a calibration board with specific patterns or relying on calibration objects with high machining precision, such as standard spheres. To simplify the calibration process, this paper proposes a hand–eye calibration method based on a highly adaptable and high-precision registration approach, using arbitrary 3D targets. The proposed method employs arbitrary spatial objects as 3D targets, obtains the relative positions of each captured data through multi-step registration, and indirectly derives the spatial relationships between cameras, thereby establishing a system of linear equations for the hand–eye relationship. By solving these equations, the hand–eye relationship of the camera system is obtained. Experimental results show that, compared to conventional calibration methods based on standard spheres for establishing spatial coordinate systems, the proposed method achieves higher precision and greater spatial coverage in 3D reconstruction error experiments, as well as higher accuracy in distance measurement error experiments.
However, the proposed method still faces some robustness and precision issues, primarily due to the reliance of the registration step on the high precision and resolution of the 3D camera itself, the dependence on tracking accuracy during data collection, and the residuals generated by the hand–eye relationship solving algorithm when the spatial relationships covered by the raw data are limited. Developing registration algorithms with higher overall feature recognition capabilities or improving the performance of measurement hardware in the system could help enhance the accuracy of the calibration results and the overall precision of the measurement system. In this paper, the experimental data acquisition was performed using a manual robot control method, which is relatively labor-intensive and time-consuming. To improve the automation of this process, further research can be conducted on the arrangement of acquisition points and the automatic planning of calibration data collection. The 3D target required by the proposed method needs to satisfy non-centrosymmetric geometric features. Furthermore, due to the precision limitations imposed by the 3D camera’s resolution, using excessively small 3D targets may reduce the calibration accuracy. Developing registration algorithms with stronger spatial feature extraction capabilities could further enhance the applicability of this method.
Acknowledgments
The authors thank HongPeng Li and ZeHao Ding for helping with solution design, conceptualization and data curation.
Author Contributions
Conceptualization, Z.S., L.Q. and Y.S.; methodology, Z.S.; algorithm and experiments, Z.S. and C.S.; original draft preparation, Z.S.; review and editing, Z.S., L.Q. and Y.S.; funding acquisition, L.Q. and Y.S. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The data presented in this study are available on request from the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
Funding Statement
This work was funded by the key scientific and technological project of HBIS Group (No. HG2023243).
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Liu S., Wang Q., Luo Y. A review of applications of visual inspection technology based on image processing in the railway industry. Transp. Saf. Environ. 2019;1:185–204. doi: 10.1093/tse/tdz007. [DOI] [Google Scholar]
- 2.Li W.L., Xie H., Zhang G., Yan S.-J., Yin Z.-P. Hand-eye calibration in visually-guided robot grinding. IEEE Trans. Cybern. 2016;46:2634–2642. doi: 10.1109/TCYB.2015.2483740. [DOI] [PubMed] [Google Scholar]
- 3.Xu J., Chen R., Chen H., Zhang S., Chen K. Fast registration methodology for fastener assembly of large-scale structure. IEEE Trans. Ind. Electron. 2017;64:717–726. doi: 10.1109/TIE.2016.2599140. [DOI] [Google Scholar]
- 4.Li T., Gao L., Li P., Pan Q. An ensemble fruit fly optimization algorithm for solving range image registration to improve quality inspection of free-form surface parts. Inf. Sci. 2016;367:953–974. doi: 10.1016/j.ins.2016.07.030. [DOI] [Google Scholar]
- 5.Lin H.W., Tai C.L., Wang G.J. A mesh reconstruction algorithm driven by an intrinsic property of a point cloud. Comput. Aided Des. 2004;36:1–9. doi: 10.1016/S0010-4485(03)00064-2. [DOI] [Google Scholar]
- 6.Chang W., Zwicker M. Global registration of dynamic range scans for articulated model reconstruction. ACM Trans. Graph. 2011;30:26. doi: 10.1145/1966394.1966405. [DOI] [Google Scholar]
- 7.Jiang J., Luo X., Luo Q., Li M. An overview of hand-eye calibration. Int. J. Adv. Manuf. Technol. 2022;119:77–97. doi: 10.1007/s00170-021-08233-6. [DOI] [Google Scholar]
- 8.Salvi J., Armangué X., Batlle J. A comparative review of camera calibrating methods with accuracy evaluation. Pattern Recognit. 2002;35:1617–1635. doi: 10.1016/S0031-3203(01)00126-1. [DOI] [Google Scholar]
- 9.Shiu Y.C., Ahmad S. Calibration of wrist-mounted robotic sensors by solving homogeneous transform equations of the form AX=XB. IEEE Trans. Robot. Autom. 1989;5:16–29. doi: 10.1109/70.88014. [DOI] [Google Scholar]
- 10.Tsai R.Y., Lenz R.K. A new technique for fully autonomous and efficient 3D robotics hand/eye calibration. IEEE Trans. Robot. Autom. 1989;5:345–358. doi: 10.1109/70.34770. [DOI] [Google Scholar]
- 11.Zhang Z.Y. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000;22:1330–1334. doi: 10.1109/34.888718. [DOI] [Google Scholar]
- 12.Fiala M. Comparing ARTag and ARToolkit plus Fiducial Marker Systems; Proceedings of the IEEE International Workshop on Haptic Audio Visual Environments and Their Applications; Ottawa, ON, Canada. 1 October 2005; pp. 148–153. [DOI] [Google Scholar]
- 13.Fiala M. ARTag, A fiducial marker system using digital techniques. IEEE Conf. Comput. Vis. Pattern Recognit. 2005;2:590–596. doi: 10.1109/cvpr.2005.74. [DOI] [Google Scholar]
- 14.Wang J., Olson E. AprilTag 2: Efficient and Robust Fiducial Detection; Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Daejeon, Republic of Korea. 9–14 October 2016; pp. 4193–4198. [DOI] [Google Scholar]
- 15.Xie Z.X., Zong P.F., Yao P., Ren P. Calibration of 6-DOF industrial robots based on line structured light. Optik. 2019;183:1166–1178. doi: 10.1016/j.ijleo.2019.02.069. [DOI] [Google Scholar]
- 16.Cajal C. A crenellated-target-based calibration method for laser triangulation sensors integration in articulated measurement arms. Robot. Comput. Integr. Manuf. 2011;27:282–291. doi: 10.1016/j.rcim.2010.07.008. [DOI] [Google Scholar]
- 17.Wagner M., Reitelshofer S., Franke J. Self-Calibration Method for a Robot Based 3D Scanning System; Proceedings of the 2015 IEEE 20th Conference on Emerging Technologies & Factory Automation (ETFA 2015); Luxembourg, Luxembourg. 8–11 September 2015; [DOI] [Google Scholar]
- 18.Lembono T.S., Suarez-Ruiz F., Pham Q.C. SCALAR: Simultaneous calibration of 2-D laser and robot kinematic parameters using planarity and distance constraints. IEEE Trans. Autom. Sci. Eng. 2019;16:1971–1979. doi: 10.1109/TASE.2019.2918141. [DOI] [Google Scholar]
- 19.Shibin Y., Yongjie R., Shenghua Y. A vision-based self-calibration method for robotic visual inspection systems. Sensors. 2013;13:16565–16582. doi: 10.3390/s131216565. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Xie H., Li W., Liu H. General geometry calibration using arbitrary free-form surface in a vision-based robot system. IEEE Trans. Ind. Electron. 2021;69:5994–6003. doi: 10.1109/TIE.2021.3090716. [DOI] [Google Scholar]
- 21.Xu G., Yan Y. A scene feature based eye-in-hand calibration method for industrial robot. Mechatron. Mach. Vis. Pract. 2021;4:179–191. doi: 10.1007/978-3-030-43703-9_15. [DOI] [Google Scholar]
- 22.Rusu R.B., Blodow N., Marton Z.C., Beetz M. Aligning Point Cloud Views Using Persistent Feature Histograms; Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems; Nice, France. 22–26 September 2008; pp. 3384–3391. [DOI] [Google Scholar]
- 23.Segal A.V., Haehnel D., Thrun S. Generalized-ICP; Proceedings of the Robotics: Science and Systems Conference; Seattle, WA, USA. 28 June–1 July 2009; pp. 347–354. [DOI] [Google Scholar]
- 24.Wang Z., Fan J., Jing F., Deng S., Zheng M., Tan M. An efficient calibration method of line structured light vision sensor in robotic eye-in-hand system. IEEE Sens. J. 2020;20:6200–6208. doi: 10.1109/JSEN.2020.2975538. [DOI] [Google Scholar]
- 25.Murali P.K., Sorrentino I., Rendiniello A., Fantacci C., Villagrossi E., Polo A. In Situ Translational Hand-Eye Calibration of Laser Profile Sensors Using Arbitrary Objects; Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA 2021); Xi’an, China. 30 May–5 June 2021; pp. 11067–11073. [DOI] [Google Scholar]
- 26.Peters A., Knoll A.C. Robot self-calibration using actuated 3D sensors. J. Field Robot. 2024;41:327–346. doi: 10.1002/rob.22259. [DOI] [Google Scholar]
- 27.Chen L., Qin Y., Zhou X., Su H. EasyHeC: Accurate and Automatic Hand-eye Calibration via Differentiable Rendering and Space Exploration. IEEE Robot. Autom. Lett. 2023;8:7234–7241. doi: 10.1109/LRA.2023.3315551. [DOI] [Google Scholar]
- 28.Hong Z., Zheng K., Chen L. EasyHeC++: Fully Automatic Hand-Eye Calibration with Pretrained Image Models; Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Detroit, MI, USA. 14–18 October 2024; pp. 1234–1240. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The data presented in this study are available on request from the corresponding author.















