Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2024 Feb 12.
Published in final edited form as: ROMAN. 2023 Nov 13;2023:2359–2365. doi: 10.1109/RO-MAN57019.2023.10309472

Human-Robot Interaction in Retinal Surgery: A Comparative Study of Serial and Parallel Cooperative Robots

Botao Zhao 1, Mojtaba Esfandiari 1, David E Usevitch 1, Peter Gehlbach 2, Iulian Iordachita 1
PMCID: PMC10861175  NIHMSID: NIHMS1923110  PMID: 38347956

Abstract

Cooperative robots for intraocular surgery allow surgeons to perform vitreoretinal surgery with high precision and stability. Several robot structural designs have shown capabilities to perform these surgeries. This research investigates the comparative performance of a serial and parallel cooperative-controlled robot in completing a retinal vessel-following task, with a focus on human-robot interaction performance and user experience. Our results indicate that despite differences in robot structure and interaction forces and torques, the two robots exhibited similar levels of performance in terms of general robot-to-patient interaction and average operating time. These findings have implications for the development and implementation of surgical robotics, suggesting that both serial and parallel cooperative-controlled robots can be effective for vitreoretinal surgery tasks.

I. INTRODUCTION

Vitreoretinal surgery (VRS) is among the most delicate and demanding eye surgery procedures in medicine. VRS deals with disorders of the vitreous and retina. The vitreous is a clear gel-like substance that fills the space between the lens and retina providing nutrients and structure to the eye. The retina is a thin layer of tissue lining the back of the eye which houses photoreceptor cells to receive photons and send electrical and chemical signals to the brain for visual perception [1]. VRS is used to treat a wide variety of conditions including retinal detachments, macular holes, epiretinal membranes, diabetic retinopathy, and age-related macular degeneration. Given the challenging nature of these procedures, various assistive eye surgical robots have been developed in recent years. These robots can be broadly categorized into three major classes: handheld robots, teleoperated robots, and cooperative-controlled robots [2].

Handheld VRS robots are miniaturized robotic instruments directly held by eye surgeons without external support to the robot. In the typical case, an active tool is attached to the robot as its end-effector. As general tool manipulation is controlled directly by the surgeon, these robots are very familiar and intuitive and require less effort to master. The Micron system of Carnegie Mellon University [3] is among the best known examples of a handheld robot designed for VRS. Teleoperated VRS robots are pairs of controller-responder interfaces, where the responder is remotely manipulated by a surgeon using a joystick or other user interface to move the controller mechanism. Benefits of this approach include motion scaling control and the elimination of hand tremor during surgery, of which both features are particularly beneficial to eye surgery tasks. The SMOS [4], Preceyes [5], and RAMS [6] robots are all notable examples of VRS teleoperated robots. Cooperative-control (CC) robots (“co-bots”) are designed to allow control by both the surgeon and robot. The surgeon holds a surgical instrument supported by the robot, connected via a force-torque sensor. As the surgeon moves the tool, the robot moves in real time according to assigned corresponding compliant motions mapped to forces and torques detected. These motions are dictated by specified robot kinematics and selected control algorithms. Co-bots can reduce or eliminate hand tremor while still maintaining the direct tool manipulation experience familiar to surgeons. Well-known examples of cooperative-control robots for VRS include the robotic systems of King’s College London and Moorfields Eye Hospital [7], KU Leuven [8] and the Steady-Hand Eye Robot (SHER) developed by Johns Hopkins University [9].

VRS robotic systems have implemented various effective robotic designs to achieve precise motions for surgery tasks including the use of serial and parallel designs. Previous research has been devoted to the design, analysis, and comparison of robot kinematics, dynamics, and control systems for both serial and parallel robotic systems [10],[11] though comparison of these two types is not a trivial matter. In general, serial robots demonstrate excellent repeatability and can be programmed with relative ease to perform a broad range of surgical procedures, but their larger size compared to parallel robots may pose challenges in certain surgical settings. Parallel robots are generally capable of faster dynamic motions, improved rigidity, and their ability to be restricted to a compact size makes them well-suited for use in tight surgical spaces. Parallel robot design and implementation, however, involves more complex mathematics, programming, and operation compared to serial robots.

Despite the extensive research in robot design and control outlying benefits and drawbacks between these two robot types, there is a lack of comparative studies in using them for medical surgery tasks. As comparing these robot types is quite complex with many factors coming into play (e.g., intricacies of kinematics, dynamics, Jacobian computations, and a slew of others), simplifying comparisons to focus primarily on human-robot interaction could prove valuable, as this component is the most essential when considering CC robots. Improved understanding of this interaction may show benefits for using a certain robot design for cooperative surgical tasks over the other. Hence a comparative study based on human-robot interaction could potentially reveal significance in performances between styles influencing further design trends for CC VRS robots of the future.

This work provides a comparative study between two kine-matically different SHER robots (SHER-2.1 and SHER-3.0), with serial and parallel robot designs, to better understand their differences for VRS use. As the serial-based SHER-2.1 has previously been verified as a reliable and effective robot for VRS [9],[13], it is used as the base reference for comparison with the SHER-3.0, the parallel robot, in this study. For human-to-robot interaction comparison, the force and torque of the user’s hand-to-tool interaction and the tool-to-eye phantom interaction force of each robot are measured and recorded. Learning curves and the cumulative summation (CUSUM) technique [14] are used to assess user proficiency and compared between manipulations of each robot.

This work contributes insight into the comparison of serial and parallel robots for human-to-robot interaction using CC VRS robots. We present a unique vessel following task which can be used to compare CC robot performance, and analysis methods for the comparison of human-to-robot and tool-to-tissue forces and torques from such robots.

II. THE STEADY-HAND EYE ROBOT

A. SHER-2.1

SHER-2.1 is a 5 degree of freedom (DoF) Cartesian serial robot as shown in Fig.1(a). The three prismatic actuators provide translation in the x, y, and z directions. The extension arm mounted on the vertical sliding rail has an additional 2-DoF of rotation about the x and y axes. The robotic extension arm includes a compact tool holder with a quick release mechanism for inserted surgical instruments, a stiff structural design, and generates a sufficient work space for tool rotation [9].

Fig. 1:

Fig. 1:

CAD models of (a) the cartesian serial robot: SHER-2.1, and (b) the delta parallel robot: SHER-3.0.

B. SHER-3.0

SHER-3.0 is a 5-DoF delta parallel robot as shown in Fig.1(b). The delta system uses three vertical arranged prismatic actuators to move the delta arms. This provides translation along the x, y, and z directions while keeping the delta platform horizontal. The robotic extension arm mounted on the delta platform has an additional 2-DoF of rotational motion about the x and y axes like SHER-2.1. SHER-3.0, however, has a simplified four-bar linkage mechanism. This mechanism offered more hand space to surgeons due to a longer extension arm comparied to SHER-2.1, thus improving comfort and workspace limitations during use. The four-bar linkage mechanism is actuated by an offset slider-crank actuated by a rotary actuator [15]. As SHER-2.1 has been used successfully in previous studies, SHER-3.0 gains were tuned before any experimentation to demonstrate a similar response and feel to that of SHER-2.1 when moving a tool.

C. FBG Force Sensing Tool

A Fiber Bragg Grating (FBG) force sensing tool mounted on the surgical instrument is used for measuring amount of force applied to the tool directly from its interaction with eye tissue. [12]. The tool shaft is made of a microsurgical light pipe with three FBG fibers (Technica S.A., Beijing, China) arranged at 120° intervals around the light pipe as shown in Fig. 2b. Each fiber contains three separate 3 mm-long FBG sensors as shown in Fig. 2c.

Fig. 2:

Fig. 2:

Design and dimensions of an FBG force sensing tool. (a) The tool shaft interacting with an eyeball phantom. (b) Section view and geometry of the tool shaft [12]. (c) The dimensions of the 3 mm-long FBG sensors.

III. ADMITTANCE CONTROL SCHEME

Both SHER robots employ an admittance control algorithm that allows surgeons to cooperatively manipulate the tool attached to the robot force-torque sensor, and guide the SHER towards a desired configuration. The 6-DoF sensor outputs a wrench vector Fhb=[fb,τb]T6, which includes the hand-to-tool interaction force fb3 and torque τb3 exerted to the robot handle, represented in the robot body frame as shown in Fig. 3. This wrench Fhb is then used in the admittance control algorithm to generate a desired end-effector velocity vector in the robot body frame denoted as Vdb=[vd,ωd]6 as follows

Vdb=KFhb (1)

in which K6×6 is a diagonal matrix of constant gains, and vd and ωd are the desired end-effector linear and angular velocities respectively, represented in the robot body frame. Equation 1 shows that the desired end-effector velocity Vdb is proportional to the hand-tool interaction force Fhb and thus the level of proportionality for each DoF can be separately adjusted by the magnitude of the diagonal entries of matrix K.

Fig. 3:

Fig. 3:

Block diagram showing the admittance control mode used in this experiments to synergically control the SHER. The transformation between the body frame {B} to the handle frame {H} is the same for both robots.

Once the desired end-effector velocity Vdb is found in task space, the desired joint space angular velocity vector can be computed as follows

Θ˙d=Jsb(Θ)Vdb (2)

where Θ5 denotes the joint variables, Θ˙d5 is the desired joint space angular velocity, Jsb6×5 is the robot Jacobian which maps joint velocities to end-effector velocities between the robot spatial frame (shown in Fig. 3) and the robot body frame, and Jsb5×6 is the Jacobian pseudoinverse which maps the end-effector velocity to joint velocities and is computed below as in [16].

Jsb=JsbT(JsbJsbT)1 (3)

At each defined time step the low-level motor controller generates desired joint velocities Θ˙d to command the robot motors for corresponding end-effector velocities Vdb comeing from the computed Jacobians.

IV. EXPERIMENT

An pilot-study was designed to quantify the cooperative-control performance between the user and each SHER by an-alyzing both user hand-to-tool interaction forces and torques, and tool-to-eye interaction forces for a vessel-following task. A right-handed, non-clinical user was used to collect data.

A. Experimental Setup

Both experimental setups are depicted in Fig.4, which included either SHER, an ATI force-torque sensor (Nano17, ATI Industrial Automation), an FBG-equipped force-sensing tool with the interrogator, a silicone eye phantom, a secondary stabilization tool, and a Zeiss surgical microscope. A desktop computer was connected to the SHERs, force-torque sensor, and FBG-equipped tool using a TCP-IP connection.

Fig. 4:

Fig. 4:

Experimental setup including the SHER, the ATI Nano17 force/torque sensor, the FBG-equipped force sensing tool, the secondary tool, the eye phantom, and the Zeiss surgical microscope.

The silicone eye phantom shown in Fig. 4 with four colored vessels painted onto the retina was placed under the surgical microscope for the user to view during tool manipulation. Eye phantoms are anatomically-similar eye models commonly used during benchtop experiments that imitate some or many properties of the human eye [17].

B. Experimental Procedure

To simulate VRS tasks, the user held the tool shaft attached to the SHER and inserted the tool into the eye phantom through a sclerotomy made on the phantom sclera tissue. The user then moved the tool tip along with the attached CC robots. Both robots used the previously described admittance control during operation.

The user was tasked to follow four colored lines mimicking eye vessels with the tip of the FBG force-sensing tool. This vessel-following task is a simplified action, inspired from motions used in retinal membrane peeling where clinicians inspect the retina, move the eye, and navigate tools to specific locations to grab tissue. Before starting each trial, the tool tip was placed “home” on the center of the circle (shown in Fig.4) at the intersection point of the four colored vessels. The user then followed a vessel one at a time, attempting to keep the central axis of the tool inline with the center axis of the followed vessel at all times. A secondary tool was used by the user to better control manipulation of the phantom eye for stabilization or rotation during vessel following.

After the FBG force sensing tool was inserted into the eyeball through the sclerotomy, the user moved the tool tip along each vessel until reaching the end of the vessel, then backtracked to the starting point along the same path. This “moving back and forth” procedure was repeated until the user finished the trajectory on all four vessels. Completion of all four vessel following tasks was recorded as one successful trial. To eliminate potential bias, the color sequence order was randomized for each trial.

To ensure the user had similar operating proficiency on both robots, 50 practice trials were conducted by the user on each robot prior to official experimentation. To quantify user proficiency, learning curves were constructed by plotting the average value of handle forces and torques of each trial. According to learning curve theory, as the the number of practice trials increases, eventual convergence of the performance curve should occur. We employed CUSUM analysis to assess the number of trials required for a participant to achieve a stable phase. CUSUM is a statistical method used to detect changes in the mean value of a process over time, and is often employed to monitor individual or group learning progress [14]. CUSUM is useful in determining whether a participant has attained a stable level of performance, with a turning point on the CUSUM curve indicating a significant improvement in performance. In this study, the CUSUM curve was used as the performance curve, and defined as CUSUM(n)=n=150(X¯Xn), where X¯ was the mean value of overall force or torque of the 50 trials, and Xn was the mean value of force or torque of the nth trial. After the learning phase, the user was asked to perform 10 trials of vessel following using each robot. To further eliminate the effect of uncertainties, the user conducted 10 trials with SHER-2.1 in the morning and with SHER-3.0 in evening and performed the same task on the next day in reverse order.

V. RESULTS AND DISCUSSION

The average values of handle force and torque for each trial of the SHER-2.1 and the SHER-3.0 are plotted in Fig. 5a and Fig. 5b for visualizing the user’s level of proficiency on each robot. Seen is a negative correlation between the number of trials and the hand-to-tool interaction force and torque for both robots. This shows tool forces decrease over time as the user learns to control the tool using the SHER. Force and torque values converged within the 50 trial learning stage demonstrating proper learning of each SHER by the user before the following experiments. The notable turning point is identified as the lowest point occuring in the CUSUM curve. In this experiment, the turning point marks the boundary between the initial and stable phases of the learning curve. This is seen in Figs. 5c and 5d as the turning points line up well with the start of the stable phases of the force and torque learning curves seen in 5a and 5b respectively for each SHER. Fig. 5c shows both the hand-to-tool interaction forces and torques reached the plateau of the learning curve after 10 trials using SHER-2.1. Similarly for the SHER-3.0, Fig.5d shows the user’s hand-to-tool interaction forces and torques reached a stable level after 15 trials and 6 trials respectively. Therefore, based on the results from the CUSUM graphs and fitted polynomial curves, we see the user reach a stable level of proficiency on both robots after 15 trials without significant change in later trials.

Fig. 5:

Fig. 5:

The original data points with third-degree polynomial fitting. (a) The learning curves of the user’s average hand-to-tool interaction force and torque using the SHER-2.1. (b) The learning curves of the user’s average hand-to-tool interaction force and torque using the SHER-3.0. (c) The CUSUM curves of handle force and torque using the SHER-2.1. (d) The CUSUM curves of handle force and torque using the SHER-3.0. For better visualization results, (c) and (d) are plotted in different scales.

Fig. 6 shows box plots representing distributions of handle torque and force data for all time intervals of the vessel following task during the experiment of 10 trials comparing both robots. Fig. 6a and Fig. 6b display handle forces, showing the distribution of the median values of the handle forces for the two robots fluctuated around 0.4 N and 0.5 N respectively. Fig. 6c and Fig. 6d display handle torques, showing the median values for SHER-2.1 to be about 0.015 Nm, which is about double that for SHER-3.0. Thus the user is required to apply greater torques to change the orientation of the SHER-2.1 tool as compared to when using SHER-3.0. The different extension arm length most likely impacted these results.

Fig. 6:

Fig. 6:

Data distribution of (a) hand-to-tool interaction force for the SHER-2.1 (b) hand-to-tool interaction force for the SHER-3.0 (c) hand-to-tool interaction torque for the SHER-2.1 (d) hand-to-tool interaction torque for the SHER-3.0

Fig. 7 shows box plots representing distributions of scleral forces for all time intervals of the vessel following task during the experiment of 10 trials comparing the SHER-2.1 and the SHER-3.0. Most values are concentrated between 100 mN and 300 mN with no observable shift in the distribution of sclera force among the 10 trials. This indicated that the scleral insertion force for both robots remained consistent across all trials.

Fig. 7:

Fig. 7:

Data distribution of (a) tool-to-eye interaction force for the SHER-2.1 (b) tool-to-eye interaction force for the SHER-3.0

Table I summarizes results from the 10 trials of vessel following using the two robots. The table also includes information on the dimensions and weights of the robots. The overall Cartesian size of each robot was measured as X×Y×Z as shown in Fig. 1, with length, width, and height measured when the robots were in their homed configurations. A two-sample t-test was performed to test the null hypothesis that the two data sets would have have no significant differences between their mean handle forces and torques, mean and max scleral forces, and average operating times [18] using a P-value of 0.05 to determine significance. The average user-robot interaction force values for SHER-2.1 were significantly lower than those using SHER-3.0 (0.437 N and 0.535 N respectively, P=3.147×108). The user applied a significantly smaller average torque using SHER-3.0 than using SHER-2.1 (0.0178 Nm versus 0.0087 Nm, P=1.649×109). Importantly, the SHER-3.0 also had a significantly smaller maximum scleral interaction force (1582.1 mN versus 1971.59, P=0.0151) although the mean scleral force, average operating time, and procedure time p-values were not found to be significantly different. Although there was variability in individual trials and some significant differences experienced between the two robots, the overall performance using both robots was generally similar. Though the reduction of maximum scleral forces is ideal as demonstrated by SHER-3.0 it is difficult to conclude that SHER-3.0 outperformed SHER-2.0 (or vice versa) though significant differences were also found between mean handle forces and torques.

Table I:

The comparison results of the ten random trials. A p-value less than 0.05 is considered to be statistically significant.

Mean handle force (N) Mean handle torque (N·m) Mean sclera force (mN) Max sclera force (mN) Average operating time (s) Cartesian size (cm) Weight (kg)
SHER-2.1 0.437 0.0178 341.271 1971.59 37.91 90×51×67 27.2
SHER-3.0 0.535 0.0087 334.503 1582.1 36.549 50×24×44 9.1
P-value 3.147×10−8 1.649×10−9 0.742 0.0151 0.328 N/A N/A

Fig. 8 shows the user hand-to-tool interaction force and torque during a single trial, using the two robots to perform the standard vessel following sequence: Red-Blue-Green-Yellow (RBGY). The handle torque of the SHER-2.1 can be seen is noticeably larger than the SHER-3.0 in this trial, though handle torque differences are not so apparent.

Fig. 8:

Fig. 8:

Handle force and torque magnitude variations for a single trial of each experiment set for the SHER-2.1 and the SHER-3.0.

Fig. 9 shows the variation of the scleral forces that each robot tool applied to the eye phantom during a single trial corresponding to Fig. 8. The scleral force of the SHER-2.1 is seen generally consistent with the SHER-3.0 during operative time but with larger maximum forces. It should be noted that the scleral forces always reached the maximum allowed peak when the user followed the green vessel. This was due to the position of the insertion point was too close to the green vessel. This forced the user to use a steep angle to follow this vessel, leading to applying a larger force to the eye phantom. During eye surgery, this situation can be avoided by changing the scleral insertion point.

Fig. 9:

Fig. 9:

Scleral force magnitude variations for a single trial of the SHER-2.1 and the SHER-3.0.

Fig. 12 visualizes the calculated tool tip trajectories (assuming the tool is a rigid body) using both robots for a single trial along RBGY vessel paths. Fig. 10 shows the insertion depth variations on the z-axis, where the zero value represents the initial insertion depth of the tool at the start point. Fig. 11 shows the magnitude of the tool tip velocity and acceleration for the single trial RBGY of the two robots. These characteristics suggest consistency in user interaction between the SHER-2.1 and the SHER-3.0.

Fig. 12:

Fig. 12:

Tool tip trajectories for a single trial calculated from the forward kinematics of the SHER-2.1 and the SHER-3.0.

Fig. 10:

Fig. 10:

Insertion depth variations for a single trial of the SHER-2.1 and the SHER-3.0.

Fig. 11:

Fig. 11:

The magnitude of tool tip velocities and accelerations for a single trial of the SHER-2.1 and the SHER-3.0.

VI. CONCLUSION AND FUTURE WORK

This study presents a preliminary comparative study of standardized vessel following performance while using both a CC serial robot (SHER-2.1) and parallel robot (SHER-3.0). Three major differences between the two robots were identified by these methods and include the mean handle torque, the mean scleral force, and the maximum sclera force applied to the eye phantom during the operation. Based on the experimental results, we can conclude that with the same amount of training effort spent using both robots, the two robots had similar performance during a standardized vessel-following task. An important aspect to consider is that SHER-3.0 is smaller and lighter than the SHER-2.1, making it easier to deploy in operating rooms. Additionally, it is important to highlight that SHER-2.1 is a well-functioning robot that has been validated through in vivo tests [19] and the similar performance observed in SHER-3.0 suggests that it holds great potential as a next-generation eye surgical platform. However, it is essential to acknowledge that further studies and surgery-specific tuning are necessary to optimize its outcomes and validate its effectiveness.

Future comparative studies could include real-time tool tip trajectory analysis with computer vision algorithms possibly applying virtual boundaries to improve following outcomes. Additionally, increasing the number of users in the study could help validate and generalize results. This would ideally evaluate outcomes between non-clinical personnel compared to experienced eye surgeons. A proper study comparing outcomes between surgeon with various experience, would also be interesting and valuable for future exploration of CC surgical robots. Testing a task for targeting various positions in the retina (i.e. touching an array of grid points using both robots) would also be valuable for evaluating the robots for tasks such as retinal membrane peeling.

Acknowledgments

This work was supported by U.S. National Institutes of Health under the grants number 2R01EB023943-04A1 and 1R01 EB025883-01A1, and partially by JHU internal funds.

References

  • [1].Nguyen KH, Patel BC, and Tadi P. (2023, Jan) Anatomy, head and neck: Eye retina. Treasure Island (FL): StatPearls Publishing. NIH, National Library of Medicine. [Online]. Available: https://www.ncbi.nlm.nih.gov/books/NBK542332/ [PubMed] [Google Scholar]
  • [2].Vander Poorten E, Riviere CN, Abbott JJ, Bergeles C, Nasseri MA, Kang JU, Sznitman R, Faridpooya K, and Iordachita I, “Robotic retinal surgery,” in Handbook of Robotic and Image-Guided Surgery. Elsevier, 2020, pp. 627–672. [Google Scholar]
  • [3].Yang S, MacLachlan RA, and Riviere CN, “Manipulator design and operation of a six-degree-of-freedom handheld tremor-canceling microsurgical instrument,” IEEE/ASME transactions on mechatronics, vol. 20, no. 2, pp. 761–772, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Guerrouad A and Vidal P, “Smos: stereotaxical microtelemanipulator for ocular surgery,” in Images of the Twenty-First Century. Proceedings of the Annual International Engineering in Medicine and Biology Society,. IEEE, 1989, pp. 879–880. [Google Scholar]
  • [5].Meenink T, Naus G, de Smet M, Beelen M, and Steinbuch M, “Robot assistance for micrometer precision in vitreoretinal surgery,” Investigative ophthalmology & visual science, vol. 54, no. 15, pp. 5808–5808, 2013. [Google Scholar]
  • [6].Nasseri MA, Eder M, Nair S, Dean E, Maier M, Zapp D, Lohmann CP, and Knoll A, “The introduction of a new robot for assistance in ophthalmic surgery,” in 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE, 2013, pp. 5682–5685. [DOI] [PubMed] [Google Scholar]
  • [7].Mablekos-Alexiou A, Ourselin S, Da Cruz L, and Bergeles C, “Requirements based design and end-to-end dynamic modeling of a robotic tool for vitreoretinal surgery,” in 2018 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2018, pp. 135–141. [Google Scholar]
  • [8].Gijbels A, Wouters N, Stalmans P, Van Brussel H, Reynaerts D, and Vander Poorten E, “Design and realisation of a novel robotic manipulator for retinal surgery,” in 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2013, pp. 3598–3603. [Google Scholar]
  • [9].He X, Roppenecker D, Gierlach D, Balicki M, Olds K, Gehlbach P, Handa J, Taylor R, and Iordachita I, “Toward clinically applicable steady-hand eye robot for vitreoretinal surgery,” in ASME International Mechanical Engineering Congress and Exposition, vol. 45189. American Society of Mechanical Engineers, 2012, pp. 145–153. [Google Scholar]
  • [10].Pandilov Z and Dukovski V, “Comparison of the characteristics between serial and parallel robots.” Acta Technica Corviniensis-Bulletin of Engineering, vol. 7, no. 1, 2014. [Google Scholar]
  • [11].Nzue R-MA, Brethé J-F, Vasselin E, and Lefebvre D, “Comparison of serial and parallel robot repeatability based on different performance criteria,” Mechanism and Machine Theory, vol. 61, pp. 136–155, 2013. [Google Scholar]
  • [12].He C, Yang E, Patel N, Ebrahimi A, Shahbazi M, Gehlbach P, and Iordachita I, “Automatic light pipe actuating system for bimanual robot-assisted retinal surgery,” IEEE/ASME Transactions on Mechatronics, vol. 25, no. 6, pp. 2846–2857, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Ebrahimi A, He C, Roizenblatt M, Patel N, Sefati S, Gehlbach P, and Iordachita I, “Real-time sclera force feedback for enabling safe robot-assisted vitreoretinal surgery,” in 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE, 2018, pp. 3650–3655. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [14].Nagakawa Y, Nakamura Y, Honda G, Gotoh Y, Ohtsuka T, Ban D, Nakata K, Sahara Y, Velasquez VVD, Takaori K et al. , “Learning curve and surgical factors influencing the surgical outcomes during the initial experience with laparoscopic pancreaticoduodenectomy,” Journal of Hepato-Biliary-Pancreatic Sciences, vol. 25, no. 11, pp. 498–507, 2018. [DOI] [PubMed] [Google Scholar]
  • [15].Roth R, Wu J, Alamdar A, Taylor RH, Gehlbach P, and Iordachita I, “Towards a clinically optimized tilt mechanism for bilateral micromanipulation with steady-hand eye robot,” in 2021 International Symposium on Medical Robotics (ISMR). IEEE, 2021, pp. 1–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Chiacchio P, Chiaverini S, Sciavicco L, and Siciliano B, “Closed-loop inverse kinematics schemes for constrained redundant manipulators with task space augmentation and task priority strategy,” The International Journal of Robotics Research, vol. 10, no. 4, pp. 410–425, 1991. [Google Scholar]
  • [17].Gupta A, Gonenc B, Balicki M, Olds K, Handa J, Gehlbach P Taylor RH, and Iordachita I, “Human eye phantom for developing computer and robot-assisted epiretinal membrane peeling,” in 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE, 2014, pp. 6864–6867. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18].Montgomery DC, Design and analysis of experiments. John wiley & sons, 2017. [Google Scholar]
  • [19].Patel N, Urias M, Ebrahimi A, Taylor RH, Gehlbach P, and Iordachita I, “Force-based control for safe robot-assisted retinal interventions: In vivo evaluation in animal studies,” IEEE Transactions on Medical Robotics and Bionics, vol. 4, no. 3, pp. 578–587, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES