Skip to main content
Frontiers in Medicine logoLink to Frontiers in Medicine
. 2026 Mar 5;13:1688748. doi: 10.3389/fmed.2026.1688748

Precision surgery in the era of 3D visualization, AR/VR, and 3D printing: current applications and future directions

Xiangxiang Ren 1, Tianhao Xie 1, Xiaoshi Jin 1, Litao Liu 1, Lingyun Liu 1, Meng Zhang 2,*
PMCID: PMC12999902  PMID: 41868228

Abstract

The advancement of precision surgery demands increased operative accuracy, underscoring the growing importance of three-dimensional (3D) visualization technology as a key tool for overcoming the limitations of two-dimensional (2D) imaging. Over the past decade, this technology has evolved from a post-processing tool into an integrated intelligent platform spanning the entire surgical workflow: preoperative assessment, surgical planning, intraoperative navigation, postoperative follow-up, and teaching/training. This report comprehensively reviews the current applications and key advances of 3D reconstruction, 3D printing, augmented reality (AR), virtual reality (VR), and mixed reality (MR) technologies across major general surgery subspecialties. These include hepatobiliary, pancreatic, gastrointestinal, thyroid/breast, hernia/abdominal wall, organ transplantation, and pediatric surgery. Through in-depth analysis, the review elucidates how these technologies facilitate precision surgery, objectively assesses current technical limitations and ethical/regulatory challenges, and explores future directions driven by artificial intelligence (AI), 5G/6G telecommunications, and digital twin technology.

Keywords: 3D printing, 3D visualization, augmented reality, general surgery, precision surgery, virtual reality

1. Introduction

The era of precision surgery demands unprecedented capabilities from surgeons to achieve the objectives of “seeing clearly, calculating precisely, and operating steadily.” Traditional 2D medical imaging modalities—such as computed tomography (CT), magnetic resonance imaging (MRI), and ultrasound—provide invaluable anatomical information. However, their planar display inherently leads to a loss of spatial depth perception and obscures critical relationships between adjacent tissues. Surgeons must therefore perform complex mental conversions from 2D to 3D representations, a process that increases cognitive load and introduces potential for misjudgment.

Three-dimensional (3D) visualization technology has emerged as a pivotal solution. Its evolution has progressed from early static reconstructions to dynamic interactive models, and now to sophisticated intelligent platforms integrating AI, AR, and VR. Within general surgery, applications have rapidly expanded beyond initial focuses like hepatobiliary planning to encompass the entire surgical workflow, including preoperative precision assessment, navigation for complex procedures, personalized treatment planning, postoperative follow-up, and teaching. This transformation is profoundly reshaping surgical paradigms (1).

3D visualization technology addresses the inherent limitation of 2D imaging by integrating multimodal data—such as CT, MRI, ultrasound, and positron emission tomography (PET)—and applying advanced computer graphics algorithms. It generates intuitive, volumetric, and interactive representations of complex anatomy, lesion morphology, and their spatial relationships with surrounding critical structures (e.g., vasculature, bile ducts) (2–6). This “god’s-eye view” provides a robust technical foundation for achieving personalized, minimally invasive, and function-preserving precision surgery.

2. Overview of key technologies

The 3D visualization platform is a sophisticated integrated system encompassing multiple cutting-edge technologies. The following sections outline the core modules and their recent progress.

2.1. 3D reconstruction and visualization

Three-dimensional reconstruction forms the foundation, aiming to precisely segment and render 3D anatomical models from sequences of 2D images. Early techniques relied on semi-automated methods like region growing based on intensity thresholding, combined with volume or surface rendering (7). These methods were operator-dependent and demonstrated limited efficiency and accuracy for soft tissues with poorly defined boundaries.

In recent years, the introduction of deep learning, particularly convolutional neural networks (CNNs), has revolutionized image segmentation by enhancing automation and accuracy. Models like the U-Net architecture and its variants have become the de facto standard in medical image segmentation (8, 9). U-Net’s encoder-decoder structure with skip connections enables effective capture of multi-scale features for precise segmentation. Building on this, the nnU-Net (no-new-Net) framework further reduces application barriers by adaptively configuring optimal preprocessing, network topologies, training schemes, and postprocessing for different datasets (10, 11) (Table 1).

Table 1.

Key performance metrics of nnU-Net in various segmentation tasks, highlighting its clinical relevance.

Segmentation task Model/method Dice similarity coefficient (DSC) 95% Hausdorff distance (95% HD, mm) References Remarks
Liver segmentation nnU-Net 0.974 (mean) 2.458 (72) Comparable to or exceeding expert-level performance
Specific organ nnU-Net Up to 0.9255 (92.55%) Significantly reduced (7, 73) Performance varies by specific organ studied
Abdominal organs nnU-Net Typically >0.90 Significantly reduced (7, 73, 74) Consistent performance across organs (e.g., liver, pancreas)

Additionally, to meet demands for real-time intraoperative interaction, rendering technologies have evolved significantly. Cloud-based GPU computing and real-time ray tracing now enable smooth manipulation—including rotation, scaling, transparency adjustment, and virtual dissection—of high-quality, large-scale 3D models. These advances provide the computational foundation for complex planning and navigation (12).

2.2. 3D printing

3D printing technology transforms digital 3D models into tangible physical models, providing a powerful platform for surgical planning, doctor-patient communication, and education. Modern multi-material, multi-color printers enable precise 1:1 replication of patient-specific anatomy. By utilizing materials with varying colors and rigidity to represent different structures (e.g., tumors, arteries, bile ducts), these models make complex spatial relationships immediately apparent (13–15). Surgeons can perform repeated simulated dissection and rehearsal on these models, optimizing strategies and anticipating risks. Beyond anatomical replicas, 3D printing shows significant potential for fabricating personalized medical devices like bioresorbable scaffolds and patient-specific implants (16–19).

2.3. Augmented reality, virtual reality, and mixed reality

Virtual Reality technology uses head-mounted displays (HMDs) to immerse users in computer-generated virtual surgical environments, providing an ideal platform for risk-free, repeatable skills training. In contrast, AR and MR technologies overlay virtual 3D anatomical models onto the real-world view. Surgeons wearing see-through HMDs, such as HoloLens (Microsoft) or Magic Leap, can visualize 3D models superimposed directly onto the patient’s anatomy, achieving a “radiographic perspective” effect (20, 21).

HoloLens (primarily v2) is widely adopted in healthcare, with substantial experience in neurosurgical and orthopedic navigation. Its strengths include a relatively mature ecosystem and validated medical applications (22). A primary limitation is its relatively narrow field of view (FOV) (23).

Magic Leap (One/Two) is characterized by a larger FOV and advanced light-field display technology. It offers more immersive virtual-physical integration and shows potential in applications requiring broader visual coverage. However, its ecosystem and medical application maturity are still developing compared to HoloLens (24).

A core challenge for AR/MR navigation systems is registration—the precise alignment of virtual models with the patient’s actual anatomy. By integrating tracking systems, leading platforms achieve target registration errors (TRE) at the millimeter level for static organs [e.g., ~1.3 mm in liver surgery (25)]. However, for abdominal soft tissues, displacement caused by respiratory motion and surgical manipulation remains a significant challenge, potentially increasing errors to 5–12 mm (26). These error ranges currently limit AR/MR to adjunctive roles for orientation and gross anatomical guidance rather than millimeter-precision tasks such as vessel dissection or nerve sparing. Clinical tasks that can tolerate such errors include identifying general tumor location, visualizing major vascular courses, or guiding dissection planes in well-exposed, relatively static fields.

Research into haptic and force feedback technologies is progressing to enhance immersion by enabling users to perceive tissue resistance and texture when manipulating virtual instruments, which is crucial for simulating advanced procedures (27).

2.4. Technology comparison and selection guidelines

Different 3D visualization technologies possess distinct advantages suited to specific clinical scenarios. Table 2 summarizes their core characteristics to guide clinical selection, aiding clinicians in choosing the most appropriate technology based on procedural needs and available resources.

Table 2.

Comparison of key 3D visualization technologies and guidelines for clinical application.

Technology Core advantages Primary limitations Typical application scenarios
3D reconstruction/visualization • Provides intuitive 3D anatomical insights.
• Enables virtual dissection, measurement, and surgical simulation.
• Forms the foundation for other technologies (3D printing, AR/VR).
• High-quality segmentation can be time-intensive (complex/low-quality images).
• Pure screen display lacks tactile realism.
• Requires high-performance hardware for real-time rendering.
• Foundational for all applications.
• Preoperative detailed assessment and planning.
• Simulation of complex procedures.
• Educational demonstrations (screen-based).
3D printing • Provides tangible, physical models.
• Excellent tool for doctor-patient communication and education.
• Enables fabrication of customized implants/devices.
• Model production is time-consuming (hours to days).
• Relatively high cost (equipment, materials).
• Models cannot dynamically reflect intraoperative changes.
• Teaching and understanding complex anatomy.
• Discussion and rehearsal of complex surgical plans (preoperative).
• Fabrication of personalized medical devices.
• Doctor-patient communication requiring physical demonstration.
Virtual reality (VR) • Delivers fully immersive experiences.
• Enables risk-free, repeatable surgical skills training.
• Can simulate rare cases and complications.
• Differs from real surgical environments.
• Lacks realistic haptic feedback (advanced systems under development).
• May cause cybersickness in some users.
• Foundational skills training for surgical residents (suturing, knot tying, basic maneuvers).
• Rehearsal of complex surgical workflows.
• Immersive anatomy learning.
Augmented/mixed reality (AR/MR) • Overlays virtual information onto the real-world/surgical field.
• Provides intraoperative real-time navigation and "radiographic perspective" capability.
• Enables hands-free operation (HMD).
• Registration accuracy significantly impacted by organ motion (especially soft tissue).
• Device comfort, field-of-view (FOV) limitations, battery life.
• High system cost; clinical validation data still accumulating.
• Intraoperative real-time navigation of anatomy (e.g., vessels, nerves, tumor location).
• Surgical pathway guidance (e.g., biopsy, liver resection planes).
• Remote expert collaboration and guidance.
• Complex anatomy teaching (overlaid on real specimens/models).

3. Application in hepatobiliary surgery

Among general surgery subspecialties, hepatobiliary surgery represents one of the most extensively validated and impactful domains for 3D visualization technology. As shown in Figure 1, it is systematically integrated across the workflow from preoperative assessment through surgical planning and intraoperative navigation to postoperative evaluation. It enables precision segmentation of complex anatomical variants and volumetric quantification of future liver remnants (FLR), fundamentally optimizing strategy formulation for hepatectomy and biliary reconstruction.

Figure 1.

Flowchart showing preoperative CT or MRI image acquisition leading to 3D reconstruction of liver, vessels, bile ducts, and tumor, then branching to virtual hepatectomy and FLR calculation for optimized surgical planning, 3D printed physical model for planning and simulation, and AR/MR intraoperative navigation for real-time guidance of resection plane, vessel, and bile duct identification.

Flowchart of the application of three-dimensional visualization technology in hepatobiliary surgery.

3.1. Living donor liver transplantation and complex hepatectomy

In living donor hepatectomy, precise assessment of graft volume and the anatomy of vasculature and bile ducts is critical. 3D visualization provides clear delineation of vascular branching patterns, enabling optimal transection plane selection to ensure donor safety and graft viability. It plays a critical role in complex procedures like right-lobe donor hepatectomy and ALPPS (Associating Liver Partition and Portal vein ligation for Staged hepatectomy) (14, 28).

A prospective single-center study of 40 living donors demonstrated that preoperative 3D reconstruction-based virtual hepatectomy and FLR volumetry achieved a mean error of <5%, compared to 15–20% with conventional 2D CT volumetry (p < 0.01) (29). A multicenter study further reported that 3D surgical planning in complex ALPPS procedures substantially reduced operative time and blood loss (30). It should be noted that these improvements are based on cohort comparisons rather than randomized trials, and further validation in larger, controlled studies is warranted.

Preoperative virtual hepatectomy with FLR calculation now achieves prediction errors consistently below 5%, markedly outperforming traditional 2D methods and contributing to reduced post-hepatectomy liver failure rates (31–34). For challenging Bismuth type IV hilar cholangiocarcinomas, preoperative 3D-printed models displaying tumor invasion and perivascular relationships have become standard practice at major centers, facilitating team discussions and rehearsal (35–38).

4. Application in pancreatic surgery

Pancreatic surgery, particularly pancreaticoduodenectomy (PD), represents a pinnacle of technical complexity due to the organ’s retroperitoneal location and vascular proximity. 3D visualization provides critical solutions. Anatomical variations of peripancreatic vessels are highly prevalent and constitute major risk factors for hemorrhage. 3D visualization (e.g., via multiphasic CT reconstruction) achieves up to 96% sensitivity in detecting these variations, directly reducing vascular injury risk (39). It further enables intraoperative navigation with real-time anatomical registration (40, 41).

Preoperative identification of hepatic arterial variants (e.g., replaced right hepatic artery) is essential for planning vascular reconstruction (42). 3D-printed models of these complex relationships enhance trainee comprehension (43). Additionally, VR surgical simulators offer an efficient training platform for PD. Repeated rehearsal in virtual environments shortens the learning curve, with studies demonstrating reduced operative time and improved management of emergencies among VR-trained surgeons (44–46). For example, resident physicians who underwent VR training achieved 29% faster gallbladder dissection and demonstrated a 6-fold reduction in the mean number of errors committed (1.19 vs. 7.38 errors per case, p < 0.008, Mann–Whitney U test). By contrast, residents without VR training were 5 times more likely to incur gallbladder injury or thermal damage to non-target tissues (χ2 = 4.27, p < 0.04) (45).

Although limited pancreatic mobility favors AR navigation, registration accuracy [typically 3–8 mm (26)] remains a barrier to widespread adoption. Current systems are not yet reliable for guiding precise vascular dissection but may assist in identifying major anatomical landmarks during initial exposure.

5. Application in gastrointestinal surgery

Figure 2 illustrates the application workflow of 3D visualization technology in gastrointestinal surgery. The emphasis on Total Mesorectal Excision (TME) and Complete Mesocolic Excision (CME) principles has heightened demands for precision. While 3D laparoscopic systems provide enhanced depth perception—improving identification of planes and reducing circumferential resection margin (CRM) positivity (47)—3D visualization offers further advancements.

Figure 2.

Flowchart illustrating the process of 3D reconstruction and surgical planning starting with preoperative CT or MRI imaging, followed by image preprocessing, multiplanar reconstruction, segmentation of anatomical structures, and model validation, leading to anatomical assessment, intraoperative navigation with AR guidance, and 3D laparoscopic integration, each branching into detailed intraoperative applications and planning steps.

Flowchart of the application of three-dimensional visualization technology in gastrointestinal surgery.

In low rectal cancer, preoperative 3D reconstruction of pelvic anatomy, particularly tumor relationships with pelvic walls and nerve plexuses, is critical for developing sphincter-preservation and neuroprotective strategies. During lateral lymph node dissection, AR navigation can superimpose preoperatively mapped pelvic plexus pathways onto the laparoscopic view, providing real-time avoidance guidance to preserve function while ensuring radicality (48, 49).

For gastric cancer, 3D visualization is equally valuable. D2 lymphadenectomy requires meticulous clearance around key vessels. Preoperative 3D models detect vascular variants with >95% accuracy—surpassing conventional imaging (~80%) (50, 51)—enabling anticipation of challenges. A study demonstrated 30% fewer intraoperative bleeding events and higher lymphadenectomy completeness scores with 3D surgical simulation (52, 53). However, this study was a single-center retrospective analysis (n = 20); prospective multicenter data are needed to confirm these benefits.

Technical Note: Clinical validation data quantifying target registration errors for AR navigation in rectal cancer [reported range: 4–7 mm (54)] remain limited. While current systems achieve errors of 1–3 mm for static structures, errors of 5–12 mm are common in abdominal soft tissues due to deformation. Such inaccuracies limit use for millimeter-precision tasks, relegating systems to adjunctive roles for orientation. For example, AR may help identify the general path of the pelvic plexus but cannot reliably guide nerve-sparing dissection at the millimeter level.

6. Application in thyroid and breast surgery

In thyroid cancer surgery, preserving the recurrent laryngeal nerve (RLN) and parathyroid glands defines procedural quality. Integrating preoperative 3D-reconstructed models with intraoperative indocyanine green (ICG) fluorescence imaging enables dual-modality navigation—combining anatomical mapping with real-time functional assessment. While 3D models delineate nerve trajectories, ICG visualizes parathyroid perfusion. This synergy reduces permanent RLN injury during central neck dissection to <1% (55, 56).

7. Application in hernia and abdominal wall surgery

Complex abdominal wall defects present significant challenges. 3D visualization enables precise preoperative quantification of defect dimensions and assessment of muscular quality, facilitating decisions regarding advanced reconstruction techniques like the Component Separation Technique (CST). Furthermore, patient-specific 3D-printed meshes, customized from CT data, achieve optimal defect adaptation, minimizing redundancy and tension to potentially reduce recurrence (57, 58). In minimally invasive repairs, AR navigation can provide real-time highlighting of key dissection planes (e.g., preperitoneal space), guiding precise separation while minimizing injury risk.

8. Application in pediatric general surgery

Figure 3 outlines the application workflow of 3D visualization technology in pediatric surgery. In this field, small organ size and intricate anatomy demand exceptional precision. For hepatoblastoma resection, 3D visualization-guided preoperative assessment and planning have become the standard to achieve maximal tumor resection while preserving functional parenchyma (59, 60).

Figure 3.

Flowchart illustrating the process from preoperative CT or MRI acquisition, leading to 3D reconstruction for organ and tumor segmentation, and branching into surgical planning, congenital malformation evaluation, and VR-based training, each with specific outcomes for assessment, decision-making, and junior surgeon education.

Flowchart of the application of three-dimensional visualization technology in pediatric surgery.

The technology is vital in managing biliary atresia (BA), where preoperative 3D reconstruction provides stereoscopic visualization of portal vein branching patterns and spatial relationships. This delineation is critical for determining optimal dissection depth during portoenterostomy (Kasai procedure), minimizing hemorrhage risk (61).

Furthermore, VR demonstrates unique advantages in pediatric surgical education. Transforming traditional anatomy learning into gamified VR modules significantly enhances engagement among junior residents. Studies confirm a 35% improvement in anatomical knowledge test scores compared to conventional methods (p < 0.01) (62, 63).

9. Application in transplant surgery

In transplantation, 3D visualization and printing technologies primarily enhance preoperative planning and communication. For donors and recipients, 1:1 physical models of livers or kidneys enable intuitive demonstration of vascular anastomotic configurations, facilitating surgical team discussions and reportedly reducing operative time (64, 65). In split-liver transplantation, AR navigation can project predefined transection planes onto the liver surface, guiding real-time execution to minimize risks (66, 67).

10. Applications in teaching and training

3D visualization technology is revolutionizing surgical education. Trainees receiving 3D visualization instruction demonstrate superior mastery of anatomical knowledge and surgical skills (68–70). The results of a multicenter randomized controlled trial (RCT) are compelling: among 120 surgical residents, those trained on a VR simulator for laparoscopic cholecystectomy had a 4% intraoperative gallbladder perforation rate in their first 10 procedures, compared to 14% in the control group trained traditionally (p < 0.05) (71). This high-level evidence supports the integration of VR into structured training curricula.

11. SWOT analysis: current status and challenges

After a decade of development, 3D visualization technology has achieved significant successes but faces multi-dimensional challenges for widespread adoption. A systematic SWOT analysis follows:

11.1. Strengths

  • Enhanced Spatial Cognition: Overcomes 2D limitations, reducing cognitive load and misjudgment risks.

  • Empowered Precision Decision-Making: Enables precise preoperative assessment, personalized planning, and intraoperative navigation.

  • Optimized Safety and Outcomes: Confirmed value in reducing complications, shortening operative time, and improving resection rates.

  • Revolutionized Teaching and Training: VR simulators provide risk-free, immersive training; 3D/AR enhances anatomy teaching.

  • Improved Doctor-Patient Communication: 3D-printed models and visualizations improve understanding and informed consent.

11.2. Weaknesses

  • Technical Bottlenecks: Segmentation efficiency in complex cases often requires manual correction (15–45 min). Organ deformation causes registration errors of 5–12 mm in abdominal surgery, limiting precision guidance. Processing ultra-large models for real-time interaction demands high computational power.

  • Clinician Acceptance and Training Gaps: Proficiency requires significant additional time and effort. Over-reliance may erode independent judgment skills. Standardized curricula and certification are not yet widespread.

  • Cost and Accessibility: High costs for hardware/software, consumables, and IT support create economic barriers.

  • Data Management and Interoperability: Challenges in integrating multi-modal data and poor interoperability between vendor systems.

11.3. Opportunities

  • Deep AI Integration: For fully automated segmentation, intelligent biomechanical models predicting deformation, and AI-assisted planning.

  • Next,-Gen Communication and Computing: 5G/6G and cloud computing enabling cloud-based MR and remote collaboration; edge computing for low-latency processing.

  • Digital Twins and Predictive Medicine: Patient-specific “digital twins” for preoperative simulation and outcome prediction (Current Status: Largely experimental; clinical implementation remains nascent and is supported primarily by proof-of-concept studies).

  • Materials Science Innovations: More realistic, accessible 3D printing materials and bioprinting.

  • Improving Policy and Reimbursement: Growing clinical evidence may lead to more regulatory approvals and insurance coverage.

11.4. Threats

  • Regulatory and Ethical Challenges: AR/MR systems face complex medical device certification processes. Data security and privacy risks require compliance with GDPR, HIPAA, etc. Liability attribution for adverse decisions based on 3D models or AI remains unclear.

  • Regional and Institutional Resource Disparities: Adoption is concentrated in top-tier centers, with low uptake in primary care and resource-limited settings, potentially widening the digital divide.

  • Ongoing Cost-Effectiveness Pressure: High costs require robust health economic evidence to justify adoption.

  • Rapid Technological Obsolescence: Fast-paced iteration risks early investments becoming outdated.

In summary, 3D visualization technology demonstrates immense strengths in general surgery but is constrained by technical bottlenecks (especially registration), training needs, high costs, regulatory hurdles, and resource disparities. Realizing its equitable value requires capitalizing on opportunities presented by AI and 5G/6G while addressing threats related to ethics, regulations, data security, and resource equity.

12. Future directions

Looking ahead, 3D visualization technology will evolve towards greater intelligence, real-time capability, integration, and accessibility, dependent on interdisciplinary innovation and systemic support.

12.1. AI-driven automation and enhanced navigation

Future developments include achieving fully automated real-time segmentation through advanced AI architectures (e.g., Transformer, graph neural networks) and self-supervised learning, enabling robust “one-click” segmentation and eliminating manual correction. It must be emphasized that while promising, these capabilities are currently experimental and require extensive clinical validation before routine use. This will be coupled with the development of multi-organ coupled deformation prediction models integrating biomechanics to sense and compensate for organ displacement in real-time. Overcoming registration inaccuracies remains a critical hurdle. Future systems must integrate real-time biomechanical modeling and intraoperative imaging to achieve dynamic registration accuracy within 2 mm, which is essential for trusting AR/MR guidance in critical dissection.

12.2. Cloud-based collaboration and digital twins

5G/6G networks will empower cloud-based mixed reality and remote collaboration by offloading intensive computing tasks to the cloud. This can create shared surgical spaces allowing geographically dispersed experts to collaborate in real-time. The ultimate goal is developing patient-specific “digital twins” that incorporate not only high-fidelity anatomy but also simulate physiological processes, mechanical feedback, and pathological evolution. Surgeons could conduct unlimited preoperative rehearsals on these twins to evaluate outcomes and prognosis of different strategies, enabling truly predictive precision surgery. These concepts remain largely in the research phase, with few clinical implementations to date.

12.3. Integrated platforms and standardization

Future platforms will evolve into unified collaborative environments seamlessly connecting surgeons, radiologists, engineers, and other specialists. These platforms will facilitate data fusion, collaborative planning, and knowledge sharing. Establishing unified technical standards and guidelines is paramount, covering data acquisition/processing, application workflows, and standardized outcome reporting. Authoritative organizations should lead the development of evidence-based clinical guidelines.

12.4. Health economics and sustainable adoption

Rigorous cost-effectiveness and cost-utility analyses are crucial to quantify technological inputs against comprehensive outcome benefits. Reliable health economic evidence should drive payers to explore innovative value-based payment models. Concurrently, innovations in materials science and 3D printing (e.g., biocompatible smart materials, faster printing techniques, bioprinting) will reduce barriers and expand applications.

The future of 3D visualization lies in building an integrated ecosystem encompassing AI, high-speed communication, biomechanics, multidisciplinary collaboration, standardization, and value-based assessment. Only through this holistic approach can current bottlenecks be overcome, enabling the transition from “technological innovation” to “accessible clinical value.”

13. Conclusion

3D visualization technology has evolved from a supplementary tool into a core platform driving the transformation of general surgery toward precision, intelligence, and personalization. It has profoundly reshaped how surgeons understand pathology, plan procedures, execute operations, and transfer expertise.

From complex hepatobiliary-pancreatic surgeries to nerve-preservation procedures in gastrointestinal and thyroid surgery, and to personalized strategies in hernia repair and transplantation, 3D visualization now permeates critical workflows across all subspecialties. It has significantly enhanced procedural safety, accuracy, and outcomes.

Nevertheless, broad clinical adoption still faces challenges, including algorithmic inefficiency, navigation inaccuracy, regulatory barriers, and high costs. It is crucial to distinguish between technologies with strong evidence and routine use (e.g., 3D printing for planning, VR for training) and those still evolving (e.g., AR/MR navigation, digital twins).

Realizing the technology’s transition from availability to utility and from pilot demonstrations to widespread adoption hinges on collaboratively addressing these bottlenecks. Continued refinement of AI-driven algorithms and breakthroughs in motion compensation are essential. Concurrently, establishing robust validation, securing regulatory approvals, implementing cost-reduction strategies, and developing standardized training frameworks are critical.

Looking forward, the deep integration of AI, next-generation communications, and digital twins holds promise to expand the scope and depth of applications. We are confident that with further evidence, validation, and standardization, this technology will ultimately forge a new surgical paradigm that bridges the virtual and physical realms, amplifying human expertise to redefine surgical care and inaugurate a patient-centered era of precision surgery.

Funding Statement

The author(s) declared that financial support was received for this work and/or its publication. This study was supported by Research Fund of Affiliated Hospital of Hebei University (No. 2025Q08).

Footnotes

Edited by: Rahul Kashyap, WellSpan Health, United States

Reviewed by: Vinod Kumar Yata, Malla Reddy University, India

Author contributions

XR: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Software, Writing – original draft, Writing – review & editing. TX: Conceptualization, Investigation, Methodology, Supervision, Writing – review & editing. XJ: Conceptualization, Data curation, Investigation, Methodology, Project administration, Writing – review & editing. LitL: Conceptualization, Formal analysis, Methodology, Supervision, Writing – review & editing. LinL: Investigation, Methodology, Project administration, Software, Writing – review & editing. MZ: Conceptualization, Formal analysis, Investigation, Methodology, Project administration, Resources, Software, Supervision, Writing – original draft, Writing – review & editing.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1.Fang C, An J, Bruno A, Cai X, Fan J, Fujimoto J, et al. Consensus recommendations of three-dimensional visualization for diagnosis and management of liver diseases. Hepatol Int. (2020) 14:437–53. doi: 10.1007/s12072-020-10052-y, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Rowe SP, Pomper MG, Leal JP, Schneider R, Krüger S, Chu LC, et al. Photorealistic three-dimensional visualization of fusion datasets: cinematic rendering of PET/CT. Abdom Radiol (NY). (2022) 47:3916–20. doi: 10.1007/s00261-022-03614-1 [DOI] [PubMed] [Google Scholar]
  • 3.Kim J, Cai W, Feng D. Dual-modality PET-CT visualization using real-time volume rendering and image fusion with interactive 3D segmentation of anatomical structures. Conf Proc IEEE Eng Med Biol Soc. (2005) 2006:642–5. doi: 10.1109/IEMBS.2005.1616495, [DOI] [PubMed] [Google Scholar]
  • 4.Marinelli M, Positano V, Nekolla SG, Marcheschi P, Todiere G, Esposito N, et al. Hybrid image visualization tool for 3D integration of CT coronary anatomy and quantitative myocardial perfusion PET. Int J Comput Assist Radiol Surg. (2013) 8:221–32. doi: 10.1007/s11548-012-0777-3, [DOI] [PubMed] [Google Scholar]
  • 5.Gao Y, Tan K, Sun J, Jiang T, Zou XW. Application of mixed reality Technology in Visualization of medical operations. Chin Med Sci J. (2019) 34:103–9. doi: 10.24920/003564, [DOI] [PubMed] [Google Scholar]
  • 6.Robb RA. 3-D visualization in biomedical applications. Annu Rev Biomed Eng. (1999) 1:377–99. doi: 10.1146/annurev.bioeng.1.1.377, [DOI] [PubMed] [Google Scholar]
  • 7.Wang T, Wen Y, Wang Z. nnU-net based segmentation and 3D reconstruction of uterine fibroids with MRI images for HIFU surgery planning. BMC Med Imaging. (2024) 24:233. doi: 10.1186/s12880-024-01385-3, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Wang S, Singh VK, Cheah E, Wang X, Li Q, Chou SH, et al. Stacked dilated convolutions and asymmetric architecture for U-net-based medical image segmentation. Comput Biol Med. (2022) 148:105891. doi: 10.1016/j.compbiomed.2022.105891, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Zhang J, Zhang Y, Jin Y, Xu J, Xu X. MDU-net: multi-scale densely connected U-net for biomedical image segmentation. Health Inf Sci Syst. (2023) 11:13. doi: 10.1007/s13755-022-00204-9, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Isensee F, Jaeger PF, Kohl SAA, Petersen J, Maier-Hein KH. nnU-net: a self-configuring method for deep learning-based biomedical image segmentation. Nat Methods. (2021) 18:203–11. doi: 10.1038/s41592-020-01008-z, [DOI] [PubMed] [Google Scholar]
  • 11.de Boer M, Kos TM, Fick T, van Doormaal J, Colombo E, Kuijf HJ, et al. NnU-net versus mesh growing algorithm as a tool for the robust and timely segmentation of neurosurgical 3D images in contrast-enhanced T1 MRI scans. Acta Neurochir. (2024) 166:92. doi: 10.1007/s00701-024-05973-8, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Rott R, Ritter DJ, Ladstätter S, Nikolić O, Hennecke ME. LiMOX-A point cloud Lidar model toolbox based on NVIDIA OptiX ray tracing engine. Sensors (Basel). (2024) 24:1846. doi: 10.3390/s24061846, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Mitsouras D, Liacouras PC, Wake N, Rybicki FJ. RadioGraphics update: medical 3D printing for the radiologist. Radiographics. (2020) 40:E21–3. doi: 10.1148/rg.2020190217, [DOI] [PubMed] [Google Scholar]
  • 14.Zein NN, Hanouneh IA, Bishop PD, Samaan M, Eghtesad B, Quintini C, et al. Three-dimensional print of a liver for preoperative planning in living donor liver transplantation. Liver Transpl. (2013) 19:1304–10. doi: 10.1002/lt.23729, [DOI] [PubMed] [Google Scholar]
  • 15.Yang L, Shang XW, Fan JN, He ZX, Wang JJ, Liu M, et al. Application of 3D printing in the surgical planning of Trimalleolar fracture and doctor-patient communication. Biomed Res Int. (2016) 2016:2482086. doi: 10.1155/2016/2482086, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Tang H, Xu Z, Qin X, Wu B, Wu L, Zhao X, et al. Chest wall reconstruction in a canine model using polydioxanone mesh, demineralized bone matrix and bone marrow stromal cells. Biomaterials. (2009) 30:3224–33. doi: 10.1016/j.biomaterials.2009.02.007, [DOI] [PubMed] [Google Scholar]
  • 17.Sándor GK, Numminen J, Wolff J, Thesleff T, Miettinen A, Tuovinen VJ, et al. Adipose stem cells used to reconstruct 13 cases with cranio-maxillofacial hard-tissue defects. Stem Cells Transl Med. (2014) 3:530–40. doi: 10.5966/sctm.2013-0173, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Fortunato GM, Sigismondi S, Nicoletta M, Condino S, Montemurro N, Vozzi G, et al. Analysis of the robotic-based in situ bioprinting workflow for the regeneration of damaged tissues through a case study. Bioengineering (Basel). (2023) 10:560. doi: 10.3390/bioengineering10050560, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Ma Y, Ji Y, Zhong T, Wan W, Yang Q, Li A, et al. Bioprinting-based PDLSC-ECM screening for in vivo repair of alveolar bone defect using cell-laden, injectable and Photocrosslinkable hydrogels. ACS Biomater Sci Eng. (2017) 3:3534–45. doi: 10.1021/acsbiomaterials.7b00601, [DOI] [PubMed] [Google Scholar]
  • 20.Pratt P, Ives M, Lawton G, Simmons J, Radev N, Spyropoulou L, et al. Through the HoloLens looking glass: augmented reality for extremity reconstruction surgery using 3D vascular models with perforating vessels. Eur Radiol Exp. (2018) 2:2. doi: 10.1186/s41747-017-0033-2, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Lu L, Wang H, Liu P, Liu R, Zhang J, Xie Y, et al. Applications of mixed reality Technology in Orthopedics Surgery: A pilot study. Front Bioeng Biotechnol. (2022) 10:740507. doi: 10.3389/fbioe.2022.740507, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Pose-Díez-de-la-Lastra A, Moreta-Martinez R, García-Sevilla M, García-Mato D, Calvo-Haro JA, Mediavilla-Santos L, et al. HoloLens 1 vs. HoloLens 2: improvements in the new model for orthopedic oncological interventions. Sensors. (2022) 22:4915. doi: 10.3390/s22134915, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Condino S, Carbone M, Piazza R, Ferrari M, Ferrari V. Perceptual limits of optical see-through visors for augmented reality guidance of manual tasks. IEEE Trans Biomed Eng. (2020) 67:411–9. doi: 10.1109/TBME.2019.2914517, [DOI] [PubMed] [Google Scholar]
  • 24.Zari G, Condino S, Cutolo F, Ferrari V. Magic leap 1 versus Microsoft HoloLens 2 for the visualization of 3D content obtained from radiological images. Sensors. (2023) 23:3040. doi: 10.3390/s23063040, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Gavaghan KA, Peterhans M, Oliveira-Santos T, Weber S. A portable image overlay projection device for computer-aided open liver surgery. IEEE Trans Biomed Eng. (2011) 58:1855–64. doi: 10.1109/TBME.2011.2126572, [DOI] [PubMed] [Google Scholar]
  • 26.Javaheri H, Ghamarnejad O, Bade R, Lukowicz P, Karolus J, Stavrou GA. Beyond the visible: preliminary evaluation of the first wearable augmented reality assistance system for pancreatic surgery. Int J Comput Assist Radiol Surg. (2025) 20:117–29. doi: 10.1007/s11548-024-03131-0, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Jourdes F, Valentin B, Allard J, Duriez C, Seeliger B. Visual haptic feedback for training of robotic suturing. Front Robot AI. (2022) 9:800232. doi: 10.3389/frobt.2022.800232, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Vohra S, Goyal N, Gupta S. Preoperative CT evaluation of potential donors in living donor liver transplantation. Indian J Radiol Imaging. (2014) 24:350–9. doi: 10.4103/0971-3026.143897, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Hori M, Suzuki K, Epstein ML, Baron RL. Computed tomography liver volumetry using 3-dimensional image data in living donor liver transplantation: effects of the slice thickness on the volume calculation. Liver Transpl. (2011) 17:1427–36. doi: 10.1002/lt.22419, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Schnitzbauer AA, Lang SA, Goessmann H, Nadalin S, Baumgart J, Farkas SA, et al. Right portal vein ligation combined with in situ splitting induces rapid left lateral liver lobe hypertrophy enabling 2-staged extended right hepatic resection in small-for-size settings. Ann Surg. (2012) 255:405–14. doi: 10.1097/SLA.0b013e31824856f5, [DOI] [PubMed] [Google Scholar]
  • 31.Wu T, Huang W, He B, Guo Y, Peng G, Li M, et al. Diagnostic accuracy of 3D imaging combined with intra-operative ultrasound in the prediction of post-hepatectomy liver failure. J Gastrointest Oncol. (2022) 13:1224–36. doi: 10.21037/jgo-22-282, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Zhao Y, Chen T, Wang H, Xue Q, Guo W, Ding G, et al. Influence of three-dimensional visual reconstruction technology combined with virtual surgical planning of CTA images on precise resection of liver cancer in hepatobiliary surgery. Comput Math Methods Med. (2022) 2022:4376654. doi: 10.1155/2022/4376654, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Guglielmi A, Ruzzenente A, Conci S, Valdegamberi A, Iacono C. How much remnant is enough in liver resection? Dig Surg. (2012) 29:6–17. doi: 10.1159/000335713, [DOI] [PubMed] [Google Scholar]
  • 34.Ribero D, Amisano M, Bertuzzo F, Langella S, Lo Tesoriere R, Ferrero A, et al. Measured versus estimated total liver volume to preoperatively assess the adequacy of the future liver remnant: which method should we use? Ann Surg. (2013) 258:801–7. doi: 10.1097/SLA.0000000000000213, [DOI] [PubMed] [Google Scholar]
  • 35.Lopez-Lopez V, Gomez-Perez B, de Vicente E, Jiménez-Galanes S, Mora-Oliver I, Sabater L, et al. Next-generation three-dimensional modelling software for personalized surgery decision-making in perihilar cholangiocarcinoma: multicentre study. Br J Surg. (2021) 108:e394–5. doi: 10.1093/bjs/znab320, [DOI] [PubMed] [Google Scholar]
  • 36.Larghi Laureiro Z, Novelli S, Lai Q, Mennini G, D’andrea V, Gaudenzi P, et al. There is a great future in plastics: personalized approach to the management of hilar cholangiocarcinoma using a 3-D-printed liver model. Dig Dis Sci. (2020) 65:2210–5. doi: 10.1007/s10620-020-06326-y, [DOI] [PubMed] [Google Scholar]
  • 37.Krauel L, Fenollosa F, Riaza L, Pérez M, Tarrado X, Morales A, et al. Use of 3D prototypes for complex surgical oncologic cases. World J Surg. (2016) 40:889–94. doi: 10.1007/s00268-015-3295-y, [DOI] [PubMed] [Google Scholar]
  • 38.Song C, Min JH, Jeong WK, Kim SH, Heo JS, Han IW, et al. Use of individualized 3D-printed models of pancreatic cancer to improve surgeons’ anatomic understanding and surgical planning. Eur Radiol. (2023) 33:7646–55. doi: 10.1007/s00330-023-09756-0, [DOI] [PubMed] [Google Scholar]
  • 39.Gaujoux S, Sauvanet A, Vullierme MP, Cortes A, Dokmak S, Sibert A, et al. Ischemic complications after pancreaticoduodenectomy: incidence, prevention, and management. Ann Surg. (2009) 249:111–7. doi: 10.1097/SLA.0b013e3181930249, [DOI] [PubMed] [Google Scholar]
  • 40.Dong H, Luo M, Ke S, Zhan J, Liu X, Li Z. Application of intraoperative navigation in 3D laparoscopic pancreaticoduodenectomy: a case report. Exp Ther Med. (2022) 24:452. doi: 10.3892/etm.2022.11380, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Miyamoto R, Oshiro Y, Sano N, Inagawa S, Ohkohchi N. Three-dimensional surgical simulation of the bile duct and vascular arrangement in pancreatoduodenectomy: A retrospective cohort study. Ann Med Surg (Lond). (2018) 36:17–22. doi: 10.1016/j.amsu.2018.09.043, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Klauss M, Schöbinger M, Wolf I, Werner J, Meinzer HP, Kauczor HU, et al. Value of three-dimensional reconstructions in pancreatic carcinoma using multidetector CT: initial results. World J Gastroenterol. (2009) 15:5827–32. doi: 10.3748/wjg.15.5827, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Torres IO, De Luccia N. A simulator for training in endovascular aneurysm repair: the use of three dimensional printers. Eur J Vasc Endovasc Surg. (2017) 54:247–53. doi: 10.1016/j.ejvs.2017.05.011, [DOI] [PubMed] [Google Scholar]
  • 44.Seymour NE, Gallagher AG, Roman SA, O'Brien MK, Bansal VK, Andersen DK, et al. Virtual reality training improves operating room performance: results of a randomized, double-blinded study. Ann Surg. (2002) 236:458–64. doi: 10.1097/00000658-200210000-00008, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Tergas AI, Sheth SB, Green IC, Giuntoli RL, 2nd, Winder AD, Fader AN. A pilot study of surgical training using a virtual robotic surgery simulator. JSLS. (2013) 17:219–26. doi: 10.4293/108680813X13654754535872, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Andreatta PB, Woodrum DT, Birkmeyer JD, Yellamanchilli RK, Doherty GM, Gauger PG, et al. Laparoscopic skills are improved with LapMentor training: results of a randomized, double-blinded study. Ann Surg. (2006) 243:854–63. doi: 10.1097/01.sla.0000219641.79092.e5, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Curtis NJ, Conti JA, Dalton R, Rockall TA, Allison AS, Ockrim JB, et al. 2D versus 3D laparoscopic total mesorectal excision: a developmental multicentre randomised controlled trial. Surg Endosc. (2019) 33:3370–83. doi: 10.1007/s00464-018-06630-9, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Solbakken AM, Sellevold S, Spasojevic M, Julsrud L, Emblemsvåg HL, Reims HM, et al. Navigation-assisted surgery for locally advanced primary and recurrent rectal Cancer. Ann Surg Oncol. (2023) 30:7602–11. doi: 10.1245/s10434-023-13964-9, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Hamabe A, Ito M. A three-dimensional pelvic model made with a three-dimensional printer: applications for laparoscopic surgery to treat rectal cancer. Tech Coloproctol. (2017) 21:383–7. doi: 10.1007/s10151-017-1622-z, [DOI] [PubMed] [Google Scholar]
  • 50.Huang Y, Mu GC, Qin XG, Chen ZB, Lin JL, Zeng YJ. Study of celiac artery variations and related surgical techniques in gastric cancer. World J Gastroenterol. (2015) 21:6944–51. doi: 10.3748/wjg.v21.i22.6944, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Liu J, Zhou H, Qin H, Ru H, Huang J, Liang S, et al. Comparative study of clinical efficacy using three-dimensional and two-dimensional laparoscopies in the treatment of distal gastric cancer. Onco Targets Ther. (2018) 11:301–6. doi: 10.2147/OTT.S153520, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Kanaji S, Suzuki S, Harada H, Nishi M, Yamamoto M, Matsuda T, et al. Comparison of two- and three-dimensional display for performance of laparoscopic total gastrectomy for gastric cancer. Langenbeck's Arch Surg. (2017) 402:493–500. doi: 10.1007/s00423-017-1574-9, [DOI] [PubMed] [Google Scholar]
  • 53.Kinoshita T, Shibasaki H, Enomoto N, Sahara Y, Sunagawa H, Nishida T. Laparoscopic splenic hilar lymph node dissection for proximal gastric cancer using integrated three-dimensional anatomic simulation software. Surg Endosc. (2016) 30:2613–9. doi: 10.1007/s00464-015-4511-4, [DOI] [PubMed] [Google Scholar]
  • 54.Kok END, Eppenga R, Kuhlmann KFD, Groen HC, van Veen R, van Dieren JM, et al. Accurate surgical navigation with real-time tumor tracking in cancer surgery. NPJ Precis Oncol. (2020) 4:8. doi: 10.1038/s41698-020-0115-0, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Iritani K, Teshima M, Shimoda H, Shinomiya H, Otsuki N, Nibu KI. Intraoperative quantitative assessment of parathyroid blood flow during total thyroidectomy using indocyanine green fluorescence imaging - surgical strategies for preserving the function of parathyroid glands. Laryngoscope Investig Otolaryngol. (2022) 7:1251–8. doi: 10.1002/lio2.868, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Moreno Llorente P, García Barrasa A, Francos Martínez JM, Alberich Prats M, Pascua SM. Intraoperative indocyanine green (ICG) angiography of the parathyroids glands in prediction of post-thyroidectomy hypocalcemia: diagnostic accuracy of the ICG score 2 versus the 4-ICG score. Cir Esp (Engl Ed). (2022) 100:274–80. doi: 10.1016/j.cireng.2021.06.012, [DOI] [PubMed] [Google Scholar]
  • 57.Song Z, Dong W, Yang D, Yang J, Wu J, Wang Y, et al. Application of 3D visualization Technology in Complex Abdominal Wall Defects. Int J Gen Med. (2021) 14:2449–57. doi: 10.2147/IJGM.S310170, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Morris LM, LeBlanc KA. Components separation technique utilizing an intraperitoneal biologic and an onlay lightweight polypropylene mesh: "a sandwich technique". Hernia. (2013) 17:45–51. doi: 10.1007/s10029-012-0949-7, [DOI] [PubMed] [Google Scholar]
  • 59.Su N, Cui K, Zhao J, Duan Y, Wu X, Zhang H, et al. Comparative study on three-dimensional versus two-dimensional imaging using a computer-assisted surgery system for preoperative planning in pediatric middle hepatic tumors. BMC Surg. (2024) 24:236. doi: 10.1186/s12893-024-02531-y, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Xiu WL, Liu J, Zhang JL, Wang JM, Wang XF, Wang FF, et al. Computer-assisted three-dimensional individualized extreme liver resection for hepatoblastoma in proximity to the major liver vasculature. World J Gastrointest Surg. (2024) 16:1066–77. doi: 10.4240/wjgs.v16.i4.1066, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Liu B, Cai J, Xu Y, Peng X, Zheng H, Huang K, et al. Three-dimensional magnetic resonance cholangiopancreatography for the diagnosis of biliary atresia in infants and neonates. PLoS One. (2014) 9:e88268. doi: 10.1371/journal.pone.0088268, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Geetha SG, Motwani R, Chandrupatla M, Raviteja P, Ariyanachi K. Transmogrifying anatomy learning by kaizen strategies and game-based education. Cureus. (2024) 16:e64073. doi: 10.7759/cureus.64073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Kuang W, Yang EJ, Truong R, Woo BKP. Bringing virtual reality to mainstream pediatric care. J Patient Cent Res Rev. (2024) 11:107–11. doi: 10.17294/2330-0698.2063, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Zhang J, Yan H, Xue W, Zheng J, Li X, Hao L, et al. 3D printing technology in open living donor nephrectomy. Chin Med J. (2022) 135:2140–1. doi: 10.1097/CM9.0000000000001996, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Raichurkar KK, Lochan R, Jacob M, Asthana S. The use of a 3D printing model in planning a donor hepatectomy for living donor liver transplantation: first in India. J Clin Exp Hepatol. (2021) 11:515–7. doi: 10.1016/j.jceh.2020.08.006, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Ma L, Liang H, Han B, Yang S, Zhang X, Liao H. Augmented reality navigation with ultrasound-assisted point cloud registration for percutaneous ablation of liver tumors. Int J Comput Assist Radiol Surg. (2022) 17:1543–52. doi: 10.1007/s11548-022-02671-7, [DOI] [PubMed] [Google Scholar]
  • 67.Zhang W, Zhu W, Yang J, Xiang N, Zeng N, Hu H, et al. Augmented reality navigation for stereoscopic laparoscopic anatomical hepatectomy of primary liver cancer: preliminary experience. Front Oncol. (2021) 11:663236. doi: 10.3389/fonc.2021.663236, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Noureldin YA, Stoica A, Kaneva P, Andonian S. Impact of training on three-dimensional versus two-dimensional laparoscopic systems on Acquisition of Laparoscopic Skills in novices: A prospective comparative pilot study. Biomed Res Int. (2016) 2016:4197693. doi: 10.1155/2016/4197693, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69.Yammine K, Violato C. A meta-analysis of the educational effectiveness of three-dimensional visualization technologies in teaching anatomy. Anat Sci Educ. (2015) 8:525–38. doi: 10.1002/ase.1510, [DOI] [PubMed] [Google Scholar]
  • 70.Salazar D, Thompson M, Rosen A, Zuniga J. Using 3D printing to improve student education of complex anatomy: a systematic review and meta-analysis. Med Sci Educ. (2022) 32:1209–18. doi: 10.1007/s40670-022-01595-w, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Ahlberg G, Enochsson L, Gallagher AG, Hedman L, Hogman C, McClusky DA, 3rd, et al. Proficiency-based virtual reality training significantly reduces the error rate for residents during their first 10 laparoscopic cholecystectomies. Am J Surg. (2007) 193:797–804. doi: 10.1016/j.amjsurg.2006.06.050, [DOI] [PubMed] [Google Scholar]
  • 72.Pettit RW, Marlatt BB, Corr SJ, Havelka J, Rana A. nnU-net deep learning method for segmenting parenchyma and determining liver volume from computed tomography images. Ann Surg Open. (2022) 3:e155. doi: 10.1097/as9.0000000000000155, [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Fukuhara K, Akashi A, Nakane S, Tomita E. Preoperative assessment of the pulmonary artery by three-dimensional computed tomography before video-assisted thoracic surgery lobectomy. Eur J Cardiothorac Surg. (2008) 34:875–7. doi: 10.1016/j.ejcts.2008.07.014, [DOI] [PubMed] [Google Scholar]
  • 74.Strotzer QD, Winther H, Utpatel K, Scheiter A, Fellner C, Doppler MC, et al. Application of A U-net for map-like segmentation and classification of discontinuous fibrosis distribution in Gd-EOB-DTPA-enhanced liver MRI. Diagnostics (Basel). (2022) 12:1938. doi: 10.3390/diagnostics12081938, [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Frontiers in Medicine are provided here courtesy of Frontiers Media SA

RESOURCES