Papers
Topics
Authors
Recent
Search
2000 character limit reached

Creating a Digital Twin of Spinal Surgery: A Proof of Concept

Published 25 Mar 2024 in cs.CV | (2403.16736v2)

Abstract: Surgery digitalization is the process of creating a virtual replica of real-world surgery, also referred to as a surgical digital twin (SDT). It has significant applications in various fields such as education and training, surgical planning, and automation of surgical tasks. In addition, SDTs are an ideal foundation for machine learning methods, enabling the automatic generation of training data. In this paper, we present a proof of concept (PoC) for surgery digitalization that is applied to an ex-vivo spinal surgery. The proposed digitalization focuses on the acquisition and modelling of the geometry and appearance of the entire surgical scene. We employ five RGB-D cameras for dynamic 3D reconstruction of the surgeon, a high-end camera for 3D reconstruction of the anatomy, an infrared stereo camera for surgical instrument tracking, and a laser scanner for 3D reconstruction of the operating room and data fusion. We justify the proposed methodology, discuss the challenges faced and further extensions of our prototype. While our PoC partially relies on manual data curation, its high quality and great potential motivate the development of automated methods for the creation of SDTs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. Fast and easy focus stacking. URL: https://github.com/PetteriAimonen/focus-stack.
  2. Robotic surgery with lean reinforcement learning. CoRR abs/2105.01006.
  3. Aracam: A rgb-d multi-view photogrammetry system for lower limb 3d reconstruction applications. Sensors 22, 2443.
  4. Toward a digital twin for arthroscopic knee surgery: A systematic review. IEEE Access 10, 45029–45052.
  5. Dense 3d reconstruction through lidar: A comparative study on ex-vivo porcine tissue. arXiv preprint arXiv:2401.10709 .
  6. Openpose: Realtime multi-person 2d pose estimation using part affinity fields. IEEE Transactions on Pattern Analysis and Machine Intelligence .
  7. Opera: Attention-regularized transformers for surgical phase recognition, in: Medical Image Computing and Computer Assisted Intervention–MICCAI 2021, Springer. pp. 604–614.
  8. Multimodal medical imaging fusion for patient specific musculoskeletal modeling of the lumbar spine system in functional posture. Journal of Medical and Biological Engineering 37, 739–749.
  9. Reinforcement learning in surgery. Surgery 170, 329–332.
  10. Hmd-egopose: Head-mounted display-based egocentric marker-less tool and hand pose estimation for augmented surgical guidance. International Journal of Computer Assisted Radiology and Surgery 17, 2253–2262.
  11. Pico lantern: Surface reconstruction and augmented reality in laparoscopic surgery using a pick-up laser projector. Medical image analysis 25, 95–102.
  12. Surgery 4.0: the natural culmination of the industrial revolution? Innovative Surgical Sciences 2, 105–108.
  13. Dynamic depth-supervised nerf for multi-view rgb-d operating room videos, in: International Workshop on PRedictive Intelligence In MEdicine, Springer. pp. 218–230.
  14. Digital twin: manufacturing excellence through virtual factory replication. White paper 1, 1–7.
  15. The advances in computer vision that are enabling more autonomous actions in surgery: A systematic review of the literature. Sensors 22, 4918.
  16. Chapter 38 - Surgical data science, in: Zhou, S.K., Rueckert, D., Fichtinger, G. (Eds.), Handbook of Medical Image Computing and Computer Assisted Intervention. Academic Press. The Elsevier and MICCAI Society Book Series, pp. 931–952. doi:https://doi.org/10.1016/B978-0-12-816176-0.00043-0.
  17. Next-generation surgical navigation: Marker-less multi-view 6dof pose estimation of surgical instruments. arXiv:2305.03535.
  18. Dynamic scene graph representation for surgical video, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 81--87.
  19. X23D - intraoperative 3d lumbar spine shape reconstruction based on sparse multi-view x-ray data. J. Imaging 8, 271.
  20. Articulated clinician detection using 3d pictorial structures on RGB-D data. Medical Image Anal. 35, 215--224.
  21. A multi-view rgb-d approach for human pose estimation in operating rooms, in: 2017 IEEE winter conference on applications of computer vision (WACV), IEEE. pp. 363--372.
  22. 3d gaussian splatting for real-time radiance field rendering. ACM Transactions on Graphics 42.
  23. Prototyping a digital twin for real time remote control over mobile networks: Application of remote surgery. Ieee Access 7, 20325--20336.
  24. Category-level articulated object pose estimation, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 3706--3715.
  25. Neuralangelo: High-fidelity neural surface reconstruction, in: CVPR.
  26. Automatic registration with continuous pose updates for marker-less surgical navigation in spine surgery. Medical Image Analysis 91, 103027.
  27. Sgt++: Improved scene graph-guided transformer for surgical report generation. IEEE Transactions on Medical Imaging .
  28. Endogaussian: Gaussian splatting for deformable surgical scene reconstruction. arXiv preprint arXiv:2401.12561 .
  29. The perioperative human digital twin. Anesthesia & Analgesia 134, 885--892.
  30. SMPL: A skinned multi-person linear model. ACM Transactions on Graphics, (Proc. SIGGRAPH Asia) 34, 248:1--248:16.
  31. Segment anything in medical images. Nature Communications 15, 654.
  32. Surgical data science--from concepts toward clinical translation. Medical image analysis 76, 102306.
  33. Surgical data science for next-generation interventions. Nature Biomedical Engineering 1, 691--696.
  34. Immersive Virtual Reality for Surgical Training: A Systematic Review. Journal of Surgical Research 268, 40--58.
  35. Computer vision in surgery: from potential to clinical value. npj Digit. Medicine 5.
  36. A structured light-based laparoscope with real-time organs’ surface reconstruction for minimally invasive surgery, in: 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, IEEE. pp. 5769--5772.
  37. Virtual reality in surgery. Bmj 323, 912--915.
  38. Resfields: Residual neural fields for spatiotemporal signals. arXiv preprint arXiv:2309.03160 .
  39. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM 65, 99--106.
  40. Surgical process modeling. Innovative surgical sciences 2, 123--137.
  41. Multimodal semantic scene graphs for holistic modeling of surgical procedures. arXiv preprint arXiv:2106.15309 .
  42. Defining digital surgery for the future. npj Digit. Medicine 5.
  43. Embodied hands: Modeling and capturing hands and bodies together. ACM Transactions on Graphics, (Proc. SIGGRAPH Asia) 36.
  44. Twin-s: a digital twin for skull base surgery. International Journal of Computer Assisted Radiology and Surgery , 1--8.
  45. Mvor: A multi-view rgb-d operating room dataset for 2d and 3d human pose estimation.
  46. Neus2: Fast learning of neural implicit surfaces for multi-view reconstruction, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3295--3306.
  47. 4d gaussian splatting for real-time dynamic scene rendering. arXiv preprint arXiv:2310.08528 .
  48. Surrol: An open-source reinforcement learning centered and dvrk compatible platform for surgical robot learning, in: 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE. pp. 1821--1828.
  49. Advancing surgical vqa with scene graph knowledge. arXiv preprint arXiv:2312.10251 .
  50. Synthesizing diverse human motions in 3d indoor scenes, in: ICCV, pp. 14692--14703.
  51. You only learn once: Universal anatomical landmark detection, in: Medical Image Computing and Computer Assisted Intervention--MICCAI 2021: 24th International Conference, Strasbourg, France, September 27--October 1, 2021, Proceedings, Part V 24, Springer. pp. 85--95.
  52. LABRAD-OR: Lightweight Memory Scene Graphs for Accurate Bimodal Reasoning in Dynamic Operating Rooms, in: Medical Image Computing and Computer Assisted Intervention – MICCAI 2023, Springer Nature Switzerland, Cham. pp. 302--311.
Citations (4)

Summary

  • The paper introduces a high-fidelity surgical digital twin for spinal surgery by integrating laser scanning and RGB-D imaging.
  • It utilizes advanced data acquisition and fusion methods to capture both static OR features and dynamic elements like surgeon movements.
  • Quantitative results indicate sub-millimeter accuracy and near-perfect alignment with real camera feeds, highlighting potential for enhanced surgical training and automation.

Creating a Digital Twin of Spinal Surgery: A Proof of Concept

Introduction

The paper explores the development of a surgical digital twin (SDT) focusing on the digitalization of spinal surgery procedures, specifically pedicle screw placement, performed in a controlled laboratory setting. The SDT serves as a comprehensive virtual replica of the operating room (OR), providing detailed geometric and semantic data crucial for visualization, analysis, and potential robotic or ML-driven task automation.

Methodology

Data Acquisition and Fusion

The methodology includes capturing static and dynamic elements within the OR using a range of technologies. Laser scanners generate a foundational 3D point cloud to represent static environment features, such as the OR infrastructure, lighting, and fixed equipment. Figure 1

Figure 1: Generation of the reference point cloud from multiple laser scans, illustrating the successful fusion from individual scans with minimized occlusion challenges.

Dynamic and Static Element Modelling

The paper employs multiple imaging technologies to accommodate the dynamic nature of surgery— RGB-D cameras were used for capturing body poses and instrument movements. Additionally, dedicated photogrammetry was applied to construct highly accurate models of the surgical tables, tools, and anatomical components before integrating these into a unified virtual environment.

Human Pose and Instrument Tracking

Human poses are estimated using a blend of RGB-D input data and machine learning algorithms utilizing the SMPL-H model alongside OpenPose. This setup allows for precise, marker-less tracking of the surgeon’s movements throughout the procedure.

Results Evaluation

Quantitative Assessment

The system demonstrated high geometric accuracy in both static and dynamic scenarios, with laser scan fusion achieving sub-millimeter errors and RGB-D camera calibration errors averaging below 1 pixel, confirming the precision of the SDT’s spatial data amassment.

Qualitative Outputs

Rendering results compared the SDT against real-world camera feeds, with near-perfect alignment evidenced in Figures 4 and 5, underscoring the fidelity of the digital models. Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2: Comparisons between the digital twin renders and real camera images, reflecting perspectives from various Kinect cameras.

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3: Renderings depicting the reconstructed operating room and estimated body poses, exhibiting the prototype's rich visual details.

Discussion

Technical Challenges and Limitations

The primary challenge arises from reflective and dynamic surfaces within the OR, which impose significant complexities for optical data acquisition systems. Moreover, while the paper makes substantial strides in static and dynamic tracking, integration with real-time intraoperative imaging or real-time anatomy deformation remains unexplored due to existing technological limitations in reconstructing dynamic, deformable anatomy during actual surgical procedures.

Future Directions

The successful realization of an SDT opens avenues for broader digital twin applications. This involves potential advancements in surgical training, real-time task automation, and simulation-based preoperative planning. There remains a clear pathway towards enhanced integration of sensor modalities, inclusion of more sophisticated AI-driven predictive analytics, and automation techniques to further refine and extend the current SDT model.

Conclusion

The research demonstrates the feasibility and practicality of developing a digital twin for surgeries by utilizing a combination of high-fidelity image capturing, dynamic modeling, and integrated data processing. The SDT establishes a platform for future explorations into educational and operational enhancements in surgical contexts. Such advancements will drive the evolution of precision surgery, provide robust platforms for robotics and AI applications in medical fields, and offer substantial improvements to surgical outcomes and preemptive problem-solving.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.