首页|面向机器人手术模拟器的神经辐射场软组织动态三维重建

面向机器人手术模拟器的神经辐射场软组织动态三维重建

扫码查看
构建了一种基于自监督的框架,该框架从单目立体内窥镜视频中提取多视图图像,利用图像中的底层三维(3D)信息构建对象的几何约束,实现软组织结构的准确重建。基于分割任意场景模型对内窥镜下的动态手术器械、静态腹腔场景及可形变软组织结构进行分割解耦。该框架利用简单的神经网络多层感知机来表示动态神经辐射场(NeRF)中运动手术器械和形变软组织结构,基于偏斜熵损失对手术场景中的手术器械、腔体场景和软组织结构进行正确分离。在通过使用单目立体内窥镜捕获机器人手术模拟器场景的数据集上,将所提方法的结果与其他方法进行定量定性比较。结果表明本文方法在处理腹腔体场景、软组织结构重建、手术器械的分割解耦,以及来自多视点的3D信息和运动对象的图像分割等方面显著优于当前的方法。
Dynamic Three-Dimensional Reconstruction of Soft Tissue in Neural Radiation Field for Robotic Surgery Simulators
Objective Reconstructing soft tissue structures based on the endoscope position with robotic surgery simulators plays an important role in robotic surgery simulator training.Traditional soft tissue structure reconstruction is mainly achieved through surface reconstruction algorithms using medical imaging data sets such as computed tomography and magnetic resonance imaging.These methods fail to reconstruct the color information of soft tissue models and are not suitable for complex surgical scenes.Therefore,we proposed a method based on neural radiation fields,combined it with classic volume rendering to segment robotic surgery simulator scenes from videos with deformable soft tissue captured by a monocular stereoscopic endoscope,and performed three-dimensional reconstruction of biological soft tissue structures to restore soft tissue.By using segmented arbitrary scene model(SASM)for segmentation modeling of time-varying objects and time-invariant objects in videos,specific dynamic occlusions in surgical scenes can be removed.Methods Inspired by recent advances in neural radiation fields,we first constructed a self-supervision-based framework that extracted multi-view images from monocular stereoscopic endoscopic videos and used the underlying 3D information in the images to construct geometric constraints of objects,so as to accurately reconstruct soft tissue structures.Then,the SASM was used to segment and decouple the dynamic surgical instruments,static abdominal scenes,and deformable soft tissue structures under the endoscope.In addition,this framework used a simple neural network multilayer perceptron(MLP)to represent moving surgical instruments and deformed soft tissue structures in dynamic neural radiation fields and proposed skew entropy loss to correctly predict surgical instruments,cavity scenes,and soft tissue structures in surgical scenes.Results and Discussions We employ MLP to represent robotic surgery simulator scenes in the neural radiation field to accommodate the inherent geometric complexity and deformable soft tissue.Furthermore,we establish a hybrid framework of the neural radiation field and SASM for efficient characterization and segmentation of endoscopic surgical scenes in an endoscopic robotic surgery simulator.To address the dynamic nature of scenes and facilitate accurate scene separation,we propose a self-supervised approach incorporating a novel loss function.For validation,we perform a comprehensive quantitative and qualitative evaluation of a data set captured using a stereoendoscope,including simulated robotic surgery scenes from different angles and distances.The results show that our method performs well in synthesizing realistic robotic surgery simulator scenes compared with existing methods,with an average improvement of 12.5%in peak signal-to-noise ratio(PSNR)and an average improvement of 8.43%in structural similarity(Table 1).It shows excellent results and performance in simulating scenes and achieving high-fidelity reconstruction of biological soft tissue structures,color,textures,and other details.Furthermore,our method shows significant efficacy in scene segmentation,enhancing overall scene understanding and accuracy.Conclusions We propose a novel NeRF-based framework for self-supervised 3D dynamic surgical scene decoupling and biological soft tissue reconstruction from arbitrary multi-viewpoint monocular stereoscopic endoscopic videos.Our method decouples dynamic surgical instrument occlusion and deformable soft tissue structures,recovers a static abdominal volume background representation,and enables high-quality new view synthesis.The key parts of our framework are the SASM and the neural radiation field.The highly segmentable module of SASM decomposes the surgical scene into dynamic,static,and deformable regions.A spatiotemporal hybrid representation is then designed to facilitate and efficiently model the decomposed neural radiation fields.Our method achieves excellent performance in various simulation scenes of robotic surgery data,such as large-scale moving surgical instruments and 3D reconstruction of deformable soft tissue structures.We believe that our method can facilitate robotic surgery simulator scene understanding and hope that emerging NeRF-based 3D reconstruction technology can provide inspiration for robotic surgery simulator scene understanding and empower various downstream clinically oriented tasks.

visual opticsneural radiation fieldthree-dimensional reconstruction of soft tissuesegmented arbitrary scene modelsegmentation decoupling

陈琪、秦芝宝、蔡晓誉、李世杰、王梓俊、石俊生、邰永航

展开 >

云南师范大学物理与电子信息学院,云南昆明 650500

云南省光电信息技术重点实验室,云南昆明 650500

视觉光学 神经辐射场 软组织三维重建 分割任意场景模型 分割解耦

国家自然科学基金国家自然科学基金国家自然科学基金云南省优秀青年基金

623650176206207062005235202301AW070001

2024

光学学报
中国光学学会 中国科学院上海光学精密机械研究所

光学学报

CSTPCD北大核心
影响因子:1.931
ISSN:0253-2239
年,卷(期):2024.44(7)
  • 5