论文标题
PA-SEG:使用上下文正则化和交叉知识蒸馏从3D医疗图像分割的点注释中学习
PA-Seg: Learning from Point Annotations for 3D Medical Image Segmentation using Contextual Regularization and Cross Knowledge Distillation
论文作者
论文摘要
3D医疗图像分割中卷积神经网络(CNN)的成功取决于大量的完全注释的3D卷,用于训练,这些训练是耗时且劳动量很大的训练。在本文中,我们建议在3D医学图像中只有7个点注释分段目标,并设计一个两个阶段弱监督的学习框架PA-SEG。在第一阶段,我们采用大地距离变换来扩展种子点以提供更多的监督信号。为了进一步处理培训期间未经注释的图像区域,我们提出了两种上下文正则化策略,即多视图条件随机场(MCRF)损失和方差最小化(VM)损失,第一个鼓励具有相似特征的像素以具有一致的标签,而第二个特征则可以最小化前置和背景的强度差异。在第二阶段,我们使用在第一阶段预先训练的模型获得的预测作为伪标签。为了克服伪标签中的噪音,我们引入了一种自我和交叉监测(SCM)策略,该策略将自我训练与跨知识蒸馏(CKD)结合在主要模型和辅助模型之间,该模型从彼此产生的软标签中学习。在公共数据集的前庭造型瘤(VS)细分和脑肿瘤细分(BRAT)上进行的实验表明,我们在第一阶段进行培训的模型优胜于现有的较大的弱势监督方法,并且使用SCM进行了SCM进行其他培训,该模型的性能与Brats DataSet的其他培训相关。
The success of Convolutional Neural Networks (CNNs) in 3D medical image segmentation relies on massive fully annotated 3D volumes for training that are time-consuming and labor-intensive to acquire. In this paper, we propose to annotate a segmentation target with only seven points in 3D medical images, and design a two-stage weakly supervised learning framework PA-Seg. In the first stage, we employ geodesic distance transform to expand the seed points to provide more supervision signal. To further deal with unannotated image regions during training, we propose two contextual regularization strategies, i.e., multi-view Conditional Random Field (mCRF) loss and Variance Minimization (VM) loss, where the first one encourages pixels with similar features to have consistent labels, and the second one minimizes the intensity variance for the segmented foreground and background, respectively. In the second stage, we use predictions obtained by the model pre-trained in the first stage as pseudo labels. To overcome noises in the pseudo labels, we introduce a Self and Cross Monitoring (SCM) strategy, which combines self-training with Cross Knowledge Distillation (CKD) between a primary model and an auxiliary model that learn from soft labels generated by each other. Experiments on public datasets for Vestibular Schwannoma (VS) segmentation and Brain Tumor Segmentation (BraTS) demonstrated that our model trained in the first stage outperformed existing state-of-the-art weakly supervised approaches by a large margin, and after using SCM for additional training, the model's performance was close to its fully supervised counterpart on the BraTS dataset.