论文标题

在CT图像中使用自动编码器的3D医疗图像分割,使用自动编码器进行标记和未标记的数据

3D medical image segmentation with labeled and unlabeled data using autoencoders at the example of liver segmentation in CT images

论文作者

Sital, Cheryl, Brosch, Tom, Tio, Dominique, Raaijmakers, Alexander, Weese, Jürgen

论文摘要

具有卷积神经网络(CNN)的解剖结构的自动分割构成了医学图像分析中的大部分研究。大多数基于CNN的方法都依赖大量的标记数据进行适当的培训。标记的医疗数据通常很少,但是未标记的数据更广泛地可用。这需要超越传统监督学习的方法,并利用未标记的数据进行细分任务。这项工作研究了自动编码器提取的特征通过CNN改善分割的潜力。考虑了两种策略。首先,将经过预处理的自动编码器特征用作分割网络中卷积层的初始化。其次,通过输入重建的分割和特征提取的任务同时学习并优化了多任务学习。卷积自动编码器用于从未标记的数据中提取特征,并使用多尺度的全卷积CNN来执行CT图像中3D肝分割的目标任务。对于这两种策略,进行了实验,并使用了不同量的标记和未标记的培训数据进行实验。与从头开始的培训相比,提议的学习策略改善了$ 75 \%的实验$ $,并将骰子得分提高了$ 0.040 $和$ 0.024 $,而无标记的培训数据的比例分别为$ 32:1 $和$ 12.5:1 $。结果表明,两种训练策略都更有效,而没有标记的培训数据比例很高。

Automatic segmentation of anatomical structures with convolutional neural networks (CNNs) constitutes a large portion of research in medical image analysis. The majority of CNN-based methods rely on an abundance of labeled data for proper training. Labeled medical data is often scarce, but unlabeled data is more widely available. This necessitates approaches that go beyond traditional supervised learning and leverage unlabeled data for segmentation tasks. This work investigates the potential of autoencoder-extracted features to improve segmentation with a CNN. Two strategies were considered. First, transfer learning where pretrained autoencoder features were used as initialization for the convolutional layers in the segmentation network. Second, multi-task learning where the tasks of segmentation and feature extraction, by means of input reconstruction, were learned and optimized simultaneously. A convolutional autoencoder was used to extract features from unlabeled data and a multi-scale, fully convolutional CNN was used to perform the target task of 3D liver segmentation in CT images. For both strategies, experiments were conducted with varying amounts of labeled and unlabeled training data. The proposed learning strategies improved results in $75\%$ of the experiments compared to training from scratch and increased the dice score by up to $0.040$ and $0.024$ for a ratio of unlabeled to labeled training data of about $32 : 1$ and $12.5 : 1$, respectively. The results indicate that both training strategies are more effective with a large ratio of unlabeled to labeled training data.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源