论文标题

乳腺组织显微镜图像的分类:基于区域重复的自我划分与货架深度表示

Classification of Microscopy Images of Breast Tissue: Region Duplication based Self-Supervision vs. Off-the Shelf Deep Representations

论文作者

Ravi, Aravind

论文摘要

乳腺癌是世界上女性死亡率的主要原因之一。在进展的早期阶段进行诊断时,可以减少这一点。此外,通过计算机辅助诊断可以显着提高该过程的效率。基于深度学习的方法已成功应用于实现这一目标。以监督方式训练深网的限制因素之一是对大量专家注释数据的依赖性。实际上,大量未标记的数据,只有少量的专家注释数据可用。在这种情况下,可以利用转移学习方法和基于自学的学习(SSL)方法。在这项研究中,我们提出了一项新颖的自我诉讼借口任务,以训练卷积神经网络(CNN)并提取特定的特定特征。将该方法与使用预训练的CNN(例如Densenet-121和Resnet-50)在Imagenet上训练的深度特征进行了比较。此外,引入了两种类型的补丁组合方法,并将其与多数投票进行了比较。该方法在BACH显微镜图像数据集上进行了验证。结果表明,对于使用RESNET50与斑块级嵌入的串联提取的深度特征,达到了99%灵敏度的最佳性能。 SSL提取域特异性特征的初步结果表明,对于仅15%的未标记数据,对于显微镜图像的四类分类,可以实现94%的高灵敏度。

Breast cancer is one of the leading causes of female mortality in the world. This can be reduced when diagnoses are performed at the early stages of progression. Further, the efficiency of the process can be significantly improved with computer aided diagnosis. Deep learning based approaches have been successfully applied to achieve this. One of the limiting factors for training deep networks in a supervised manner is the dependency on large amounts of expert annotated data. In reality, large amounts of unlabelled data and only small amounts of expert annotated data are available. In such scenarios, transfer learning approaches and self-supervised learning (SSL) based approaches can be leveraged. In this study, we propose a novel self-supervision pretext task to train a convolutional neural network (CNN) and extract domain specific features. This method was compared with deep features extracted using pre-trained CNNs such as DenseNet-121 and ResNet-50 trained on ImageNet. Additionally, two types of patch-combination methods were introduced and compared with majority voting. The methods were validated on the BACH microscopy images dataset. Results indicated that the best performance of 99% sensitivity was achieved for the deep features extracted using ResNet50 with concatenation of patch-level embedding. Preliminary results of SSL to extract domain specific features indicated that with just 15% of unlabelled data a high sensitivity of 94% can be achieved for a four class classification of microscopy images.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源