论文标题
眼睛眼底图像的强大深度学习:桥接实际和合成数据以增强概括
Robust deep learning for eye fundus images: Bridging real and synthetic data for enhancing generalization
论文作者
论文摘要
评估医学图像的深度学习应用程序受到限制,因为数据集通常很小且不平衡。在文献中已经提出了合成数据的使用,但是既没有对不同方法的可靠比较也没有报道。我们的方法集成了视网膜图像质量评估模型和StyleGAN 2体系结构,以增强与年龄相关的黄斑变性(AMD)检测能力并提高可推广性。这项工作比较了十个不同的生成对抗网络(GAN)架构,以产生有或没有AMD的合成眼镜图像。我们合并了三个公共数据库(Ichallenge-AMD,ODIR-2019和RIADD)的子集,以形成一个单个培训和测试集。我们采用了凝视数据集进行外部验证,以确保对拟议方法的全面评估。结果表明,stylegan2达到了最低的特征距离(166.17),而临床医生无法准确区分真实图像和合成图像。 RESNET-18体系结构以85%的精度获得了最佳性能,并且在检测AMD眼底图像方面优于两个人类专家(80%和75%)。测试集的精度率为82.8%,凝视数据集为81.3%,证明了该模型的可推广性。拟议的合成医学图像生成方法已通过鲁棒性和准确性进行了验证,并免费访问其代码以在该领域进行进一步的研究和开发。
Deep learning applications for assessing medical images are limited because the datasets are often small and imbalanced. The use of synthetic data has been proposed in the literature, but neither a robust comparison of the different methods nor generalizability has been reported. Our approach integrates a retinal image quality assessment model and StyleGAN2 architecture to enhance Age-related Macular Degeneration (AMD) detection capabilities and improve generalizability. This work compares ten different Generative Adversarial Network (GAN) architectures to generate synthetic eye-fundus images with and without AMD. We combined subsets of three public databases (iChallenge-AMD, ODIR-2019, and RIADD) to form a single training and test set. We employed the STARE dataset for external validation, ensuring a comprehensive assessment of the proposed approach. The results show that StyleGAN2 reached the lowest Frechet Inception Distance (166.17), and clinicians could not accurately differentiate between real and synthetic images. ResNet-18 architecture obtained the best performance with 85% accuracy and outperformed the two human experts (80% and 75%) in detecting AMD fundus images. The accuracy rates were 82.8% for the test set and 81.3% for the STARE dataset, demonstrating the model's generalizability. The proposed methodology for synthetic medical image generation has been validated for robustness and accuracy, with free access to its code for further research and development in this field.