分割
人工智能
计算机科学
深度学习
正电子发射断层摄影术
病变
图像分割
模式识别(心理学)
计算机视觉
放射科
医学
病理
作者
Ine Dirks,Marleen Keyaerts,Bart Neyns,Jef Vandemeulebroucke
摘要
Automated lesion segmentation is essential to provide fast, reproducible tumor load estimates. Though deep learning methods have achieved unprecedented results in this field, they are often difficult to interpret, hampering their potential integration in the clinic. An interpretable deep learning approach is proposed for segmenting melanoma lesions on whole-body fluorine-18 fluorodeoxyglucose ([18F]FDG) positron emission tomography (PET) / computed tomography (CT). This consists of an automated PET thresholding step to identify FDGavid regions, followed by a three-channel nnU-Net considering the binary mask in addition to the PET and CT images. This segmentation step differentiates healthy from malignant tissue and removes the restriction on lesion boundaries imposed by the thresholding. The proposed method, trained on 267 images and evaluated on two sets acquired at the same institute, achieved mean Dice similarity coefficients (DSC) of 0.779 and 0.638 with mean absolute volume differences of 15.2mL and 22.0 mL. The DSC proved significantly higher compared to a direct, two-channel nnU-Net considering only the PET and CT. The same was observed when retraining and testing on subsets of the public data of the autoPET challenge, containing melanoma, lung cancer and lymphoma patients. In addition, overall results proved superior to a previously proposed two-step approach, where a classification network categorized each component of increased tracer uptake as healthy or malignant. The proposed lesion segmentation method for whole-body [18F]FDG PET/CT incorporates prior thresholding information while allowing more flexibility in the lesion delineation than a pure thresholding approach and increased interpretability over a direct segmentation network.
科研通智能强力驱动
Strongly Powered by AbleSci AI