论文标题
脑肿瘤分割有不完整的成像数据
Brain tumour segmentation with incomplete imaging data
论文作者
论文摘要
越来越多地认识到,只有从常规临床护理中得出的完全包容的大规模收集,脑部肿瘤的复杂异质性越来越多地要求大小和丰富性数据。这是当代机器学习可以促进的一项任务,尤其是在神经影像方面,但是它处理在现实世界中临床实践中常见数据的能力仍然未知。在这里,我们将最新方法应用于大规模的多站点MRI数据,以量化自动化肿瘤分割模型的比较保真度,以复制临床现实中观察到的各种序列可用性。我们将深度学习(NNU-NET衍生的)分割模型与T1,对比增强的T1,T2和Flair序列的所有可能组合进行了比较,对2021 BRATS-RSNA胶质瘤人群的1251例患者进行了五倍的交叉验证,对1251名患者进行了培训和验证,并在实际上进行了进一步的测试,不仅对现实的患者进行了进一步的测试。术后成像也是如此。在不完整的成像数据进行培训的模型中,经常等效地等效于对完整数据培训的模型,该系数的骰子系数为0.907(单个序列)至0.945(完整数据集),用于整个肿瘤,以及0.701(单个序列)(单个序列)至0.891(完整的数据集(完整数据集),用于组件组织类型。在没有对比度成像的情况下,不完整的数据分割模型可以准确地检测到增强肿瘤,从而用R2在0.95-0.97之间量化其体积,并且对于病变形态计量学是不变的。当缺少数据时,深度学习分割模型可以很好地表征肿瘤,甚至可以在不使用对比度的情况下检测增强的组织。这表明将数据不完整的临床实践转换为临床实践,可能比迄今为止认为的要容易得多,并且在减少对对比度使用的依赖性方面可能具有价值。
The complex heterogeneity of brain tumours is increasingly recognized to demand data of magnitudes and richness only fully-inclusive, large-scale collections drawn from routine clinical care could plausibly offer. This is a task contemporary machine learning could facilitate, especially in neuroimaging, but its ability to deal with incomplete data common in real world clinical practice remains unknown. Here we apply state-of-the-art methods to large scale, multi-site MRI data to quantify the comparative fidelity of automated tumour segmentation models replicating the various levels of sequence availability observed in the clinical reality. We compare deep learning (nnU-Net-derived) segmentation models with all possible combinations of T1, contrast-enhanced T1, T2, and FLAIR sequences, trained and validated with five-fold cross-validation on the 2021 BraTS-RSNA glioma population of 1251 patients, with further testing on a real-world 50 patient sample diverse in not only MRI scanner and field strength, but a random selection of pre- and post-operative imaging also. Models trained on incomplete imaging data segmented lesions well, often equivalently to those trained on complete data, exhibiting Dice coefficients of 0.907 (single sequence) to 0.945 (full datasets) for whole tumours, and 0.701 (single sequence) to 0.891 (full datasets) for component tissue types. Incomplete data segmentation models could accurately detect enhancing tumour in the absence of contrast imaging, quantifying its volume with an R2 between 0.95-0.97, and were invariant to lesion morphometry. Deep learning segmentation models characterize tumours well when missing data and can even detect enhancing tissue without the use of contrast. This suggests translation to clinical practice, where incomplete data is common, may be easier than hitherto believed, and may be of value in reducing dependence on contrast use.