Skip to main content

High-resolution 3D abdominal segmentation with random patch network fusion

Posted by on Tuesday, December 14, 2021 in Abdomen Imaging, Deep Learning, Image Processing.

Y. Tang,R.Gao,S.Han, Y.Chen, D.Gao, V.Nath, C.Bermudez, M.R. Savona, R.G. Abramson, S.Bao,I.Lyu, Y.Huo and B.A. Landman,“High-resolution 3D Abdominal Segmentation with Random PatchNetworkFusion”,Medical Image Analysis, 2021.

Full Text:

https://www.sciencedirect.com/science/article/pii/S1361841520302589

Abstract

Deep learning for three dimensional (3D) abdominal organ segmentation on high-resolution computed to- mography (CT) is a challenging topic, in part due to the limited memory provide by graphics processing units (GPU) and large number of parameters and in 3D fully convolutional networks (FCN). Two preva- lent strategies, lower resolution with wider field of view and higher resolution with limited field of view, have been explored but have been presented with varying degrees of success. In this paper, we propose a novel patch-based network with random spatial initialization and statistical fusion on overlapping regions of interest (ROIs). We evaluate the proposed approach using three datasets consisting of 260 subjects with varying numbers of manual labels. Compared with the canonical “coarse-to-fine” baseline methods, the proposed method increases the performance on multi-organ segmentation from 0.799 to 0.856 in terms of mean DSC score (p-value < 0.01 with paired t -test). The effect of different numbers of patches is eval- uated by increasing the depth of coverage (expected number of patches evaluated per voxel). In addition, our method outperforms other state-of-the-art methods in abdominal organ segmentation. In conclusion, the approach provides a memory-conservative framework to enable 3D segmentation on high-resolution CT. The approach is compatible with many base network structures, without substantially increasing the complexity during inference. Given a CT scan with at high resolution, a low-res section (left panel) is trained with multi-channel seg- mentation. The low-res part contains down-sampling and normalization in order to preserve the complete spatial information. Interpolation and random patch sampling (mid panel) is employed to collect patches. The high-dimensional probability maps are acquired (right panel) from integration of all patches on field of views.

Method framework. Given a CT scan with at high resolution of ~0.8 x ~0.8 x ~2 mm, a low-res section (left panel) is trained with multi-channel segmentation. The low-res part contains down-sampling and normalization in order to preserve the complete spatial information. After the coarse segmentations are acquired from low- res UNet, we interpolate the mask to match the image’s original resolution. Next, random patch sampling (mid panel) is employed to collect patches, and patches are concatenated with corresponding coarse segmentation masks. Finally, we trained a patch-based high-res (right panel) segmentation model, the high-dimensional probability maps are acquired from integration of all patches on field of views. Majority vote is used to merge estimates into a final segmentation.
Method framework. Given a CT scan with at high resolution of ~0.8 x ~0.8 x ~2 mm, a low-res section (left panel) is trained with multi-channel segmentation. The low-res part contains down-sampling and normalization in order to preserve the complete spatial information. After the coarse segmentations are acquired from low- res UNet, we interpolate the mask to match the image’s original resolution. Next, random patch sampling (mid panel) is employed to collect patches, and patches are concatenated with corresponding coarse segmentation masks. Finally, we trained a patch-based high-res (right panel) segmentation model, the high-dimensional probability maps are acquired from integration of all patches on field of views. Majority vote is used to merge estimates into a final segmentation.

Tags: , ,