• Keine Ergebnisse gefunden

Proceedings of the OAGM Workshop 2018 DOI: 10.3217/978-3-85125-603-1-02 6

N/A
N/A
Protected

Academic year: 2022

Aktie "Proceedings of the OAGM Workshop 2018 DOI: 10.3217/978-3-85125-603-1-02 6"

Copied!
1
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)Proceedings of the OAGM Workshop 2018. DOI: 10.3217/978-3-85125-603-1-02. Multi-Label Whole Heart Segmentation Using Anatomical Label Configurations and CNNs Christian Payer1 , Darko Štern2 , Horst Bischof1 , Martin Urschler2. Dr af t. I. E XTENDED A BSTRACT The accurate analysis of the whole heart substructures, i.e., left and right ventricle, left and right atrium, myocardium, pulmonary artery and the aorta, is highly relevant for cardiovascular applications. Therefore, automatic segmentation of these substructures from CT or MRI volumes is an important topic in medical image analysis [4]. Challenges for segmenting the heart substructures are their large anatomical variability in shape among subjects, the potential indistinctive boundaries between substructures and, especially for MRI data, artifacts and intensity inhomogeneities resulting from the acquisition process. To objectively compare and analyze whole heart substructure segmentation approaches, efforts like the MICCAI 2017 Multi-Modality Whole Heart Segmentation (MM-WHS) challenge are necessary and important for potential future application of semi-automated and fully automatic methods in clinical practice. We participated in the MM-WHS challenge, where we proposed a deep learning framework for fully automatic multi-label segmentation [2]. Evaluated on the MM-WHS challenge test data, we rank first for CT and second for MRI with a whole heart segmentation Dice score of 90.8% and 87%, respectively, leading to an overall first ranking among all participants. Our proposed method [2] performs fully automatic multilabel whole heart segmentation with CNNs using volumetric kernels. Due to the extensive memory and runtime requirements of volumetric CNNs, we use a pipeline of two CNNs that first localizes the heart in lower resolution volumes to crop a standardized region around the heart, followed by obtaining the final segmentation in higher resolution within this region (see Fig. 1). The localization CNN based on the U-Net [3] performs landmark localization using heatmap regression [1] to predict the approximate center of the bounding box around all heart substructures. Then, we crop a region with fixed physical size around the predicted center, ensuring that the region encloses all segmentation labels. Within the cropped region, the multilabel segmentation CNN predicts the heart substructure labels of each voxel. For this task, we use an adaptation of the fully convolutional end-to-end trained SpatialConfigurationNet (SCN) from [1]. The main idea of the three-component *This work was supported by the Austrian Science Fund (FWF): P 28078N33. 1 Christian Payer and Horst Bischof are with the Institute of Computer Graphics and Vision, Graz University of Technology, Austria. christian.payer@icg.tugraz.at. 2 Darko Štern and Martin Urschler are with the Ludwig Boltzmann Institute for Clinical Forensic Imaging, Graz, Austria. martin.urschler@cfi.lbg.ac.at. 6. Fig. 1. Overview of the fully automatic two-step multi-label segmentation pipeline. The first CNN uses a low resolution volume as input to localize the center of the bounding box around all heart substructures. The second CNN crops a region around this center and performs the multi-label segmentation. The figure is adapted from [2].. SCN is to learn from relative positions among structures to focus on anatomically feasible configurations as seen in the training data. In the first component of the SCN, a U-Net-like architecture [3] generates the first intermediate label predictions, corresponding to a voxel-wise probability of all labels. Then, the second component models the spatial configuration among labels, by using consecutive convolution layers to transform these probabilities to positions of other labels, generating the second intermediate label predictions. Finally, the third component multiplies both intermediate predictions, which results in the combined label predictions. Note that only when trained in an end-to-end manner, this last multiplication ensures that the first and second network component perform as expected. Without any further postprocessing, choosing the maximum value among the label predictions for each voxel leads to the final multi-label segmentation. R EFERENCES. [1] C. Payer, D. Štern, H. Bischof, and M. Urschler, “Regressing Heatmaps for Multiple Landmark Localization Using CNNs,” in Proc. Med. Image Comput. Comput. Interv. Springer, 2016, pp. 230–238. [2] C. Payer, D. Štern, H. Bischof, and M. Urschler, “Multi-label Whole Heart Segmentation Using CNNs and Anatomical Label Configurations,” in Stat. Atlases Comput. Model. Hear. ACDC MMWHS Challenges. STACOM 2017. Springer, 2018, pp. 190–198. [3] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” in Proc. Med. Image Comput. Comput. Interv. Springer, 2015, pp. 234–241. [4] X. Zhuang and J. Shen, “Multi-scale patch and multi-modality atlases for whole heart segmentation of MRI,” Med. Image Anal., vol. 31, pp. 77–87, 2016..

(2)

Referenzen

ÄHNLICHE DOKUMENTE

Multivarite Manifold Modeling of Functional Connectivity in Developing Language Networks Ernst Schwartz1,2 , Karl-Heinz Nenning1 , Gregor Kasprian1 , Anna-Lisa Schuller2 ,

Prediction of Local Lesion Evolving Risk In this work we demonstrate the application of predicting future lesions and mark corresponding high risk locations, by incorporating

TABLE II: Segmentation performance comparison between training on real data, generated data, and mixed data, using either no additional data augmentation, or standard data

Starting from a small colonic polyp endoscopic image database, we increase the number of images by tracking the content shown in the images through the endoscopic videos and

In particular, we introduce a network using a global shape descriptor and a local descriptor that is directly applicable to a point cloud, and that keeps the robustness of the

As soon as the sensor generates a new frame from a new position, the formerly mapped surface elements are used to render a depth map in the current sensor position.. This

In the online phase of the inspection with VAE, the decision, if an image contains a defect or is similar to the trained valid image distribution, could be made by means of the

Using this trained network to compute the local image similarities, a cumulative match score of about 70% at a retrieval rate of 20% can be achieved for the toolmark images.. In