KEYWORDS: Image segmentation, Education and training, Feature fusion, Deep learning, Medical imaging, Heart, Spleen, Data modeling, Parkinson disease, Fourier transforms
Deep models have been shown to tend to fit the target function from low to high frequencies (a phenomenon called the frequency principle of deep learning). One may hypothesize that such property can be leveraged for better training of deep learning models, in particular for segmentation tasks where annotated datasets are often small. In this paper, we exploit this property to propose a new training method based on frequency-domain disentanglement. It consists of three main stages. First, it disentangles the image into high- and low-frequency components. Then, the segmentation network model learns them separately (the approach is general and can use any segmentation network as backbone). Finally, feature fusion is performed to complete the downstream task. The method was applied to the segmentation of the red and dentate nuclei in Quantitative Susceptibility Mapping (QSM) data and to three tasks of the Medical Segmentation Decathlon (MSD) challenge under different training sample sizes. For segmenting the red and dentate nuclei and the heart, the proposed approach resulted in considerable improvements over the baseline (respectively between 8 and 16 points of Dice and between 5 and 8 points). On the other hand, there was no improvement for the spleen and the hippocampus. We believe that these intriguing results, which echo theoretical work on the frequency principle of deep learning, are of interest for discussion at the conference. The source code is publicly available at: https://github.com/GuanghuiFU/frequency_disentangled_learning.
When performing manual segmentations, experts heavily rely on prior anatomical knowledge. Topology is an important prior information due to its stability across patients. Recently, several losses based on persistent homology were proposed to constrain topology. However, such approaches are computationally expensive and complex to implement, in particular in 3D. In this paper, we propose a novel loss function to introduce topological priors in deep learning-based segmentation, which is fast to compute and easy to implement. Our approach was evaluated in several medical datasets (spleen, heart, hippocampus, red nucleus). It allowed reducing topological errors and, in some cases, improving voxel-level accuracy.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.