Medical images often consist of multiple modalities, such as multimodal MRI images commonly used in diagnosing and studying brain tumors., and multimodal images provide rich complementary information. In the past, multimodal image segmentation usually directly added or connected modal features in the early or middle stage, which made it difficult to obtain the connection between modal features. In addition, there is a difference in information between modals and modals, and the previous method did not dealign modal features, which is likely to lead to reduced the effect of modal fusion. Thus, we propose a Multiscale dual dynamic feature fusion transformer (MdcFormer) model to explore the effects of multi-scale features, spatial and channel dynamic fusion and modal feature alignment on the segmentation effect of multimodal medical images. Utilizing a multi-encoder configuration and a single decoder, we gather characteristics from various modes at various levels and blend them in a dynamic manner across both spatial and channel domains. The proposed approach was evaluated using the BraTS2020 benchmark dataset. Empirical findings indicate that the model enhances the precision of segmentation.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.