Disentangle, Align and Fuse for Multimodal and Semi-Supervised Image Segmentation

dc.contributor.authorChartsias, Agisilaos
dc.contributor.authorPapanastasiou, Giorgos
dc.contributor.authorWang, Chengjia
dc.contributor.authorSemple, Scott
dc.contributor.authorNewby, David E.
dc.contributor.authorDharmakumar, Rohan
dc.contributor.authorTsaftaris, Sotirios A.
dc.contributor.departmentMedicine, School of Medicine
dc.date.accessioned2024-11-21T11:33:52Z
dc.date.available2024-11-21T11:33:52Z
dc.date.issued2021
dc.description.abstractMagnetic resonance (MR) protocols rely on several sequences to assess pathology and organ status properly. Despite advances in image analysis, we tend to treat each sequence, here termed modality, in isolation. Taking advantage of the common information shared between modalities (an organ's anatomy) is beneficial for multi-modality processing and learning. However, we must overcome inherent anatomical misregistrations and disparities in signal intensity across the modalities to obtain this benefit. We present a method that offers improved segmentation accuracy of the modality of interest (over a single input model), by learning to leverage information present in other modalities, even if few (semi-supervised) or no (unsupervised) annotations are available for this specific modality. Core to our method is learning a disentangled decomposition into anatomical and imaging factors. Shared anatomical factors from the different inputs are jointly processed and fused to extract more accurate segmentation masks. Image misregistrations are corrected with a Spatial Transformer Network, which non-linearly aligns the anatomical factors. The imaging factor captures signal intensity characteristics across different modality data and is used for image reconstruction, enabling semi-supervised learning. Temporal and slice pairing between inputs are learned dynamically. We demonstrate applications in Late Gadolinium Enhanced (LGE) and Blood Oxygenation Level Dependent (BOLD) cardiac segmentation, as well as in T2 abdominal segmentation. Code is available at https://github.com/vios-s/multimodal_segmentation.
dc.eprint.versionAuthor's manuscript
dc.identifier.citationChartsias A, Papanastasiou G, Wang C, et al. Disentangle, Align and Fuse for Multimodal and Semi-Supervised Image Segmentation. IEEE Transactions on Medical Imaging. 2021;40(3):781-792. doi:10.1109/TMI.2020.3036584
dc.identifier.urihttps://hdl.handle.net/1805/44642
dc.language.isoen_US
dc.publisherIEEE
dc.relation.isversionof10.1109/TMI.2020.3036584
dc.relation.journalIEEE Transactions on Medical Imaging
dc.rightsPublisher Policy
dc.sourceArXiv
dc.subjectAnnotations
dc.subjectBiomedical imaging
dc.subjectDecoding
dc.subjectDisentanglement
dc.subjectImage segmentation
dc.subjectMagnetic resonance imaging
dc.subjectMultimodal segmentation
dc.subjectSemantics
dc.subjectTraining
dc.titleDisentangle, Align and Fuse for Multimodal and Semi-Supervised Image Segmentation
dc.typeArticle
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Chartsias2021Disentangle-AAM.pdf
Size:
4.42 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.04 KB
Format:
Item-specific license agreed upon to submission
Description: