Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Equivariant Multi-Modality Image Fusion
126
Zitationen
9
Autoren
2024
Jahr
Abstract
Multi-modality image fusion is a technique that combines information from different sensors or modalities, en-abling the fused image to retain complementary features from each modality, such as functional highlights and texture details. However, effective training of such fusion models is challenging due to the scarcity of ground truth fusion data. To tackle this issue, we propose the Equivariant Multi-Modality imAge fusion (EMMA) paradigm for end-to-end self-supervised learning. Our approach is rooted in the prior knowledge that natural imaging responses are equiv-ariant to certain transformations. Consequently, we introduce a novel training paradigm that encompasses a fusion module, a pseudo-sensing module, and an equivariant fusion module. These components enable the net training to follow the principles of the natural sensing-imaging process while satisfying the equivariant imaging prior. Extensive experiments confirm that EMMA yields high-quality fusion results for infraredvisible and medical images, concurrently facilitating downstream multi-modal segmentation and detection tasks. The code is available at https://github.com/Zhaozixiang1228/MMIF-EMMA.
Ähnliche Arbeiten
Image quality assessment: from error visibility to structural similarity
2004 · 54.278 Zit.
A theory for multiresolution signal decomposition: the wavelet representation
1989 · 20.843 Zit.
Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering
2007 · 8.993 Zit.
Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising
2017 · 8.498 Zit.
Orthonormal bases of compactly supported wavelets
1988 · 8.129 Zit.