Infrared and Visible Image Fusion: From Data Compatibility to Task Adaption

📅 2025-01-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Infrared–visible image fusion (IVIF) has long suffered from a lack of systematic surveys and faces persistent bottlenecks in data compatibility, perceptual fidelity, and downstream task adaptability. To address these challenges, we propose the first unified analytical framework that jointly emphasizes data alignment capability and task-driven generalization. Our framework establishes a multidimensional evaluation protocol encompassing registration robustness, fusion quality (measured by EN, SSIM, and VIF), and performance on high-level vision tasks. We further present the first comprehensive taxonomy of deep learning–based IVIF methods, accompanied by a comparative methodology table. Additionally, we open-source IVIF_ZOO—the most extensive, reproducible benchmark codebase to date. Extensive experiments demonstrate that our framework significantly improves algorithm selection efficiency and cross-task generalization. This work delivers an authoritative survey, a standardized, reproducible benchmark, and novel research directions for the IVIF community.

Technology Category

Application Category

📝 Abstract
Infrared-visible image fusion (IVIF) is a critical task in computer vision, aimed at integrating the unique features of both infrared and visible spectra into a unified representation. Since 2018, the field has entered the deep learning era, with an increasing variety of approaches introducing a range of networks and loss functions to enhance visual performance. However, challenges such as data compatibility, perception accuracy, and efficiency remain. Unfortunately, there is a lack of recent comprehensive surveys that address this rapidly expanding domain. This paper fills that gap by providing a thorough survey covering a broad range of topics. We introduce a multi-dimensional framework to elucidate common learning-based IVIF methods, from visual enhancement strategies to data compatibility and task adaptability. We also present a detailed analysis of these approaches, accompanied by a lookup table clarifying their core ideas. Furthermore, we summarize performance comparisons, both quantitatively and qualitatively, focusing on registration, fusion, and subsequent high-level tasks. Beyond technical analysis, we discuss potential future directions and open issues in this area. For further details, visit our GitHub repository: https://github.com/RollingPlain/IVIF_ZOO.
Problem

Research questions and friction points this paper is trying to address.

Infrared-Visible Image Fusion
Image Quality
Task Flexibility
Innovation

Methods, ideas, or system contributions that make the work stand out.

Learning-based IVIF Methodology
Image Quality Enhancement
Task Adaptability and Fusion Strategies
🔎 Similar Papers
No similar papers found.
J
Jinyuan Liu
School of Software Technology, Dalian University of Technology, Dalian, 116024, China
G
Guanyao Wu
School of Software Technology, Dalian University of Technology, Dalian, 116024, China
Z
Zhu Liu
School of Software Technology, Dalian University of Technology, Dalian, 116024, China
D
Di Wang
School of Software Technology, Dalian University of Technology, Dalian, 116024, China
Zhiying Jiang
Zhiying Jiang
University of Waterloo
Natural Language ProcessingMachine Learning
Long Ma
Long Ma
Dalian University of Technology
Computer VisionImage Processing
Wei Zhong
Wei Zhong
Department of Statistics, Xiamen University
Statistics
X
Xin Fan
School of Software Technology, Dalian University of Technology, Dalian, 116024, China
Risheng Liu
Risheng Liu
Professor, Dalian University of Technology
computer visionmachine learningoptimization