CodeEnhance: A Codebook-Driven Approach for Low-Light Image Enhancement

📅 2024-04-08
🏛️ arXiv.org
📈 Citations: 4
Influential: 0
📄 PDF
🤖 AI Summary
Low-light image enhancement (LLIE) faces two key challenges: ambiguity in restoration due to diverse illumination degradations, and trade-offs between noise suppression and brightness enhancement that compromise texture fidelity and color accuracy. To address these, we propose the first codebook-driven LLIE paradigm, formulating enhancement as a semantic-guided mapping from low-light inputs to high-quality discrete codebook representations. Our approach introduces three core innovations: (1) a Semantic Embedding Module (SEM) for hierarchical fusion of high-level semantics and low-level features; (2) a Customizable Codebook Shift (CS) mechanism enabling user-preference adaptation; and (3) an Interactive Feature Transformation (IFT) module for fine-grained texture and chromatic reconstruction. Integrating discrete vector quantization (VQ) with dynamic codebook adaptation, our method achieves state-of-the-art performance on both synthetic and real-world benchmarks—yielding substantial PSNR/SSIM gains—and demonstrates superior robustness to non-uniform illumination, noise, and color bias, along with enhanced visual quality and color fidelity.

Technology Category

Application Category

📝 Abstract
Low-light image enhancement (LLIE) aims to improve low-illumination images. However, existing methods face two challenges: (1) uncertainty in restoration from diverse brightness degradations; (2) loss of texture and color information caused by noise suppression and light enhancement. In this paper, we propose a novel enhancement approach, CodeEnhance, by leveraging quantized priors and image refinement to address these challenges. In particular, we reframe LLIE as learning an image-to-code mapping from low-light images to discrete codebook, which has been learned from high-quality images. To enhance this process, a Semantic Embedding Module (SEM) is introduced to integrate semantic information with low-level features, and a Codebook Shift (CS) mechanism, designed to adapt the pre-learned codebook to better suit the distinct characteristics of our low-light dataset. Additionally, we present an Interactive Feature Transformation (IFT) module to refine texture and color information during image reconstruction, allowing for interactive enhancement based on user preferences. Extensive experiments on both real-world and synthetic benchmarks demonstrate that the incorporation of prior knowledge and controllable information transfer significantly enhances LLIE performance in terms of quality and fidelity. The proposed CodeEnhance exhibits superior robustness to various degradations, including uneven illumination, noise, and color distortion.
Problem

Research questions and friction points this paper is trying to address.

Addressing uncertainty in restoration from diverse brightness degradations
Preventing loss of texture and color information during enhancement
Adapting pre-learned codebook to low-light image characteristics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages quantized priors for image enhancement
Uses Semantic Embedding Module for feature integration
Introduces Interactive Feature Transformation for refinement
🔎 Similar Papers
No similar papers found.
X
Xu Wu
Computer Vision Institute, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China, Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen 518060, China, and Guangdong Key Laboratory of Intelligent Information Processing, Shenzhen University, Shenzhen 518060, China
Xianxu Hou
Xianxu Hou
Xi'an Jiaotong-Liverpool University
Deep LearningComputer Vision
Zhihui Lai
Zhihui Lai
Shenzhen University
J
Jie Zhou
National Engineering Laboratory for Big Data System Computing Technology, Shenzhen University, and SZU Branch, Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen, Guangdong 518060, China
Y
Ya-nan Zhang
Computer Vision Institute, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China, Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen 518060, China, and Guangdong Key Laboratory of Intelligent Information Processing, Shenzhen University, Shenzhen 518060, China
W
W. Pedrycz
Department of Electrical & Computer Engineering, University of Alberta, University of Alberta, Canada
Linlin Shen
Linlin Shen
Shenzhen University
Deep LearningComputer VisionFacial Analysis/RecognitionMedical Image Analysis