Deep Unfolding with Kernel-based Quantization in MIMO Detection

📅 2025-05-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the accuracy degradation of deep-unfolding MIMO detection models (e.g., PGD-Nets, ADMM-Nets) under quantization on resource-constrained edge devices—caused by inaccurate activation distribution assumptions and static quantization step-size mismatches—this paper proposes a prior-free dynamic quantization method. First, kernel density estimation (KDE) and maximum mean discrepancy (MMD) are jointly employed to align inter-layer activation distributions, eliminating reliance on predefined assumptions (e.g., Gaussian or uniform). Second, a channel-adaptive dynamic step-size update mechanism is introduced, enabling fine-grained, input-aware trade-offs between detection accuracy and computational efficiency. Experiments demonstrate that the proposed method significantly improves quantized model accuracy while maintaining low inference latency, consistently outperforming conventional quantization-aware training (QAT) approaches across diverse channel conditions and bit-widths.

Technology Category

Application Category

📝 Abstract
The development of edge computing places critical demands on energy-efficient model deployment for multiple-input multiple-output (MIMO) detection tasks. Deploying deep unfolding models such as PGD-Nets and ADMM-Nets into resource-constrained edge devices using quantization methods is challenging. Existing quantization methods based on quantization aware training (QAT) suffer from performance degradation due to their reliance on parametric distribution assumption of activations and static quantization step sizes. To address these challenges, this paper proposes a novel kernel-based adaptive quantization (KAQ) framework for deep unfolding networks. By utilizing a joint kernel density estimation (KDE) and maximum mean discrepancy (MMD) approach to align activation distributions between full-precision and quantized models, the need for prior distribution assumptions is eliminated. Additionally, a dynamic step size updating method is introduced to adjust the quantization step size based on the channel conditions of wireless networks. Extensive simulations demonstrate that the accuracy of proposed KAQ framework outperforms traditional methods and successfully reduces the model's inference latency.
Problem

Research questions and friction points this paper is trying to address.

Energy-efficient MIMO detection for edge computing deployment
Overcoming performance loss in quantization for deep unfolding models
Dynamic quantization step adjustment for varying wireless channel conditions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Kernel-based adaptive quantization for deep unfolding
Joint KDE and MMD aligns activation distributions
Dynamic step size updates based on channel conditions
🔎 Similar Papers
2024-01-15arXiv.orgCitations: 1
Zeyi Ren
Zeyi Ren
MPhil, The University of Hong Kong
Model-driven Deep LearningWireless CommunicationsAutonomous Driving
Jingreng Lei
Jingreng Lei
The University of Hong Kong
AI for Wireless Commun.OptimizationDistributed AlgorithmsRobotics Perception
Y
Yichen Jin
Department of Electrical and Electronic Engineering, The University of Hong Kong, HKSAR, China
Ermo Hua
Ermo Hua
Tsinghua University
Physics-driven Foundation Model
Qingfeng Lin
Qingfeng Lin
Postdoctoral Scholar, University of Southern California
Opticselectronicsenergynanofabricationnanodevices
C
Chen Zhang
Department of Electrical and Electronic Engineering, The University of Hong Kong, HKSAR, China
B
Bowen Zhou
Shanghai AI Laboratory
Y
Yik-Chung Wu
Department of Electrical and Electronic Engineering, The University of Hong Kong, HKSAR, China