Towards Efficient Privacy-Preserving Machine Learning: A Systematic Review from Protocol, Model, and System Perspectives

📅 2025-07-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Privacy-preserving machine learning (PPML) suffers from severe efficiency bottlenecks and poor scalability due to cryptographic protocols (e.g., MPC, HE). To address this, we propose a cross-layer co-optimization paradigm integrating protocol design, model architecture, and system implementation. Our approach combines lightweight cryptographic protocols, model sparsification and quantization, hardware acceleration (FPGA/GPU), and adaptive scheduling. Through systematic qualitative and quantitative analysis, we demonstrate that inter-layer coupling optimization reduces computational overhead by orders of magnitude. We establish an open-source knowledge repository to track state-of-the-art advances, synthesize mainstream technical pathways, and characterize the fundamental trade-offs among efficiency, accuracy, and security. Furthermore, we identify promising future directions—including heterogeneous trusted execution environment integration and dynamic protocol selection—to advance practical, cloud-based privacy-preserving computation. This work provides both theoretical foundations and engineering guidelines for scalable PPML deployment.

Technology Category

Application Category

📝 Abstract
Privacy-preserving machine learning (PPML) based on cryptographic protocols has emerged as a promising paradigm to protect user data privacy in cloud-based machine learning services. While it achieves formal privacy protection, PPML often incurs significant efficiency and scalability costs due to orders of magnitude overhead compared to the plaintext counterpart. Therefore, there has been a considerable focus on mitigating the efficiency gap for PPML. In this survey, we provide a comprehensive and systematic review of recent PPML studies with a focus on cross-level optimizations. Specifically, we categorize existing papers into protocol level, model level, and system level, and review progress at each level. We also provide qualitative and quantitative comparisons of existing works with technical insights, based on which we discuss future research directions and highlight the necessity of integrating optimizations across protocol, model, and system levels. We hope this survey can provide an overarching understanding of existing approaches and potentially inspire future breakthroughs in the PPML field. As the field is evolving fast, we also provide a public GitHub repository to continuously track the developments, which is available at https://github.com/PKU-SEC-Lab/Awesome-PPML-Papers.
Problem

Research questions and friction points this paper is trying to address.

Efficiency and scalability issues in privacy-preserving machine learning
Cross-level optimizations for PPML protocols, models, and systems
Comprehensive review and future directions for PPML advancements
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cryptographic protocols for privacy protection
Cross-level optimizations for efficiency
Public GitHub repository for updates
🔎 Similar Papers
No similar papers found.
Wenxuan Zeng
Wenxuan Zeng
Peking University
Efficient Deep LearningLarge Language Model
T
Tianshi Xu
Peking University, China
Y
Yi Chen
Peking University, China
Y
Yifan Zhou
Peking University, China
Mingzhe Zhang
Mingzhe Zhang
Computing System Lab@Ant Research
Computer Architecture
Jin Tan
Jin Tan
Principal Engineer, National Renewable Energy Laboratory
power systems stability and operationrenewables integrations
C
Cheng Hong
Ant Group, China
M
Meng Li
Peking University, China