Exploring Scaling Laws of CTR Model for Online Performance Improvement

📅 2025-08-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the performance plateau of CTR models and the challenge of balancing prediction accuracy with low-latency online serving, this paper proposes SUAN, a scalable CTR modeling framework, and its lightweight variant LightSUAN. Methodologically, it innovatively adapts large-model scaling laws to the CTR domain and introduces a Unified Action Behavior encoder (UAB), enabling, for the first time, joint modeling of sequential and non-sequential features. By stacking UAB modules, employing sparse self-attention, and leveraging parallel inference, SUAN demonstrates stable scalability across three orders of magnitude in parameter count. Furthermore, an online knowledge distillation mechanism is proposed, allowing LightSUAN to retain identical inference latency while approximating the performance of significantly larger SUAN variants. Online deployment results show a 2.81% lift in CTR and a 1.69% improvement in CPM.

Technology Category

Application Category

📝 Abstract
CTR models play a vital role in improving user experience and boosting business revenue in many online personalized services. However, current CTR models generally encounter bottlenecks in performance improvement. Inspired by the scaling law phenomenon of LLMs, we propose a new paradigm for improving CTR predictions: first, constructing a CTR model with accuracy scalable to the model grade and data size, and then distilling the knowledge implied in this model into its lightweight model that can serve online users. To put it into practice, we construct a CTR model named SUAN (Stacked Unified Attention Network). In SUAN, we propose the UAB as a behavior sequence encoder. A single UAB unifies the modeling of the sequential and non-sequential features and also measures the importance of each user behavior feature from multiple perspectives. Stacked UABs elevate the configuration to a high grade, paving the way for performance improvement. In order to benefit from the high performance of the high-grade SUAN and avoid the disadvantage of its long inference time, we modify the SUAN with sparse self-attention and parallel inference strategies to form LightSUAN, and then adopt online distillation to train the low-grade LightSUAN, taking a high-grade SUAN as a teacher. The distilled LightSUAN has superior performance but the same inference time as the LightSUAN, making it well-suited for online deployment. Experimental results show that SUAN performs exceptionally well and holds the scaling laws spanning three orders of magnitude in model grade and data size, and the distilled LightSUAN outperforms the SUAN configured with one grade higher. More importantly, the distilled LightSUAN has been integrated into an online service, increasing the CTR by 2.81% and CPM by 1.69% while keeping the average inference time acceptable. Our source code is available at https://github.com/laiweijiang/SUAN.
Problem

Research questions and friction points this paper is trying to address.

Improving CTR model performance beyond current bottlenecks
Scaling model accuracy with increased model grade and data
Distilling high-grade model knowledge into lightweight online deployment
Innovation

Methods, ideas, or system contributions that make the work stand out.

Stacked Unified Attention Network for scalable CTR modeling
User Attention Block unifying sequential and non-sequential features
Online distillation from high-grade to lightweight model deployment
🔎 Similar Papers
No similar papers found.
W
Weijiang Lai
Institute of Software, Chinese Academy of Sciences, University of Chinese Academy of Sciences, Beijing, China
Beihong Jin
Beihong Jin
Institute of Software, Chinese Academy of Sciences
Pervasive ComputingDistributed Computing
J
Jiongyan Zhang
Meituan, Beijing, China
Y
Yiyuan Zheng
Institute of Software, Chinese Academy of Sciences, University of Chinese Academy of Sciences, Beijing, China
Jian Dong
Jian Dong
Shopee
Computer VisionMachine Learning
Jia Cheng
Jia Cheng
Meituan, Beijing, China
Jun Lei
Jun Lei
Meituan, Beijing, China
X
Xingxing Wang
Meituan, Beijing, China