DCN^2: Interplay of Implicit Collision Weights and Explicit Cross Layers for Large-Scale Recommendation

📅 2025-06-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address three critical limitations of DCNv2 in large-scale recommendation systems—(i) information attenuation in deep cross layers, (ii) coarse-grained implicit handling of feature hash collisions, and (iii) absence of explicit second-order feature interaction modeling—this paper proposes Triple-DCN. Our method introduces: (1) an enhanced Cross layer to mitigate gradient and information vanishing in deep cross operations; (2) learnable lookup-level embedding weights for fine-grained suppression of hash collision bias; and (3) an FFM-style explicit pairwise similarity modeling layer to capture highly discriminative second-order interactions. Triple-DCN achieves >500 million predictions per second in online inference. Extensive offline evaluations and production A/B tests consistently outperform DCNv2, establishing new state-of-the-art results on four major benchmarks—Criteo, Avazu, KuaiRec, and Alibaba-CCP.

Technology Category

Application Category

📝 Abstract
The Deep and Cross architecture (DCNv2) is a robust production baseline and is integral to numerous real-life recommender systems. Its inherent efficiency and ability to model interactions often result in models that are both simpler and highly competitive compared to more computationally demanding alternatives, such as Deep FFMs. In this work, we introduce three significant algorithmic improvements to the DCNv2 architecture, detailing their formulation and behavior at scale. The enhanced architecture we refer to as DCN^2 is actively used in a live recommender system, processing over 0.5 billion predictions per second across diverse use cases where it out-performed DCNv2, both offline and online (ab tests). These improvements effectively address key limitations observed in the DCNv2, including information loss in Cross layers, implicit management of collisions through learnable lookup-level weights, and explicit modeling of pairwise similarities with a custom layer that emulates FFMs' behavior. The superior performance of DCN^2 is also demonstrated on four publicly available benchmark data sets.
Problem

Research questions and friction points this paper is trying to address.

Address information loss in Cross layers
Manage collisions via learnable lookup weights
Model pairwise similarities like FFMs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Enhanced Cross layers to reduce information loss
Learnable lookup-level weights for collision management
Custom layer emulating FFMs for pairwise similarities
Blaž Škrlj
Blaž Škrlj
Jožef Stefan Institute
Machine Learning
Y
Yonatan Karni
Teads
G
Grega Gašperšič
Teads
B
Blaž Mramor
Teads
Y
Yulia Stolin
Teads
M
Martin Jakomin
Teads
J
Jasna Urbančič
Teads
Y
Yuval Dishi
Teads
Natalia Silberstein
Natalia Silberstein
Outbrain
coding theoryrecommender systemsmachine learning
Ophir Friedler
Ophir Friedler
Ph.D at Tel Aviv University
Algorithmic game theoryAlgorithmic mechanism design
A
Assaf Klein
Teads