๐ค AI Summary
To address the degradation in ranking performance caused by inconsistent feature scaling between training and online inference in Learning-to-Rank (LTR), this paper proposes a mathematically rigorous deep-and-wide joint modeling paradigm that guarantees scale invariance. It is the first LTR method to achieve end-to-end, normalization-free theoretical scale invariance, enabling automatic adaptation to arbitrary feature rescalings throughout both training and inferenceโthereby eliminating reliance on standardization or batch normalization. The approach integrates co-designed architecture, explicit scale-invariance constraints, and a perturbation-robustness evaluation framework. Under simulated scale-mismatch scenarios, it significantly outperforms baselines: Ranking Loss decreases by 12.7%, NDCG@10 improves by 8.3%, with zero additional inference latency.
๐ Abstract
At Expedia, learning-to-rank (LTR) models plays a key role on our website in sorting and presenting information more relevant to users, such as search filters, property rooms, amenities, and images. A major challenge in deploying these models is ensuring consistent feature scaling between training and production data, as discrepancies can lead to unreliable rankings when deployed. Normalization techniques like feature standardization and batch normalization could address these issues but are impractical in production due to latency impacts and the difficulty of distributed real-time inference. To address consistent feature scaling issue, we introduce a scale-invariant LTR framework which combines a deep and a wide neural network to mathematically guarantee scale-invariance in the model at both training and prediction time. We evaluate our framework in simulated real-world scenarios with injected feature scale issues by perturbing the test set at prediction time, and show that even with inconsistent train-test scaling, using framework achieves better performance than without.