Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Pretrained deep models outperform GBDTs in Learning-To-Rank under label scarcity

Created by
  • Haebom

Author

Charlie Hou, Kiran Koshy Thekumparampil, Michael Shavlovsky, Giulia Fanti, Yesh Dattatreya, Sujay Sanghavi

Outline

This paper demonstrates that deep learning (DL) models underperform gradient boosting decision trees (GBDTs) on outlier data in existing research, pointing out that this phenomenon is limited to ideal settings. Considering the complexity of real-world scenarios, we demonstrate that DL models can outperform GBDTs in label-sparse tabular learning-ranking (LTR) problems. Specifically, tabular LTR applications, such as search and recommendation, often lack labels but are also rich in unlabeled data. This paper demonstrates that DL ranking models can leverage this unlabeled data through unsupervised pretraining. Extensive experiments on public and proprietary datasets demonstrate that pretrained DL ranking models consistently outperform GBDT ranking models on ranking metrics (up to a 38% improvement).

Takeaways, Limitations

Takeaways: We experimentally demonstrate that a pre-trained DL model outperforms a GBDT model on label-poor tabular LTR problems, expanding the practical applicability of DL models. We also present a method for effectively utilizing unlabeled data through unsupervised pre-training.
Limitations: The experimental results in this paper are limited to a specific data type (unlabeled tabular LTR data), requiring further research on generalizability. Detailed information on the proprietary dataset used may be lacking, requiring verification of the reproducibility of the results. Because this is a comparison between specific models, rather than a comparison of various DL models with GBDT models, broader model comparison studies may be necessary.
👍