Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

RedOne: Revealing Domain-specific LLM Post-Training in Social Networking Services

Created by
  • Haebom

Author

Fei Zhao, Chonggang Lu, Yue Wang, Zheyong Xie, Ziyan Liu, Haofu Qian, JianZhao Huang, Fangcheng Shi, Zijie Meng, Hongcheng Guo, Mingqian He, Wang, Shaosheng Cao

Outline

RedOne, a domain-specific large-scale language model (LLM), was developed to improve content management and interaction quality in social networking services (SNS). It was developed using a large, real-world dataset through a three-stage training strategy: continuous pretraining, supervised fine-tuning, and preference optimization. It achieved an average performance improvement of 14.02% on eight key social networking tasks and 7.56% on the social networking bilingual evaluation benchmark. It also reduced the exposure rate for harmful content detection by 11.23% and improved the click-through rate for post search by 14.95%.

Takeaways, Limitations

Takeaways:
Overcoming the performance bottlenecks of single-task-based models and establishing a comprehensive foundation for SNS.
Demonstrates excellent generalization ability across a variety of tasks.
Contributes to detecting harmful content and improving user experience.
Demonstrates applicability in real-world scenarios.
Limitations:
This paper does not specifically address Limitations (though it does mention that it overcomes the limitations of previous studies that focus on a single task).
👍