This page organizes papers related to artificial intelligence published around the world. This page is summarized using Google Gemini and is operated on a non-profit basis. The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.
RedOne, a domain-specific large-scale language model (LLM), was developed to improve content management and interaction quality in social networking services (SNS). It was developed using a large, real-world dataset through a three-stage training strategy: continuous pretraining, supervised fine-tuning, and preference optimization. It achieved an average performance improvement of 14.02% on eight key social networking tasks and 7.56% on the social networking bilingual evaluation benchmark. It also reduced the exposure rate for harmful content detection by 11.23% and improved the click-through rate for post search by 14.95%.
Takeaways, Limitations
•
Takeaways:
◦
Overcoming the performance bottlenecks of single-task-based models and establishing a comprehensive foundation for SNS.
◦
Demonstrates excellent generalization ability across a variety of tasks.
◦
Contributes to detecting harmful content and improving user experience.
◦
Demonstrates applicability in real-world scenarios.
•
Limitations:
◦
This paper does not specifically address Limitations (though it does mention that it overcomes the limitations of previous studies that focus on a single task).