Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

ToxiFrench: Benchmarking and Enhancing Language Models via CoT Fine-Tuning for French Toxicity Detection

Created by
  • Haebom

Author

Axel Delaval, Shujian Yang, Haicheng Wang, Han Qiu, Jialiang Lu

Outline

This paper presents TOXIFRENCH, a new public benchmark to address the challenges of detecting toxic content in French. TOXIFRENCH consists of 53,622 French online comments and was generated through a semi-automated annotation pipeline that reduces manual labeling to 10% through LLM-based pre-annotation and human validation. Benchmarking various models revealed that small language models (SLMs) outperform large models in terms of robustness and generalization on toxicity detection tasks. Building on this, we propose a novel Chain-of-Thought (CoT) fine-tuning strategy that utilizes a dynamic weighted loss that progressively emphasizes the model's final decision. The fine-tuned 4B model achieves state-of-the-art performance, outperforming LLMs such as GPT-40 and Gemini-2.5. Furthermore, TOXIFRENCH demonstrates robust multilingual capabilities on cross-lingual toxicity benchmarks, suggesting that our methodology can be effectively extended to other languages and safety-critical classification tasks.

Takeaways, Limitations

Takeaways:
We are releasing TOXIFRENCH, a large-scale dataset specialized for detecting toxic content in French, to contribute to future research.
We show that small language models can outperform large models in toxicity detection tasks.
We propose a new Chain-of-Thought (CoT) fine-tuning strategy and verify its effectiveness.
The developed model achieves state-of-the-art performance and demonstrates multilingual capabilities.
Limitations:
Further validation of the bias and generalizability of the TOXIFRENCH dataset is needed.
Further research is needed to determine the generalizability of the proposed CoT fine-tuning strategy and its applicability to other tasks.
Because the focus is on performance evaluation for specific languages and domains, further research is needed on generalizability to other languages and domains.
👍