Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Pragmatic Inference Chain (PIC) Improving LLMs' Reasoning of Authentic Implicit Toxic Language

Created by
  • Haebom

Author

Xi Chen, Shuo Wang

Outline

This paper addresses the ethical challenges of large-scale language models (LLMs) and raises new possibilities for developing toxic language detection technologies. While previous studies have used data based on simple semantic associations (e.g., biased associations between "he" and "programmer" and "she" and "housewife"), this study collects real-world toxic interaction data, which avoids online censorship and has been identified by human evaluators as requiring inference. Drawing on this data, we propose a novel prompting method, Pragmatic Inference Chain (PIC), leveraging research in cognitive science and linguistics. We demonstrate that PIC prompting significantly improves the success rate of identifying implicit toxic language compared to existing prompting methods (e.g., CoT, rule-based), in models such as GPT-4o, Llama-3.1-70B-Instruct, DeepSeek-v2.5, and DeepSeek-v3, and produces clearer and more consistent inference processes. This suggests that our method could generalize to other inference-intensive tasks, such as humor and metaphor comprehension.

Takeaways, Limitations

Takeaways:
A novel prompting method (PIC) is presented that is effective in detecting suggestive and sophisticated toxic language.
Improving toxic language detection performance by enhancing LLM's inference capabilities.
Suggesting the generalizability of PIC prompting to other inference-intensive tasks (e.g., humor, metaphor comprehension)
Building and Utilizing a Dataset of Toxic Interactions in Real-World Online Environments
Limitations:
The generalizability of the proposed PIC prompting should be verified through further research.
Further discussion is needed regarding the scope and representativeness of the dataset used.
A more in-depth comparative analysis of the performance of PIC prompting for various LLM models is needed.
👍