Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

SI-FACT: Mitigating Knowledge Conflict via Self-Improving Faithfulness-Aware Contrastive Tuning

Created by
  • Haebom

Author

Shengqiang Fu

Outline

This paper proposes a self-improving faith-aware contrastive tuning (SI FACT) framework to address the "knowledge conflict" problem, where large-scale language models (LLMs) rely on their internal parameter knowledge to ignore the context provided in knowledge-intensive tasks. SI FACT automatically generates high-quality, structured contrastive training data (anchor samples, semantically equivalent positive samples, and negative samples mimicking unfaithful scenarios) through a self-directed mechanism, significantly reducing the cost of manual annotation. The model is then trained by placing reliable responses close together and unfaithful responses farther apart through contrastive learning. Based on Llama3 8B Instruct, the SI FACT model improves contextual recall by 6.2% and significantly reduces internal memory dependency compared to the best baseline model on the ECARE KRE and COSE KRE benchmarks. This demonstrates that SI FACT is an effective, data-efficient, and practical method for building more proactive and reliable language models.

Takeaways, Limitations

Takeaways:
Presenting an effective self-improvement framework to address knowledge conflicts.
Achieving high data efficiency by reducing manual annotation costs.
Improved contextual reliability and reduced internal memory dependency of LLM.
Practical suggestions for more reliable and proactive LLM development
Limitations:
Further research is needed on the generalization performance of the proposed framework.
Applicability evaluation for various LLM architectures and sizes is needed.
Limitations of self-direction mechanisms and the need to explore ways to improve them.
Since the results are based on a specific benchmark, generalizability to other domains or tasks needs to be verified.
👍