Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

PruneCD: Contrasting Pruned Self Model to Improve Decoding Factuality

Created by
  • Haebom

Author

Byeongho Yu, Changhun Lee, Jungyu Jin, Eunhyeok Park

Outline

This paper proposes PruneCD, a novel contrastive decoding method, to mitigate the hallucination problem of large-scale language models (LLMs). We address the limitations of the existing contrastive learning method using early-terminating logits (DoLa) and propose PruneCD. DoLa suffers from the drawback that early-terminating logits, due to their small size and insufficient information, hinder effective contrastive learning. PruneCD constructs an amateur model through hierarchical pruning, generating more information-rich and aligned logits, thereby enhancing the efficiency of contrastive decoding. Experimental results demonstrate that PruneCD is an effective method for improving realism while minimizing inference overhead.

Takeaways, Limitations

Takeaways:
Presenting an effective and practical method to alleviate hallucination problems in LLM.
Generating more information-rich logits through hierarchical pruning
Improved realism with minimal inference overhead
Limitations:
Further research is needed on the generalization performance of the proposed method.
Experimental results for various LLM architectures and sizes are required.
Further research is needed to optimize pruning strategies.
👍