Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

LaDiR: Latent Diffusion Enhances LLMs for Text Reasoning

Created by
  • Haebom

Author

Haoqiang Kang, Yizhe Zhang, Nikki Lijing Kuang, Nicklas Majamaki, Navdeep Jaitly, Yi-An Ma, Lianhui Qin

Outline

This paper proposes LaDiR (Latent Diffusion Reasoner), a novel inference framework that combines the expressive power of continuous latent representations with the iterative refinement of the latent diffusion model to enhance the inference performance of large-scale language models (LLMs). LaDiR constructs a structured latent inference space using a Variational Autoencoder (VAE), which preserves semantic information while providing a concise representation for text inference. The latent diffusion model is then utilized to denoise blocks of latent thought tokens, and a block-by-block bidirectional attention mask enables iterative refinement through a long-term horizon and adaptive runtime computation. LaDiR presents a novel paradigm that improves accuracy, diversity, and interpretability over existing autoregressive, diffusion-based, and latent inference methods on mathematical inference and planning benchmarks.

Takeaways, Limitations

Takeaways:
A novel framework for improving the reasoning ability of LLM (LaDiR).
Building a structured latent inference space using VAE.
Iterative refinement and parallel inference using latent diffusion models.
Outperforms existing methodologies on mathematical reasoning and planning benchmarks.
Increased diversity and interpretability of the reasoning process.
Limitations:
Limitations presented in the paper is not directly mentioned.
👍