Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Fast and Fluent Diffusion Language Models via Convolutional Decoding and Rejective Fine-tuning

Created by
  • Haebom

Author

Yeongbin Seo, Dongha Lee, Jaehyung Kim, Jinyoung Yeo

Outline

To address the inference speed limitations of autoregressive (AR) language models, we propose a diffusion-based language model capable of parallel decoding of multiple tokens. To address the long decoding window problem (lack of relevance and repetition of tokens located far from the input context), a key issue with existing diffusion language models, we propose convolutional decoding (Conv), a boundary-free regularization-based method, to improve fluency and flexibility. Furthermore, we introduce Rejecting Rule-based Fine-Tuning (R2FT) to better align tokens located far from the context. The proposed methods outperform existing diffusion language models on open generative benchmarks, improving both speed and quality.

Takeaways, Limitations

Takeaways:
Improves generation quality and speed by solving the long decoding window problem of diffusion models.
Improving the performance of diffusion language models with Conv and R2FT.
Achieves superior performance compared to existing diffusion language models in open generative benchmarks.
Limitations:
The specific Limitations of this paper is not stated in the Abstract.
👍