Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Ultra-Fast Language Generation via Discrete Diffusion Divergence Instruct

Created by
  • Haebom

Author

Haoyang Zheng, Xinyang Liu, Cindy Xiangrui Kong, Nan Jiang, Zheyuan Hu, Weijian Luo, Wei Deng, Guang Lin

Outline

This paper introduces DiDi-Instruct, a training-based method that initializes a pre-trained (masked) Discrete Diffusion Language Model (dLLM) and distills a student model through several stages to achieve fast, high-quality language generation, a key goal in the AI era. DiDi-Instruct achieves performance comparable to or superior to the dLLM teacher model and the GPT-2 base model, while achieving up to 64x speedup. The theoretical foundation of this method is a novel framework based on integral KL-divergence minimization, and grouped reward regularization, intermediate state matching, and a reward-based ancestor sampler are introduced to improve training stability, model coverage, and inference quality. On OpenWebText, DiDi-Instruct achieves confusion scores ranging from 62.2 (8 NFEs) to 18.4 (128 NFEs), outperforming existing accelerated dLLMs and the GPT-2 base model. This improvement is achieved with negligible entropy loss (approximately 1%) and a training time reduction of more than 20x compared to competing dLLM distillation methods. Furthermore, we validated the robustness and effectiveness of DiDi-Instruct through extensive elimination studies, model expansion, and discrete protein sequence generation.

Takeaways, Limitations

Takeaways:
Development of a fast and efficient language generation model
Accelerate generation speed while maintaining dLLM performance
Outperforms existing models and the GPT-2 base model on the OpenWebText dataset.
Improved training stability, model coverage, and inference quality
Software and model release planned (github.com/haoyangzheng-ai/didi-instruct)
Limitations:
Limitations is not explicitly mentioned in the paper.
👍