Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

AdaMCoT: Rethinking Cross-Lingual Factual Reasoning through Adaptive Multilingual Chain-of-Thought

Created by
  • Haebom

Author

Weihua Zheng, Xin Huang, Zhengyuan Liu, Tarun Kumar Vangani, Bowei Zou, Xiyan Tao, Yuhao Wu, Ai Ti Aw, Nancy F. Chen, Roy Ka-Wei Lee

Outline

This paper presents the Adaptive Multilingual Chain-of-Thought (AdaMCOT) framework for enhancing factual inference performance of multilingual large-scale language models (LLMs). To address the scalability issues and difficulties in capturing subtle inference processes associated with existing multilingual dictionary training and cross-language tuning approaches, AdaMCOT dynamically routes thought processes from an intermediate "thought language" to generate target language responses. It selects the optimal inference path through an adaptive reward-based mechanism without additional dictionary training. Through various benchmark evaluations, we demonstrate that it significantly improves factual inference quality and cross-language consistency, particularly in low-resource language environments. Analysis of the model's hidden state and semantic space elucidates the underlying mechanisms of this approach, suggesting that the adaptive inference path effectively bridges the performance gap between high- and low-resource languages while preserving cultural and linguistic nuances.

Takeaways, Limitations

Takeaways:
Effective in improving factual reasoning performance of low-resource languages.
Improved cross-language consistency.
Adaptive inference path selection without additional pre-training.
Maintaining cultural and linguistic subtleties.
Limitations:
Further analysis is needed to determine the specific factors contributing to the performance improvement of AdaMCOT.
Generalization performance verification is needed for various languages and tasks.
The need for greater transparency and interpretability of the “thought language” selection mechanism.
👍