Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

FinBERT2: A Specialized Bidirectional Encoder for Bridging the Gap in Finance-Specific Deployment of Large Language Models

Created by
  • Haebom

Author

Xuan Xu, Fufang Wen, Beilin Chu, Zhibing Fu, Qinhong Lin, Jiaqi Liu, Binjie Fei, Yu Li, Linna Zhou, Zhongliang Yang

Outline

This paper points out the limitations of applying large-scale language models (LLMs) in the financial field, and proposes FinBERT2, a Chinese finance-specific BERT model, to solve this problem. Despite its high computational cost, LLM underperforms fine-tuned BERT in discriminative tasks such as financial sentiment analysis, relies heavily on the Retrieval Augmented Generation (RAG) method for providing special information in generative tasks, and is also deficient in other feature-based scenarios such as topic modeling. FinBERT2 is a bidirectional encoder model pre-trained on a high-quality financial-specific corpus of 32 billion tokens, and outperforms existing (Fin)BERT models and LLM in five financial classification tasks. In addition, Fin-Retrievers based on FinBERT2 outperform existing embedding models in financial retrieval tasks, and Fin-TopicModel enables excellent clustering and topic representation for financial titles. In conclusion, FinBERT2 suggests an effective way to utilize finance-specific models in the LLM era.

Takeaways, Limitations

Takeaways:
We demonstrate that FinBERT2, a finance-specific BERT model, overcomes the limitations of LLM and achieves superior performance in the financial field.
FinBERT2 provides improved performance over existing models in a variety of financial-related tasks, including classification, retrieval, and topic modeling.
We present a method to effectively utilize the BERT model in the financial field in the LLM era.
Limitations:
Since FinBERT2 is specialized for Chinese financial data, its generalizability to other languages or domains may be limited.
Since this paper presents results for a corpus and model of a certain size, generalizability to other scales of data or models requires further study.
Discussion of the performance improvement of the RAG method is limited. Further research may be needed to investigate the performance improvement when FinBERT2 is combined with RAG.
👍