Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Rethinking Table Instruction Tuning

Created by
  • Haebom

Author

Naihao Deng, Rada Mihalcea

Outline

This paper points out the lack of comprehensive evaluation of the influence of hyperparameter selection on the fine-tuning of large-scale language models (LLMs) for tabular understanding and their ability to generalize and generalize outside the domain. We evaluate existing tabular LLMs and find that their ability to generalize and understand outside the domain is significantly worse than the baseline models. We show that hyperparameters such as learning rate have a significant impact on both tabular-specific and general features, and unlike previous studies, we show that small learning rates and few training instances can improve tabular understanding while maintaining general features. Based on these results, we present a tabular LLM called TAMA, fine-tuned on LLaMA 3.1 8B Instruct, which achieves performance comparable to or better than GPT-3.5 and GPT-4 on tabular tasks while maintaining strong outside-domain generalization and general features. This demonstrates the potential of reducing data annotation costs and improving model development efficiency through careful hyperparameter selection. We open source our project and model in this paper.

Takeaways, Limitations

Takeaways:
We emphasize the importance of hyperparameters (especially learning rates) in fine-tuning LLM's instructions for table understanding.
We present an efficient fine-tuning method that improves table understanding ability and maintains general features using less data and small learning rate.
We present and open source a new high-performance tabular LLM called TAMA.
It suggests the potential to reduce data annotation costs and improve model development efficiency.
Limitations:
The types of LLMs evaluated in this study may be limited.
Generalization performance for specific types of tabular data requires further study.
Further validation is needed to determine whether TAMA's performance is superior across all types of tabular operations.
👍