Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Do Code Semantics Help? A Comprehensive Study on Execution Trace-Based Information for Code Large Language Models

Created by
  • Haebom

Author

Jian Wang, Xiaofei Xie, Qiang Hu, Shangqing Liu, Yi Li

Outline

This paper addresses the limitations of Code LLMs in their ability to infer runtime behavior and understand program functionality. Code LLMs suffer from a lack of inference capabilities regarding program execution behavior and the inconsistent and fragmented representation of semantic information, such as execution traces. To address these challenges, we present a general framework that integrates semantic information (e.g., execution traces) into code task-related prompts and comprehensively study the impact of semantic information on improving the inference performance of Code LLMs. Specifically, we investigate the impact of trace-based semantic information on the supervised fine-tuning (SFT) and inference stages of Code LLMs. Our experimental results demonstrate that, unlike previous studies, semantic information has limited utility in improving the test time of SFT and Code LLMs.

Takeaways, Limitations

Takeaways: A new framework for improving the reasoning ability of Code LLMs was presented, and an experimental study was conducted to examine the utility of semantic information. The results, which contradict previous research, prompted a rethinking of the use of semantic information.
Limitations: The results of this study may be limited to specific Code LLMs and datasets. Further research is needed on various Code LLMs and different types of code tasks. A more detailed analysis of the types of semantic information and how they are integrated is needed.
👍