[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Fine-Tune an SLM or Prompt an LLM? The Case of Generating Low-Code Workflows

Created by
  • Haebom

Author

Orlando Marquez Ayala, Patrice Bechard, Emily Chen, Maggie Baird, Jingfei Chen

Outline

In this paper, we present evidence that SLMs still have a quality advantage for certain domain tasks that require structured output, despite the emergence of large-scale language models (LLMs) such as GPT-4 making the benefits (faster inference, lower cost) of fine-tuning small-scale language models (SLMs) for real-world applications less clear. By comparing SLM fine-tuning with LLM prompting on a low-code JSON workflow generation task, we find that good prompting can produce reasonable results, but fine-tuning only improves quality by an average of 10%. In addition, we reveal the limitations of the model through systematic error analysis.

Takeaways, Limitations

Takeaways: We experimentally demonstrate that SLM fine-tuning produces higher quality results than LLM prompting for specific domain tasks that require structured output. Despite the reduced token cost, we confirm the practical value of SLM fine-tuning.
Limitations: This study is limited to a specific domain (low-code workflow generation) and a specific task (JSON format output). Generalizability to other domains or tasks requires further research. Systematic error analysis revealed the limitations of the model, but did not suggest specific measures to overcome these limitations.
👍