Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Is In-Context Learning Learning?

Created by
  • Haebom

Author

Adrian de Wynter

Outline

This paper validates the claim that context-based learning (ICL) can learn new tasks. While ICL enables task solving by predicting the next token without additional training, this may be due to inference based on the model's prior knowledge and presented examples rather than explicit learning. The paper argues that while ICL can be mathematically considered learning, a complete characterization requires experimental research. Through a large-scale analysis, we analyze the effectiveness and limitations of ICL by removing or accounting for factors such as memorization, pretraining, distribution shift, prompt style, and phrasing. Our results show that while ICL is an effective learning paradigm, its ability to learn and generalize to new tasks is limited. As examples increase, accuracy is less affected by example distribution, model, prompt style, and linguistic features of the input. However, distribution sensitivity is observed, particularly for prompt styles like chain-of-thought, as it infers patterns from prompt regularities. The variation in accuracy across formally similar tasks suggests that the temporal encoding of autoregression is not a robust mechanism and that its generalization ability is limited.

Takeaways, Limitations

Takeaways: While ICL is an effective learning paradigm, it has been shown to have limited generalization to new tasks. While accuracy improves with increasing examples, it suffers from limitations due to its reliance on inferences about the regularity of prompts.
Limitations: ICL's autoregressive encoding is not a robust and universal mechanism, and it exhibits distributional sensitivity, particularly in prompting styles such as thought chaining. Given that accuracy varies significantly even across formally similar tasks, its generalizability across all tasks is limited. Further experimental research is needed to fully characterize ICL.
👍