Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Pre-training Limited Memory Language Models with Internal and External Knowledge

Created by
  • Haebom

Author

Linxi Zhao, Sofian Zalouk, Christian K. Belardi, Justin Lovelace, Jin Peng Zhou, Ryan Thomas Noonan, Dongyoung Go, Kilian Q. Weinberger, Yoav Artzi, Jennifer J. Sun

Outline

Neural language models are black boxes, with linguistic patterns and factual knowledge distributed across numerous opaque parameters. This entangled encoding makes it difficult to reliably inspect, verify, or update specific facts. In this paper, we introduce Limited Memory Language Models (LMLM), which store factual knowledge externally in an external database rather than memorizing it during pretraining. Through a pretraining approach, the authors strategically mask externally retrieved factual values from the training loss, allowing the model to learn to perform target lookups rather than relying on model weights. Experimental results demonstrate that LMLM achieves competitive performance on standard benchmarks compared to much larger LLMs, while offering the advantage of an explicit, editable, and verifiable knowledge base.

Takeaways, Limitations

Takeaways:
LMLM utilizes external databases to explicitly manage factual knowledge, reducing the model's reliance on memorization and facilitating knowledge modification, verification, and management while maintaining performance.
LMLM can achieve competitive performance even with smaller models, thereby improving computational efficiency.
Limitations:
This paper does not directly address Limitations. (This may be inferred from the dependence on the quality and accessibility of external databases, and the need for further research on the interaction and integration between external databases and models.)
👍