Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Foundations of LLM Knowledge Materialization: Termination, Reproducibility, Robustness

Created by
  • Haebom

Author

Luca Giordano, Simon Razniewski

Outline

This paper focuses on systematically measuring and organizing the knowledge encoding of large-scale language models (LLMs). Specifically, we address the shortcomings of research on converting LLM knowledge into a structured format using recursive extraction methods, such as the GPTKB methodology. We analyze the termination rate, reproducibility, and robustness to fluctuations of this extraction process. Using miniGPTKB (domain-specific, tractable subcrawl), we measure termination rates, reproducibility, and three categories of metrics (yield, lexical similarity, and semantic similarity). The study was conducted across four variants (seed, language, randomness, and model) and three representative domains (history, entertainment, and finance).

Takeaways, Limitations

Takeaways:
LLM knowledge implementation can reliably reveal core knowledge.
The termination rate is high, but varies by model.
High robustness to seed and temperature variations.
Limitations:
Reproducibility shows mixed results.
Robustness to language and model variations is low.
There are important limitations to the implementation of LLM knowledge.
👍