Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Extracting Probabilistic Knowledge from Large Language Models for Bayesian Network Parameterization

Created by
  • Haebom

Author

Aliakbar Nafar, Kristen Brent Venable, Zijun Cui, Parisa Kordjamshidi

Outline

This paper evaluates the feasibility of building Bayesian networks (BNs) using large-scale language models (LLMs). While LLMs have demonstrated potential as factual knowledge bases, their ability to generate probabilistic knowledge about real-world events remains underexplored. This study explores how to leverage the probabilistic knowledge inherent in LLMs to derive probability estimates of statements about events and their relationships within BNs. LLMs enable parameterization of BNs, enabling probabilistic modeling within specific domains. Experiments on 80 publicly available BNs ranging from healthcare to finance demonstrate that querying LLMs for the conditional probabilities of events yields meaningful results compared to baselines, including random and uniform distributions and approaches based on the probability of the next token generation. Specifically, we explore how distributions extracted from LLMs can be used as expert priors to improve data-derived distributions, particularly when data is scarce. Overall, this study presents a promising strategy for automatically constructing Bayesian networks by combining probabilistic knowledge extracted from LLMs with real-world data. Furthermore, it establishes the first comprehensive baseline for evaluating the performance of LLMs in probabilistic knowledge extraction.

Takeaways, Limitations

Takeaways:
We present a novel method for effectively estimating the parameters of Bayesian networks by utilizing LLM.
We show that LLM can improve the accuracy of Bayesian networks by acting as expert prior information in situations where data is scarce.
We present a comprehensive baseline for evaluating the probabilistic knowledge extraction performance of LLM.
It provides an efficient method for automatically building Bayesian networks, increasing their applicability in various fields.
Limitations:
Further research is needed to determine the accuracy of the probability estimates generated by LLM.
The performance of LLM may be biased towards specific domains or datasets.
Further analysis is needed on the computational cost and efficiency of building Bayesian networks using LLM.
The generalizability of the 80 BN datasets used needs to be examined.
👍