Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Associative Knowledge Graphs for Efficient Sequence Storage and Retrieval

Created by
  • Haebom

Author

Przemys{\l}aw Stok{\l}osa, Janusz A. Starzyk, Pawe{\l} Raif, Adrian Horzyk, Marcin Kowalik

Outline

This paper presents a novel method utilizing Sequential Structure-Associative Knowledge Graphs (SSAKGs) to address the challenges of sequence storage and retrieval in applications such as anomaly detection, behavior prediction, and genetic information analysis. SSAKGs encode sequences as trended tournaments, with nodes representing objects and edges defining their order. We develop four ordering algorithms (Simple Sort, Node Ordering, Enhanced Node Ordering, and Weighted Edges Node Ordering) and evaluate their performance using synthetic and real-world datasets (sentence sequences from the NLTK library and miRNA sequences). Precision, sensitivity, and specificity are used as evaluation metrics. SSAKGs exhibit a memory capacity that grows quadratically with graph size, require no learning, are flexible in context-based reconstruction, and offer high efficiency in sparse memory graphs. This provides a scalable solution for sequence-based memory operations with broad applications in computational neuroscience and bioinformatics.

Takeaways, Limitations

Takeaways:
Presenting a new, efficient and scalable method for sequence storage and retrieval.
Context-based sequence reconstruction without the need for learning
Improving memory efficiency by leveraging sparse memory graphs.
Applicable to various fields such as computational neuroscience and bioinformatics
Limitations:
Memory capacity increases quadratically with graph size.
The scope of the dataset used may be limited (further experiments with datasets of more diverse types and sizes are needed).
The performance of the algorithm may vary depending on the characteristics of the dataset (further research is needed on algorithm optimization and generalization).
👍