Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

A Survey of Event Causality Identification: Taxonomy, Challenges, Assessment, and Prospects

Created by
  • Haebom

Author

Qing Cheng, Zefan Zeng, Xingchen Hu, Yuehang Si, Zhong Liu

Outline

This paper presents a comprehensive survey of event causality identification (ECI), an essential task in natural language processing (NLP) that automatically detects causal relationships between events in text. We define key concepts, formalize the ECI problem, and present a standard evaluation protocol. We develop a classification framework that categorizes ECI models into two main tasks: sentence-level event causality identification (SECI) and document-level event causality identification (DECI). For SECI, we review models that utilize feature pattern-based matching, machine learning classifiers, deep semantic encoding, prompt-based fine-tuning, and causal knowledge pretraining and data augmentation strategies. For DECI, we focus on approaches that leverage deep semantic encoding, event graph inference, and prompt-based fine-tuning. We pay particular attention to recent advances in multilingual and cross-lingual ECI and zero-shot ECI using large-scale language models (LLMs). We analyze the strengths, limitations, and challenges associated with each approach, and perform extensive quantitative evaluations on four benchmark datasets to rigorously evaluate the performance of various ECI models. Finally, we discuss future research directions and highlight opportunities to further advance this field.

Takeaways, Limitations

Takeaways:
The ECI model is systematically classified into SECI and DECI to enable a comprehensive understanding of the research status.
Provides a comprehensive review of various ECI models (feature pattern-based matching, machine learning, deep semantic encoding, prompt-based fine-tuning, causal knowledge pretraining, etc.).
Reflecting the latest research trends including multilingual and cross-lingual ECI, zero-shot ECI.
We present rigorous experimental evaluations using four benchmark datasets.
Contribute to the development of the ECI field by suggesting future research directions.
Limitations:
The types of ECI models covered in this paper may not be completely comprehensive.
As new ECI models and data sets are continuously emerging, there may be limitations in reflecting the latest research trends after the publication of a paper.
Difficulty in applying uniform evaluation criteria to all ECI models.
Possible poor generalization performance due to using datasets biased towards a specific language or domain.
👍