Analysing zero-shot temporal relation extraction on clinical notes using temporal consistency

Analysing zero-shot temporal relation extraction on clinical notes using temporal consistency

Abstract

This paper presents the first study for temporal relation extraction in a zero-shot setting focusing on biomedical text. We employ two types of prompts and five Large Language Models (LLMs; GPT-3.5, Mixtral, Llama 2, Gemma, and PMC-LLaMA) to obtain responses about the temporal relations between two events. Our experiments demonstrate that LLMs struggle in the zero-shot setting, performing worse than fine-tuned specialized models in terms of F1 score. This highlights the challenging nature of this task and underscores the need for further research to enhance the performance of LLMs in this context. We further contribute a novel comprehensive temporal analysis by calculating consistency scores for each LLM. Our findings reveal that LLMs face challenges in providing responses consistent with the temporal properties of uniqueness and transitivity. Moreover, we study the relation between the temporal consistency of an LLM and its accuracy, and whether the latter can be improved by solving temporal inconsistencies. Our analysis shows that even when temporal consistency is achieved, the predictions can remain inaccurate.

Grafik Top
Authors
  • Kougia, Vasiliki
  • Sedova, Anastasiia
  • Stephan, Andreas
  • Zaporojets, Klim
  • Roth, Benjamin
Grafik Top
Shortfacts
Category
Paper in Conference Proceedings or in Workshop Proceedings (Paper)
Event Title
62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024)
Divisions
Data Mining and Machine Learning
Subjects
Kuenstliche Intelligenz
Event Location
Bangkok, Thailand
Event Type
Workshop
Event Dates
11-16 August 2024
Series Name
Proceedings of the 23rd Workshop on Biomedical Natural Language Processing
Page Range
pp. 72-84
Date
2024
Export
Grafik Top