Analysing zero-shot temporal relation extraction on clinical notes using temporal consistency
This paper presents the first study for temporal relation extraction in a zero-shot setting focusing on biomedical text. We employ two types of prompts and five Large Language Models (LLMs; GPT-3.5, Mixtral, Llama 2, Gemma, and PMC-LLaMA) to obtain responses about the temporal relations between two events. Our experiments demonstrate that LLMs struggle in the zero-shot setting, performing worse than fine-tuned specialized models in terms of F1 score. This highlights the challenging nature of this task and underscores the need for further research to enhance the performance of LLMs in this context. We further contribute a novel comprehensive temporal analysis by calculating consistency scores for each LLM. Our findings reveal that LLMs face challenges in providing responses consistent with the temporal properties of uniqueness and transitivity. Moreover, we study the relation between the temporal consistency of an LLM and its accuracy, and whether the latter can be improved by solving temporal inconsistencies. Our analysis shows that even when temporal consistency is achieved, the predictions can remain inaccurate.
Top- Kougia, Vasiliki
- Sedova, Anastasiia
- Stephan, Andreas
- Zaporojets, Klim
- Roth, Benjamin
Category |
Paper in Conference Proceedings or in Workshop Proceedings (Paper) |
Event Title |
62nd Annual Meeting of the Association for Computational Linguistics (ACL 2024) |
Divisions |
Data Mining and Machine Learning |
Subjects |
Kuenstliche Intelligenz |
Event Location |
Bangkok, Thailand |
Event Type |
Workshop |
Event Dates |
11-16 August 2024 |
Series Name |
Proceedings of the 23rd Workshop on Biomedical Natural Language Processing |
Page Range |
pp. 72-84 |
Date |
2024 |
Export |