Abstract
Laboratory tests generate structured numerical data, which a clinician must interpret to justify diagnoses and help patients understand the outcomes of the tests. LLMs have the potential to assist with the generation of interpretative comments, but legitimate concerns remain about the accuracy and reliability of the generation process. This work introduces LAB-KG, which conditions the generation process of an LLM on information retrieved from a knowledge graph of relevant patient conditions and lab test results. This helps to ground the text-generation process in accurate medical knowledge and enables generated text to be traced back to the knowledge graph. Given a dataset of laboratory test results and associated interpretive comments, we show how an LLM can build a KG of the relationships between laboratory test results, reference ranges, patient conditions and demographic information. We further show that the interpretive comments produced by an LLM conditioned on information retrieved from the KG are of higher quality than those from a standard RAG method. Finally, we show how our KG approach can improve the interpretability of the LLM generated text.
Original language | English |
---|---|
Title of host publication | Proceedings of the 31st International Conference on Computational Linguistics |
Publisher | ACL |
Publication status | Accepted - 09 Dec 2024 |
Event | The 31st International Conference on Computational Linguistics - Abu Dhabi, UAE Duration: 19 Jan 2025 → 24 Jan 2025 https://coling2025.org/ |
Conference
Conference | The 31st International Conference on Computational Linguistics |
---|---|
Abbreviated title | COLING 2025 |
Period | 19/01/2025 → 24/01/2025 |
Internet address |