Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

CLINES: Clinical LLM-based Information Extraction and Structuring Agent

View through CrossRef
Abstract Background Clinical narratives in electronic health records (EHRs) contain essential diagnostic, therapeutic, and temporal information that is often missing from structured fields, leaving manual chart review as the de facto standard for high-quality labels, but slow, costly, and variable, thereby constraining accurate cohort construction for clinical trials, large-scale epidemiologic studies, and the development of robust machine-learning models. Methods We developed CLINES, a modular agentic pipeline that extracts and structures clinical concepts: semantic chunking of long notes; extraction by reasoningcapable large language models; assignment of attributes (assertion/experiencer, numerical values with SI units); normalization to the Unified Medical Language System (UMLS); resolution of explicit and relative dates; and aggregation into an i2b2-style schema. Zero-shot evaluation was conducted on de-identified EHR: MIMIC-III notes, 4CE notes, and CORAL oncology reports (breast, pancreas). Comparators included rule/lexicon systems, transformer encoders, and single-prompt LLM baselines. Outcomes were F1 scores for entity extraction, assertion status, value&unit extraction, and date processing. Findings Across all datasets, CLINES led every baseline. F1 scores (entity / assertion / value&unit / date) were: MIMIC-III 0.69 / 0.93 / 0.90 (date not evaluated); 4CE 0.87 / 0.88 / 0.79 / 0.79; CORAL–Breast 0.81 / 0.84 / 0.77 / 0.73; CORAL–Pancreas 0.85 / 0.87 / 0.90 / 0.78. Gains over the strongest single-prompt LLM were +0.21–0.38 across tasks, and transformer encoders trailed by +0.28–0.68 F1 on entity extraction. Performance remained stable across note-length quantiles, while transformer baselines lost recall as notes lengthened. Interpretation CLINES translates narrative text from electronic health records into ontology-grounded, auditable, and schema-ready data, offering a practical route to scale chart-review-like extraction for cohort discovery and real-world evidence. CLINES is model agnostic–different open and close models can be substituted to achieve specific cost, performance, and privacy goals. Future work aims to quantify inter-annotator agreements and explore adaptive feedback and domain-specific fine-tuning.
Title: CLINES: Clinical LLM-based Information Extraction and Structuring Agent
Description:
Abstract Background Clinical narratives in electronic health records (EHRs) contain essential diagnostic, therapeutic, and temporal information that is often missing from structured fields, leaving manual chart review as the de facto standard for high-quality labels, but slow, costly, and variable, thereby constraining accurate cohort construction for clinical trials, large-scale epidemiologic studies, and the development of robust machine-learning models.
Methods We developed CLINES, a modular agentic pipeline that extracts and structures clinical concepts: semantic chunking of long notes; extraction by reasoningcapable large language models; assignment of attributes (assertion/experiencer, numerical values with SI units); normalization to the Unified Medical Language System (UMLS); resolution of explicit and relative dates; and aggregation into an i2b2-style schema.
Zero-shot evaluation was conducted on de-identified EHR: MIMIC-III notes, 4CE notes, and CORAL oncology reports (breast, pancreas).
Comparators included rule/lexicon systems, transformer encoders, and single-prompt LLM baselines.
Outcomes were F1 scores for entity extraction, assertion status, value&unit extraction, and date processing.
Findings Across all datasets, CLINES led every baseline.
F1 scores (entity / assertion / value&unit / date) were: MIMIC-III 0.
69 / 0.
93 / 0.
90 (date not evaluated); 4CE 0.
87 / 0.
88 / 0.
79 / 0.
79; CORAL–Breast 0.
81 / 0.
84 / 0.
77 / 0.
73; CORAL–Pancreas 0.
85 / 0.
87 / 0.
90 / 0.
78.
Gains over the strongest single-prompt LLM were +0.
21–0.
38 across tasks, and transformer encoders trailed by +0.
28–0.
68 F1 on entity extraction.
Performance remained stable across note-length quantiles, while transformer baselines lost recall as notes lengthened.
Interpretation CLINES translates narrative text from electronic health records into ontology-grounded, auditable, and schema-ready data, offering a practical route to scale chart-review-like extraction for cohort discovery and real-world evidence.
CLINES is model agnostic–different open and close models can be substituted to achieve specific cost, performance, and privacy goals.
Future work aims to quantify inter-annotator agreements and explore adaptive feedback and domain-specific fine-tuning.

Related Results

Exploring Large Language Models Integration in the Histopathologic Diagnosis of Skin Diseases: A Comparative Study
Exploring Large Language Models Integration in the Histopathologic Diagnosis of Skin Diseases: A Comparative Study
Abstract Introduction The exact manner in which large language models (LLMs) will be integrated into pathology is not yet fully comprehended. This study examines the accuracy, bene...
Human-AI Collaboration in Clinical Reasoning: A UK Replication and Interaction Analysis
Human-AI Collaboration in Clinical Reasoning: A UK Replication and Interaction Analysis
Abstract Objective A paper from Goh et al found that a large language model (LLM) working alone outperformed American clinicians assisted...
Development and Evaluation of a Retrieval-Augmented Large Language Model Framework for Ophthalmology
Development and Evaluation of a Retrieval-Augmented Large Language Model Framework for Ophthalmology
ImportanceAlthough augmenting large language models (LLMs) with knowledge bases may improve medical domain–specific performance, practical methods are needed for local implementati...
CAT-LLM: Style-enhanced Large Language Models with Text Style Definition for Chinese Article-style Transfer
CAT-LLM: Style-enhanced Large Language Models with Text Style Definition for Chinese Article-style Transfer
Text style transfer plays a vital role in online entertainment and social media. However, existing models struggle to handle the complexity of Chinese long texts, such as rhetoric,...
Utilizing Large Language Models for Geoscience Literature Information Extraction
Utilizing Large Language Models for Geoscience Literature Information Extraction
Extracting information from unstructured and semi-structured geoscience literature is a crucial step in conducting geological research. The traditional machine learning extraction ...
APPLICATION OF INTELLIGENT MULTIAGENT APPROACH TO LYME DISEASE SIMULATION
APPLICATION OF INTELLIGENT MULTIAGENT APPROACH TO LYME DISEASE SIMULATION
ObjectiveThe objective of this research is to develop the model for calculating the forecast of the Lyme disease dynamics what will help to take effective preventive and control me...
Emerging Practices in LLM-integrated Game Writing
Emerging Practices in LLM-integrated Game Writing
Abstract This article examines emerging practices in large language model (LLM) integration within game writing, focusing on how these technologies reshape narrat...

Back to Top