Javascript must be enabled to continue!
CLINES: Clinical LLM-based Information Extraction and Structuring Agent
View through CrossRef
Abstract
Background
Clinical narratives in electronic health records (EHRs) contain essential diagnostic, therapeutic, and temporal information that is often missing from structured fields, leaving manual chart review as the
de facto
standard for high-quality labels, but slow, costly, and variable, thereby constraining accurate cohort construction for clinical trials, large-scale epidemiologic studies, and the development of robust machine-learning models.
Methods
We developed CLINES, a modular agentic pipeline that extracts and structures clinical concepts: semantic chunking of long notes; extraction by reasoningcapable large language models; assignment of attributes (assertion/experiencer, numerical values with SI units); normalization to the Unified Medical Language System (UMLS); resolution of explicit and relative dates; and aggregation into an i2b2-style schema. Zero-shot evaluation was conducted on de-identified EHR: MIMIC-III notes, 4CE notes, and CORAL oncology reports (breast, pancreas). Comparators included rule/lexicon systems, transformer encoders, and single-prompt LLM baselines. Outcomes were F1 scores for entity extraction, assertion status, value&unit extraction, and date processing.
Findings
Across all datasets, CLINES led every baseline. F1 scores (entity / assertion / value&unit / date) were: MIMIC-III 0.69 / 0.93 / 0.90 (date not evaluated); 4CE 0.87 / 0.88 / 0.79 / 0.79; CORAL–Breast 0.81 / 0.84 / 0.77 / 0.73; CORAL–Pancreas 0.85 / 0.87 / 0.90 / 0.78. Gains over the strongest single-prompt LLM were +0.21–0.38 across tasks, and transformer encoders trailed by +0.28–0.68 F1 on entity extraction. Performance remained stable across note-length quantiles, while transformer baselines lost recall as notes lengthened.
Interpretation
CLINES translates narrative text from electronic health records into ontology-grounded, auditable, and schema-ready data, offering a practical route to scale chart-review-like extraction for cohort discovery and real-world evidence. CLINES is model agnostic–different open and close models can be substituted to achieve specific cost, performance, and privacy goals. Future work aims to quantify inter-annotator agreements and explore adaptive feedback and domain-specific fine-tuning.
Title: CLINES: Clinical LLM-based Information Extraction and Structuring Agent
Description:
Abstract
Background
Clinical narratives in electronic health records (EHRs) contain essential diagnostic, therapeutic, and temporal information that is often missing from structured fields, leaving manual chart review as the
de facto
standard for high-quality labels, but slow, costly, and variable, thereby constraining accurate cohort construction for clinical trials, large-scale epidemiologic studies, and the development of robust machine-learning models.
Methods
We developed CLINES, a modular agentic pipeline that extracts and structures clinical concepts: semantic chunking of long notes; extraction by reasoningcapable large language models; assignment of attributes (assertion/experiencer, numerical values with SI units); normalization to the Unified Medical Language System (UMLS); resolution of explicit and relative dates; and aggregation into an i2b2-style schema.
Zero-shot evaluation was conducted on de-identified EHR: MIMIC-III notes, 4CE notes, and CORAL oncology reports (breast, pancreas).
Comparators included rule/lexicon systems, transformer encoders, and single-prompt LLM baselines.
Outcomes were F1 scores for entity extraction, assertion status, value&unit extraction, and date processing.
Findings
Across all datasets, CLINES led every baseline.
F1 scores (entity / assertion / value&unit / date) were: MIMIC-III 0.
69 / 0.
93 / 0.
90 (date not evaluated); 4CE 0.
87 / 0.
88 / 0.
79 / 0.
79; CORAL–Breast 0.
81 / 0.
84 / 0.
77 / 0.
73; CORAL–Pancreas 0.
85 / 0.
87 / 0.
90 / 0.
78.
Gains over the strongest single-prompt LLM were +0.
21–0.
38 across tasks, and transformer encoders trailed by +0.
28–0.
68 F1 on entity extraction.
Performance remained stable across note-length quantiles, while transformer baselines lost recall as notes lengthened.
Interpretation
CLINES translates narrative text from electronic health records into ontology-grounded, auditable, and schema-ready data, offering a practical route to scale chart-review-like extraction for cohort discovery and real-world evidence.
CLINES is model agnostic–different open and close models can be substituted to achieve specific cost, performance, and privacy goals.
Future work aims to quantify inter-annotator agreements and explore adaptive feedback and domain-specific fine-tuning.
Related Results
Exploring Large Language Models Integration in the Histopathologic Diagnosis of Skin Diseases: A Comparative Study
Exploring Large Language Models Integration in the Histopathologic Diagnosis of Skin Diseases: A Comparative Study
Abstract
Introduction
The exact manner in which large language models (LLMs) will be integrated into pathology is not yet fully comprehended. This study examines the accuracy, bene...
Human-AI Collaboration in Clinical Reasoning: A UK Replication and Interaction Analysis
Human-AI Collaboration in Clinical Reasoning: A UK Replication and Interaction Analysis
Abstract
Objective
A paper from Goh et al found that a large language model (LLM) working alone outperformed American clinicians assisted...
Financial Advisory LLM Model for Modernizing Financial Services and Innovative Solutions for Financial Literacy in India
Financial Advisory LLM Model for Modernizing Financial Services and Innovative Solutions for Financial Literacy in India
Abstract
Dynamically evolving financial conditions in India place sophisticated models of financial advisory services relative to its own peculiar conditions more in demand...
Development and Evaluation of a Retrieval-Augmented Large Language Model Framework for Ophthalmology
Development and Evaluation of a Retrieval-Augmented Large Language Model Framework for Ophthalmology
ImportanceAlthough augmenting large language models (LLMs) with knowledge bases may improve medical domain–specific performance, practical methods are needed for local implementati...
Optimization of ultrasonic extraction of
Lycium barbarum
polysaccharides using response surface methodology
Optimization of ultrasonic extraction of
Lycium barbarum
polysaccharides using response surface methodology
Abstract
Ultrasonic extraction was a new development method to achieve high-efficiency extraction of
Lycium barbarum
...
CAT-LLM: Style-enhanced Large Language Models with Text Style Definition for Chinese Article-style Transfer
CAT-LLM: Style-enhanced Large Language Models with Text Style Definition for Chinese Article-style Transfer
Text style transfer plays a vital role in online entertainment and social media. However, existing models struggle to handle the complexity of Chinese long texts, such as rhetoric,...
Utilizing Large Language Models for Geoscience Literature Information Extraction
Utilizing Large Language Models for Geoscience Literature Information Extraction
Extracting information from unstructured and semi-structured geoscience literature is a crucial step in conducting geological research. The traditional machine learning extraction ...
APPLICATION OF INTELLIGENT MULTIAGENT APPROACH TO LYME DISEASE SIMULATION
APPLICATION OF INTELLIGENT MULTIAGENT APPROACH TO LYME DISEASE SIMULATION
ObjectiveThe objective of this research is to develop the model for calculating the forecast of the Lyme disease dynamics what will help to take effective preventive and control me...

