TY - GEN
T1 - LATA – Label attention transformer architectures for ICD-10 coding of unstructured clinical notes
AU - Mayya, Veena
AU - Kamath, Sowmya S.
AU - Sugumaran, Vijayan
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Effective code assignment for patient clinical records in a hospital plays a significant role in the process of standardizing medical records, mainly for streamlining clinical care delivery, billing, and managing insurance claims. The current practice employed is manual coding, usually carried out by trained medical coders, making the process subjective, error-prone, inexact, and time-consuming. To alleviate this cost-intensive process, intelligent coding systems built on patients’ structured electronic medical records are critical. Classification of medical diagnostic codes, like ICD-10, is widely employed to categorize patients’ clinical conditions and associated diagnoses. In this work, we present a neural model LATA, built on Label Attention Transformer Architectures for automatic assignment of ICD-10 codes. Our work is benchmarked on the CodiEsp dataset, a dataset for automatic clinical coding systems for multilingual medical documents, used in the eHealth CLEF 2020-Multilingual Information Extraction Shared Task. The experimental results reveal that the proposed LATA variants outperform their basic BERT counterparts by 33-49% in terms of standard metrics like precision, recall, F1-score and mean average precision. The label attention mechanism also enables direct extraction of textual evidence in medical documents that map to the clinical ICD-10 diagnostic codes.
AB - Effective code assignment for patient clinical records in a hospital plays a significant role in the process of standardizing medical records, mainly for streamlining clinical care delivery, billing, and managing insurance claims. The current practice employed is manual coding, usually carried out by trained medical coders, making the process subjective, error-prone, inexact, and time-consuming. To alleviate this cost-intensive process, intelligent coding systems built on patients’ structured electronic medical records are critical. Classification of medical diagnostic codes, like ICD-10, is widely employed to categorize patients’ clinical conditions and associated diagnoses. In this work, we present a neural model LATA, built on Label Attention Transformer Architectures for automatic assignment of ICD-10 codes. Our work is benchmarked on the CodiEsp dataset, a dataset for automatic clinical coding systems for multilingual medical documents, used in the eHealth CLEF 2020-Multilingual Information Extraction Shared Task. The experimental results reveal that the proposed LATA variants outperform their basic BERT counterparts by 33-49% in terms of standard metrics like precision, recall, F1-score and mean average precision. The label attention mechanism also enables direct extraction of textual evidence in medical documents that map to the clinical ICD-10 diagnostic codes.
UR - http://www.scopus.com/inward/record.url?scp=85126461409&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85126461409&partnerID=8YFLogxK
U2 - 10.1109/CIBCB49929.2021.9562815
DO - 10.1109/CIBCB49929.2021.9562815
M3 - Conference contribution
AN - SCOPUS:85126461409
T3 - 2021 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology, CIBCB 2021
BT - 2021 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology, CIBCB 2021
A2 - Hallinan, Jennifer
A2 - Chetty, Madhu
A2 - Heredia, Gonzalo Ruz
A2 - Shatte, Adrian
A2 - Lim, Suryani
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2021 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology, CIBCB 2021
Y2 - 13 October 2021 through 15 October 2021
ER -