Juyong Kim
2022
AnEMIC: A Framework for Benchmarking ICD Coding Models
Juyong Kim
|
Abheesht Sharma
|
Suhas Shanbhogue
|
Jeremy Weiss
|
Pradeep Ravikumar
Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
Diagnostic coding, or ICD coding, is the task of assigning diagnosis codes defined by the ICD (International Classification of Diseases) standard to patient visits based on clinical notes. The current process of manual ICD coding is time-consuming and often error-prone, which suggests the need for automatic ICD coding. However, despite the long history of automatic ICD coding, there have been no standardized frameworks for benchmarking ICD coding models. We open-source an easy-to-use tool named AnEMIC, which provides a streamlined pipeline for preprocessing, training, and evaluating for automatic ICD coding. We correct errors in preprocessing by existing works, and provide key models and weights trained on the correctly preprocessed datasets. We also provide an interactive demo performing real-time inference from custom inputs, and visualizations drawn from explainable AI to analyze the models. We hope the framework helps move the research of ICD coding forward and helps professionals explore the potential of ICD coding. The framework and the associated code are available here.
2021
Improving Compositional Generalization in Classification Tasks via Structure Annotations
Juyong Kim
|
Pradeep Ravikumar
|
Joshua Ainslie
|
Santiago Ontanon
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers)
Compositional generalization is the ability to generalize systematically to a new data distribution by combining known components. Although humans seem to have a great ability to generalize compositionally, state-of-the-art neural models struggle to do so. In this work, we study compositional generalization in classification tasks and present two main contributions. First, we study ways to convert a natural language sequence-to-sequence dataset to a classification dataset that also requires compositional generalization. Second, we show that providing structural hints (specifically, providing parse trees and entity links as attention masks for a Transformer model) helps compositional generalization.
Search
Co-authors
- Abheesht Sharma 1
- Jeremy Weiss 1
- Joshua Ainslie 1
- Pradeep Ravikumar 2
- Santiago Ontanon 1
- show all...