Hongbo Fan


Fixing paper assignments

  1. Please select all papers that belong to the same person.
  2. Indicate below which author they should be assigned to.
Provide a valid ORCID iD here. This will be used to match future papers to this author.
Provide the name of the school or the university where the author has received or will receive their highest degree (e.g., Ph.D. institution for researchers, or current affiliation for students). This will be used to form the new author page ID, if needed.

TODO: "submit" and "cancel" buttons here


2022

pdf bib
BERT for Long Documents: A Case Study of Automated ICD Coding
Arash Afkanpour | Shabir Adeel | Hansenclever Bassani | Arkady Epshteyn | Hongbo Fan | Isaac Jones | Mahan Malihi | Adrian Nauth | Raj Sinha | Sanjana Woonna | Shiva Zamani | Elli Kanal | Mikhail Fomitchev | Donny Cheung
Proceedings of the 13th International Workshop on Health Text Mining and Information Analysis (LOUHI)

Transformer models have achieved great success across many NLP problems. However, previous studies in automated ICD coding concluded that these models fail to outperform some of the earlier solutions such as CNN-based models. In this paper we challenge this conclusion. We present a simple and scalable method to process long text with the existing transformer models such as BERT. We show that this method significantly improves the previous results reported for transformer models in ICD coding, and is able to outperform one of the prominent CNN-based methods.