Generating Mammography Reports from Multi-view Mammograms with BERT
Alexander Yalunin | Elena Sokolova | Ilya Burenko | Alexander Ponomarchuk | Olga Puchkova | Dmitriy Umerenkov
Findings of the Association for Computational Linguistics: EMNLP 2021
Writing mammography reports can be error-prone and time-consuming for radiologists. In this paper we propose a method to generate mammography reports given four images, corresponding to the four views used in screening mammography. To the best of our knowledge our work represents the first attempt to generate the mammography report using deep-learning. We propose an encoder-decoder model that includes an EfficientNet-based encoder and a Transformer-based decoder. We demonstrate that the Transformer-based attention mechanism can combine visual and semantic information to localize salient regions on the input mammograms and generate a visually interpretable report. The conducted experiments, including an evaluation by a certified radiologist, show the effectiveness of the proposed method.