Fine-tuning BERT Models for Summarizing German Radiology Findings

Abstract:

Writing the conclusion section of radiology reports is essential for communicating the radiology findings and its assessment to physician in a condensed form. In this work, we employ a transformer-based Seq2Seq model for generating the conclusion section of German radiology reports. The model is initialized with the pretrained parameters of a German BERT model and fine-tuned in our downstream task on our domain data. We proposed two strategies to improve the factual correctness of the model. In the first method, next to the abstractive learning objective, we introduce an extraction learning objective to train the decoder in the model to both generate one summary sequence and extract the key findings from the source input. The second approach is to integrate the pointer mechanism into the transformer-based Seq2Seq model. The pointer network helps the Seq2Seq model to choose between generating tokens from the vocabulary or copying parts from the source input during generation. The results of the automatic and human evaluations show that the enhanced Seq2Seq model is capable of generating human-like radiology conclusions and that the improved models effectively reduce the factual errors in the generations despite the small amount of training data.

SEEK ID: https://publications.h-its.org/publications/1498

DOI: 10.18653/v1/2022.clinicalnlp-1.4

Research Groups: Natural Language Processing

Publication type: InProceedings

Journal: Proceedings of the 4th Clinical Natural Language Processing Workshop

Book Title: Proceedings of the 4th Clinical Natural Language Processing Workshop, Seattle, Washington, July 2022

Publisher: Association for Computational Linguistics

Citation: Proceedings of the 4th Clinical Natural Language Processing Workshop, Seattle, Washington, July 2022

Date Published: 14th Jul 2022

URL: https://aclanthology.org/2022.clinicalnlp-1.4.pdf

Registered Mode: manually

Authors: Siting Liang, Klaus Kades, Matthias Fink, Peter Full, Tim Weber, Jens Kleesiek, Michael Strube, Klaus Maier-Hein

help Submitter
Citation
Liang, S., Kades, K., Fink, M., Full, P., Weber, T., Kleesiek, J., Strube, M., & Maier-Hein, K. (2022). Fine-tuning BERT Models for Summarizing German Radiology Findings. In Proceedings of the 4th Clinical Natural Language Processing Workshop. Proceedings of the 4th Clinical Natural Language Processing Workshop. Association for Computational Linguistics. https://doi.org/10.18653/v1/2022.clinicalnlp-1.4
Activity

Views: 3205

Created: 19th Jul 2022 at 14:53

Last updated: 5th Mar 2024 at 21:24

help Tags

This item has not yet been tagged.

help Attributions

None

Powered by
(v.1.14.2)
Copyright © 2008 - 2023 The University of Manchester and HITS gGmbH