The University of Southampton
University of Southampton Institutional Repository

IIHT: medical report generation with image-to-indicator hierarchical transformer

IIHT: medical report generation with image-to-indicator hierarchical transformer
IIHT: medical report generation with image-to-indicator hierarchical transformer

Automated medical report generation has become increasingly important in medical analysis. It can produce computer-aided diagnosis descriptions and thus significantly alleviate the doctors’ work. Inspired by the huge success of neural machine translation and image captioning, various deep learning methods have been proposed for medical report generation. However, due to the inherent properties of medical data, including data imbalance and the length and correlation between report sequences, the generated reports by existing methods may exhibit linguistic fluency but lack adequate clinical accuracy. In this work, we propose an image-to-indicator hierarchical transformer (IIHT) framework for medical report generation. It consists of three modules, i.e., a classifier module, an indicator expansion module and a generator module. The classifier module first extracts image features from the input medical images and produces disease-related indicators with their corresponding states. The disease-related indicators are subsequently utilised as input for the indicator expansion module, incorporating the “data-text-data” strategy. The transformer-based generator then leverages these extracted features along with image features as auxiliary information to generate final reports. Furthermore, the proposed IIHT method is feasible for radiologists to modify disease indicators in real-world scenarios and integrate the operations into the indicator expansion module for fluent and accurate medical report generation. Extensive experiments and comparisons with state-of-the-art methods under various evaluation metrics demonstrate the great performance of the proposed method.

Chest X-Ray, Deep neural networks, Medical report generation, Transformers
0302-9743
57-71
Springer Singapore
Fan, Keqiang
0b1613e0-0167-425e-9ab9-986054928dd2
Cai, Xiaohao
de483445-45e9-4b21-a4e8-b0427fc72cee
Niranjan, Mahesan
5cbaeea8-7288-4b55-a89c-c43d212ddd4f
Luo, Biao
Cheng, Long
Wu, Zheng-Guang
Li, Hongyi
Li, Chaojie
Fan, Keqiang
0b1613e0-0167-425e-9ab9-986054928dd2
Cai, Xiaohao
de483445-45e9-4b21-a4e8-b0427fc72cee
Niranjan, Mahesan
5cbaeea8-7288-4b55-a89c-c43d212ddd4f
Luo, Biao
Cheng, Long
Wu, Zheng-Guang
Li, Hongyi
Li, Chaojie

Fan, Keqiang, Cai, Xiaohao and Niranjan, Mahesan (2023) IIHT: medical report generation with image-to-indicator hierarchical transformer. Luo, Biao, Cheng, Long, Wu, Zheng-Guang, Li, Hongyi and Li, Chaojie (eds.) In Neural Information Processing. ICONIP 2023. vol. 14452 LNCS, Springer Singapore. pp. 57-71 . (doi:10.1007/978-981-99-8076-5_5).

Record type: Conference or Workshop Item (Paper)

Abstract

Automated medical report generation has become increasingly important in medical analysis. It can produce computer-aided diagnosis descriptions and thus significantly alleviate the doctors’ work. Inspired by the huge success of neural machine translation and image captioning, various deep learning methods have been proposed for medical report generation. However, due to the inherent properties of medical data, including data imbalance and the length and correlation between report sequences, the generated reports by existing methods may exhibit linguistic fluency but lack adequate clinical accuracy. In this work, we propose an image-to-indicator hierarchical transformer (IIHT) framework for medical report generation. It consists of three modules, i.e., a classifier module, an indicator expansion module and a generator module. The classifier module first extracts image features from the input medical images and produces disease-related indicators with their corresponding states. The disease-related indicators are subsequently utilised as input for the indicator expansion module, incorporating the “data-text-data” strategy. The transformer-based generator then leverages these extracted features along with image features as auxiliary information to generate final reports. Furthermore, the proposed IIHT method is feasible for radiologists to modify disease indicators in real-world scenarios and integrate the operations into the indicator expansion module for fluent and accurate medical report generation. Extensive experiments and comparisons with state-of-the-art methods under various evaluation metrics demonstrate the great performance of the proposed method.

This record has no associated files available for download.

More information

Published date: 14 November 2023
Venue - Dates: 30th International Conference on Neural Information Processing, ICONIP 2023, , Changsha, China, 2023-11-20 - 2023-11-23
Keywords: Chest X-Ray, Deep neural networks, Medical report generation, Transformers

Identifiers

Local EPrints ID: 491924
URI: http://eprints.soton.ac.uk/id/eprint/491924
ISSN: 0302-9743
PURE UUID: 428806d2-6e15-4780-9e3b-56ff7cd26073
ORCID for Xiaohao Cai: ORCID iD orcid.org/0000-0003-0924-2834
ORCID for Mahesan Niranjan: ORCID iD orcid.org/0000-0001-7021-140X

Catalogue record

Date deposited: 08 Jul 2024 16:55
Last modified: 11 Jul 2024 02:06

Export record

Altmetrics

Contributors

Author: Keqiang Fan
Author: Xiaohao Cai ORCID iD
Author: Mahesan Niranjan ORCID iD
Editor: Biao Luo
Editor: Long Cheng
Editor: Zheng-Guang Wu
Editor: Hongyi Li
Editor: Chaojie Li

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×