The University of Southampton
University of Southampton Institutional Repository

Multi-target landmark detection with incomplete images via reinforcement learning and shape prior embedding

Multi-target landmark detection with incomplete images via reinforcement learning and shape prior embedding
Multi-target landmark detection with incomplete images via reinforcement learning and shape prior embedding

Medical images are generally acquired with limited field-of-view (FOV), which could lead to incomplete regions of interest (ROI), and thus impose a great challenge on medical image analysis. This is particularly evident for the learning-based multi-target landmark detection, where algorithms could be misleading to learn primarily the variation of background due to the varying FOV, failing the detection of targets. Based on learning a navigation policy, instead of predicting targets directly, reinforcement learning (RL)-based methods have the potential to tackle this challenge in an efficient manner. Inspired by this, in this work we propose a multi-agent RL framework for simultaneous multi-target landmark detection. This framework is aimed to learn from incomplete or (and) complete images to form an implicit knowledge of global structure, which is consolidated during the training stage for the detection of targets from either complete or incomplete test images. To further explicitly exploit the global structural information from incomplete images, we propose to embed a shape model into the RL process. With this prior knowledge, the proposed RL model can not only localize dozens of targets simultaneously, but also work effectively and robustly in the presence of incomplete images. We validated the applicability and efficacy of the proposed method on various multi-target detection tasks with incomplete images from practical clinics, using body dual-energy X-ray absorptiometry (DXA), cardiac MRI and head CT datasets. Results showed that our method could predict whole set of landmarks with incomplete training images up to 80% missing proportion (average distance error 2.29 cm on body DXA), and could detect unseen landmarks in regions with missing image information outside FOV of target images (average distance error 6.84 mm on 3D half-head CT). Our code will be released via https://zmiclab.github.io/projects.html.

Incomplete image, Landmark detection, Multi-agent reinforcement learning, Shape prior
1361-8415
Wan, Kaiwen
5dbb17b4-f186-49ea-9ee0-eaf3ac897dcc
Li, Lei
2da88502-0bd8-4e6b-8f7d-0c01a48b399e
Jia, Dengqiang
2dcae8a2-7a41-43e6-990e-fa8129ffe6e6
Gao, Shangqi
eee7fca1-9aa7-42ff-93d2-64c47d5b7955
Qian, Wei
46aa5579-4bf9-4dba-a696-033e74713842
Wu, Yingzhi
b114c190-20d4-4d0d-bb5c-5a78666e126e
Lin, Huandong
c4bd00ac-8727-4fea-8146-9d1482a8a5b1
Mu, Xiongzheng
b269f32b-0d1b-48b4-911c-f74e3b8035f8
Gao, Xin
50c863fb-52d2-4ff3-bc77-da2ff9e314a6
Wang, Sijia
99c46d0e-0622-4b65-96e7-5d1cf563b1de
Wu, Fuping
cee10409-dfca-4d75-9b28-ed5747aab173
Zhuang, Xiahai
c58e977b-e70e-4b37-9acd-b7f8070d98a8
et al.
Wan, Kaiwen
5dbb17b4-f186-49ea-9ee0-eaf3ac897dcc
Li, Lei
2da88502-0bd8-4e6b-8f7d-0c01a48b399e
Jia, Dengqiang
2dcae8a2-7a41-43e6-990e-fa8129ffe6e6
Gao, Shangqi
eee7fca1-9aa7-42ff-93d2-64c47d5b7955
Qian, Wei
46aa5579-4bf9-4dba-a696-033e74713842
Wu, Yingzhi
b114c190-20d4-4d0d-bb5c-5a78666e126e
Lin, Huandong
c4bd00ac-8727-4fea-8146-9d1482a8a5b1
Mu, Xiongzheng
b269f32b-0d1b-48b4-911c-f74e3b8035f8
Gao, Xin
50c863fb-52d2-4ff3-bc77-da2ff9e314a6
Wang, Sijia
99c46d0e-0622-4b65-96e7-5d1cf563b1de
Wu, Fuping
cee10409-dfca-4d75-9b28-ed5747aab173
Zhuang, Xiahai
c58e977b-e70e-4b37-9acd-b7f8070d98a8

Wan, Kaiwen, Li, Lei and Jia, Dengqiang , et al. (2023) Multi-target landmark detection with incomplete images via reinforcement learning and shape prior embedding. Medical Image Analysis, 89, [102875]. (doi:10.1016/j.media.2023.102875).

Record type: Article

Abstract

Medical images are generally acquired with limited field-of-view (FOV), which could lead to incomplete regions of interest (ROI), and thus impose a great challenge on medical image analysis. This is particularly evident for the learning-based multi-target landmark detection, where algorithms could be misleading to learn primarily the variation of background due to the varying FOV, failing the detection of targets. Based on learning a navigation policy, instead of predicting targets directly, reinforcement learning (RL)-based methods have the potential to tackle this challenge in an efficient manner. Inspired by this, in this work we propose a multi-agent RL framework for simultaneous multi-target landmark detection. This framework is aimed to learn from incomplete or (and) complete images to form an implicit knowledge of global structure, which is consolidated during the training stage for the detection of targets from either complete or incomplete test images. To further explicitly exploit the global structural information from incomplete images, we propose to embed a shape model into the RL process. With this prior knowledge, the proposed RL model can not only localize dozens of targets simultaneously, but also work effectively and robustly in the presence of incomplete images. We validated the applicability and efficacy of the proposed method on various multi-target detection tasks with incomplete images from practical clinics, using body dual-energy X-ray absorptiometry (DXA), cardiac MRI and head CT datasets. Results showed that our method could predict whole set of landmarks with incomplete training images up to 80% missing proportion (average distance error 2.29 cm on body DXA), and could detect unseen landmarks in regions with missing image information outside FOV of target images (average distance error 6.84 mm on 3D half-head CT). Our code will be released via https://zmiclab.github.io/projects.html.

This record has no associated files available for download.

More information

Accepted/In Press date: 13 June 2023
e-pub ahead of print date: 24 June 2023
Published date: 11 July 2023
Keywords: Incomplete image, Landmark detection, Multi-agent reinforcement learning, Shape prior

Identifiers

Local EPrints ID: 488808
URI: http://eprints.soton.ac.uk/id/eprint/488808
ISSN: 1361-8415
PURE UUID: 8726a024-9644-479d-afe0-ce97a8166df1
ORCID for Lei Li: ORCID iD orcid.org/0000-0003-1281-6472

Catalogue record

Date deposited: 05 Apr 2024 16:45
Last modified: 10 Apr 2024 02:14

Export record

Altmetrics

Contributors

Author: Kaiwen Wan
Author: Lei Li ORCID iD
Author: Dengqiang Jia
Author: Shangqi Gao
Author: Wei Qian
Author: Yingzhi Wu
Author: Huandong Lin
Author: Xiongzheng Mu
Author: Xin Gao
Author: Sijia Wang
Author: Fuping Wu
Author: Xiahai Zhuang
Corporate Author: et al.

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×