Predicting seafloor visual classes from multimodal remote sensed priors using location-guided self-supervised learning
Predicting seafloor visual classes from multimodal remote sensed priors using location-guided self-supervised learning
Remote sensed mapping data and seafloor in-situ imagery are often gathered to infer benthic habitat distributions. However, leveraging multimodal data is challenging because of inherent inconsistencies between measurement modes (e.g., resolution, positional offsets, shape discrepancies). We investigate the impact of using location metadata in multimodal, self-supervised feature learning on habitat classification. Experiments were carried out on a multimodal dataset gathered using and Autonomous Underwater Vehicle (AUV) at the Darwin Mounds Marine Protected Area (MPA). Introducing location metadata improved F1 classification performance of a Bayesian classifier by an average of 27.7% over all conditions tested in this work, with a larger improvement of 32.9% achieved when multiple remote sensing data modes are combined for the analysis.
Liang, Cailei
f9a26dcf-539b-42c0-8b54-e266c89cf6ea
Cappelletto, Jose Cruz
a6620d58-0abe-4f9d-9fd9-9ac474de9230
Bodenmann, Adrian
070a668f-cc2f-402a-844e-cdf207b24f50
Turnock, Stephen
d6442f5c-d9af-4fdb-8406-7c79a92b26ce
Huvenne, Veerle
f22be3e2-708c-491b-b985-a438470fa053
Wardell, Catherine
fca74e02-0488-4c35-8d0f-c29633afb913
Thornton, Blair
8293beb5-c083-47e3-b5f0-d9c3cee14be9
18 September 2024
Liang, Cailei
f9a26dcf-539b-42c0-8b54-e266c89cf6ea
Cappelletto, Jose Cruz
a6620d58-0abe-4f9d-9fd9-9ac474de9230
Bodenmann, Adrian
070a668f-cc2f-402a-844e-cdf207b24f50
Turnock, Stephen
d6442f5c-d9af-4fdb-8406-7c79a92b26ce
Huvenne, Veerle
f22be3e2-708c-491b-b985-a438470fa053
Wardell, Catherine
fca74e02-0488-4c35-8d0f-c29633afb913
Thornton, Blair
8293beb5-c083-47e3-b5f0-d9c3cee14be9
Liang, Cailei, Cappelletto, Jose Cruz, Bodenmann, Adrian, Turnock, Stephen, Huvenne, Veerle, Wardell, Catherine and Thornton, Blair
(2024)
Predicting seafloor visual classes from multimodal remote sensed priors using location-guided self-supervised learning.
In AUV Symposium 2024.
IEEE.
6 pp
.
Record type:
Conference or Workshop Item
(Paper)
Abstract
Remote sensed mapping data and seafloor in-situ imagery are often gathered to infer benthic habitat distributions. However, leveraging multimodal data is challenging because of inherent inconsistencies between measurement modes (e.g., resolution, positional offsets, shape discrepancies). We investigate the impact of using location metadata in multimodal, self-supervised feature learning on habitat classification. Experiments were carried out on a multimodal dataset gathered using and Autonomous Underwater Vehicle (AUV) at the Darwin Mounds Marine Protected Area (MPA). Introducing location metadata improved F1 classification performance of a Bayesian classifier by an average of 27.7% over all conditions tested in this work, with a larger improvement of 32.9% achieved when multiple remote sensing data modes are combined for the analysis.
Text
Cailei_AUV2024_conference
- Accepted Manuscript
More information
Submitted date: 8 May 2024
Accepted/In Press date: 6 July 2024
Published date: 18 September 2024
Venue - Dates:
IEEE AUV2024, Northeastern University, Boston, United States, 2024-09-18 - 2024-09-20
Identifiers
Local EPrints ID: 495706
URI: http://eprints.soton.ac.uk/id/eprint/495706
PURE UUID: 44f492f4-f6fc-4755-8f2f-8477ab549337
Catalogue record
Date deposited: 20 Nov 2024 17:50
Last modified: 21 Nov 2024 02:52
Export record
Contributors
Author:
Cailei Liang
Author:
Jose Cruz Cappelletto
Author:
Veerle Huvenne
Author:
Catherine Wardell
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics