The University of Southampton
University of Southampton Institutional Repository

Dense material segmentation with context-aware network

Dense material segmentation with context-aware network
Dense material segmentation with context-aware network
The dense material segmentation task aims at recognising the material for every pixel in daily images. It is beneficial to applications such as robot manipulation and spatial audio synthesis. Modern deep-learning methods combine material features with contextual features. Material features can generalise to unseen images regardless of appearance properties such as material shape and colour. Contextual features can reduce the segmentation uncertainty by providing extra global or semi-global information about the image. Recent studies proposed to crop the images into patches, which forces the network to learn material features from local visual clues. Typical contextual information includes extracted feature maps from networks targeting object and place related tasks. However, due to the lack of contextual labels, existing methods use pre-trained networks to provide contextual features. As a consequence, the trained networks do not give a promising performance. Their accuracy is below 70%, and the predicted segments have coarse boundaries. Considering this problem, this chapter introduces the Context-Aware Material Segmentation Network (CAM-SegNet). The CAM-SegNet is a hybrid network architecture to simultaneously learn from contextual and material features jointly with labelled materials. The effectiveness of the CAM-SegNet is demonstrated by training the network to learn boundary-related contextual features. Since the existing material datasets are sparsely labelled, a self-training approach is adopted to fill in the unlabelled pixels. Experiments show that CAM-SegNet can identify materials correctly, even with similar appearances. The network improves the pixel accuracy by 3–20% and raises the Mean IoU by 6–28%.
1865-0929
66-88
Springer Cham
Heng, Yuwen
a3edf9da-2d3b-450c-8d6d-85f76c861849
Wu, Yihong
2876bede-25f1-47a5-9e08-b98be99b2d31
Dasmahapatra, Srinandan
eb5fd76f-4335-4ae9-a88a-20b9e2b3f698
Kim, Hansung
2c7c135c-f00b-4409-acb2-85b3a9e8225f
Augusto de Sousa, A.
Debattista, Kurt
Paljic, Alexis
Ziat, Mounia
Hurter, Christophe
Purchase, Helen
Farinella, Giovanni Maria
Radeva, Petia
Bouatouch, Kadi
Heng, Yuwen
a3edf9da-2d3b-450c-8d6d-85f76c861849
Wu, Yihong
2876bede-25f1-47a5-9e08-b98be99b2d31
Dasmahapatra, Srinandan
eb5fd76f-4335-4ae9-a88a-20b9e2b3f698
Kim, Hansung
2c7c135c-f00b-4409-acb2-85b3a9e8225f
Augusto de Sousa, A.
Debattista, Kurt
Paljic, Alexis
Ziat, Mounia
Hurter, Christophe
Purchase, Helen
Farinella, Giovanni Maria
Radeva, Petia
Bouatouch, Kadi

Heng, Yuwen, Wu, Yihong, Dasmahapatra, Srinandan and Kim, Hansung (2023) Dense material segmentation with context-aware network. In, Augusto de Sousa, A., Debattista, Kurt, Paljic, Alexis, Ziat, Mounia, Hurter, Christophe, Purchase, Helen, Farinella, Giovanni Maria, Radeva, Petia and Bouatouch, Kadi (eds.) Computer Vision, Imaging and Computer Graphics Theory and Applications: 17th International Joint Conference, VISIGRAPP 2022, Virtual Event, February 6–8, 2022, Revised Selected Papers. (Communications in Computer and Information Science, 1815) 1 ed. Springer Cham, pp. 66-88. (doi:10.1007/978-3-031-45725-8_4).

Record type: Book Section

Abstract

The dense material segmentation task aims at recognising the material for every pixel in daily images. It is beneficial to applications such as robot manipulation and spatial audio synthesis. Modern deep-learning methods combine material features with contextual features. Material features can generalise to unseen images regardless of appearance properties such as material shape and colour. Contextual features can reduce the segmentation uncertainty by providing extra global or semi-global information about the image. Recent studies proposed to crop the images into patches, which forces the network to learn material features from local visual clues. Typical contextual information includes extracted feature maps from networks targeting object and place related tasks. However, due to the lack of contextual labels, existing methods use pre-trained networks to provide contextual features. As a consequence, the trained networks do not give a promising performance. Their accuracy is below 70%, and the predicted segments have coarse boundaries. Considering this problem, this chapter introduces the Context-Aware Material Segmentation Network (CAM-SegNet). The CAM-SegNet is a hybrid network architecture to simultaneously learn from contextual and material features jointly with labelled materials. The effectiveness of the CAM-SegNet is demonstrated by training the network to learn boundary-related contextual features. Since the existing material datasets are sparsely labelled, a self-training approach is adopted to fill in the unlabelled pixels. Experiments show that CAM-SegNet can identify materials correctly, even with similar appearances. The network improves the pixel accuracy by 3–20% and raises the Mean IoU by 6–28%.

This record has no associated files available for download.

More information

Published date: 18 October 2023

Identifiers

Local EPrints ID: 490626
URI: http://eprints.soton.ac.uk/id/eprint/490626
ISSN: 1865-0929
PURE UUID: 4973b9c7-d1f4-4f1b-b5fa-78347162e435
ORCID for Yuwen Heng: ORCID iD orcid.org/0000-0003-3793-4811
ORCID for Yihong Wu: ORCID iD orcid.org/0000-0003-3340-2535
ORCID for Hansung Kim: ORCID iD orcid.org/0000-0003-4907-0491

Catalogue record

Date deposited: 31 May 2024 16:46
Last modified: 01 Jun 2024 01:59

Export record

Altmetrics

Contributors

Author: Yuwen Heng ORCID iD
Author: Yihong Wu ORCID iD
Author: Srinandan Dasmahapatra
Author: Hansung Kim ORCID iD
Editor: A. Augusto de Sousa
Editor: Kurt Debattista
Editor: Alexis Paljic
Editor: Mounia Ziat
Editor: Christophe Hurter
Editor: Helen Purchase
Editor: Giovanni Maria Farinella
Editor: Petia Radeva
Editor: Kadi Bouatouch

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×