The University of Southampton
University of Southampton Institutional Repository

LInKs “lifting independent keypoints” - partial pose lifting for occlusion handling with improved accuracy in 2D-3D human pose estimation

LInKs “lifting independent keypoints” - partial pose lifting for occlusion handling with improved accuracy in 2D-3D human pose estimation
LInKs “lifting independent keypoints” - partial pose lifting for occlusion handling with improved accuracy in 2D-3D human pose estimation
We present LInKs, a novel unsupervised learning method to recover 3D human poses from 2D kinematic skeletons obtained from a single image, even when occlusions are present. Our approach follows a unique two-step process, which involves first lifting the occluded 2D pose to the 3D domain, followed by filling in the occluded parts using the partially reconstructed 3D coordinates. This lift-then-fill approach leads to significantly more accurate results compared to models that complete the pose in 2D space alone. Additionally, we improve the stability and likelihood estimation of normalising flows through a custom sampling function replacing PCA dimensionality reduction used in prior work. Furthermore, we are the first to investigate if different parts of the 2D kinematic skeleton can be lifted independently which we find by itself reduces the error of current lifting approaches. We attribute this to the reduction of long-range keypoint correlations. In our detailed evaluation, we quantify the error under various realistic occlusion scenarios, showcasing the versatility and applicability of our model. Our results consistently demonstrate the superiority of handling all types of occlusions in 3D space when compared to others that complete the pose in 2D space. Our approach also exhibits consistent accuracy in scenarios without occlusion, as evidenced by a 7.9% reduction in reconstruction error compared to prior works on the Human3.6M dataset. Furthermore, our method excels in accurately retrieving complete 3D poses even in the presence of occlusions, making it highly applicable in situations where complete 2D pose information is unavailable.
3D computer vision, Algorithms, Biometrics, Machine learning architectures, and algorithms, body pose, face, formulations, gesture
3414-3423
IEEE
Hardy, Peter
361a5d48-51cf-4eaf-9b60-1de78f2f2f20
Kim, Hansung
2c7c135c-f00b-4409-acb2-85b3a9e8225f
Hardy, Peter
361a5d48-51cf-4eaf-9b60-1de78f2f2f20
Kim, Hansung
2c7c135c-f00b-4409-acb2-85b3a9e8225f

Hardy, Peter and Kim, Hansung (2024) LInKs “lifting independent keypoints” - partial pose lifting for occlusion handling with improved accuracy in 2D-3D human pose estimation. In Proceedings - 2024 IEEE Winter Conference on Applications of Computer Vision, WACV 2024. IEEE. pp. 3414-3423 . (doi:10.1109/WACV57701.2024.00339).

Record type: Conference or Workshop Item (Paper)

Abstract

We present LInKs, a novel unsupervised learning method to recover 3D human poses from 2D kinematic skeletons obtained from a single image, even when occlusions are present. Our approach follows a unique two-step process, which involves first lifting the occluded 2D pose to the 3D domain, followed by filling in the occluded parts using the partially reconstructed 3D coordinates. This lift-then-fill approach leads to significantly more accurate results compared to models that complete the pose in 2D space alone. Additionally, we improve the stability and likelihood estimation of normalising flows through a custom sampling function replacing PCA dimensionality reduction used in prior work. Furthermore, we are the first to investigate if different parts of the 2D kinematic skeleton can be lifted independently which we find by itself reduces the error of current lifting approaches. We attribute this to the reduction of long-range keypoint correlations. In our detailed evaluation, we quantify the error under various realistic occlusion scenarios, showcasing the versatility and applicability of our model. Our results consistently demonstrate the superiority of handling all types of occlusions in 3D space when compared to others that complete the pose in 2D space. Our approach also exhibits consistent accuracy in scenarios without occlusion, as evidenced by a 7.9% reduction in reconstruction error compared to prior works on the Human3.6M dataset. Furthermore, our method excels in accurately retrieving complete 3D poses even in the presence of occlusions, making it highly applicable in situations where complete 2D pose information is unavailable.

This record has no associated files available for download.

More information

Published date: 9 April 2024
Additional Information: Publisher Copyright: © 2024 IEEE.
Venue - Dates: IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa , HAWAII, United States, 2024-01-04 - 2024-01-08
Keywords: 3D computer vision, Algorithms, Biometrics, Machine learning architectures, and algorithms, body pose, face, formulations, gesture

Identifiers

Local EPrints ID: 490628
URI: http://eprints.soton.ac.uk/id/eprint/490628
PURE UUID: d2d29300-2409-4881-a2db-792c2113f52a
ORCID for Hansung Kim: ORCID iD orcid.org/0000-0003-4907-0491

Catalogue record

Date deposited: 31 May 2024 16:46
Last modified: 12 Dec 2024 02:59

Export record

Altmetrics

Contributors

Author: Peter Hardy
Author: Hansung Kim ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×