The University of Southampton
University of Southampton Institutional Repository

EchoMotion: enhancing exercise analysis with acoustic sensing

EchoMotion: enhancing exercise analysis with acoustic sensing
EchoMotion: enhancing exercise analysis with acoustic sensing
EchoMotion revolutionizes exercise monitoring with its innovative acoustic-based smart speaker system, designed to perform human pose estimation using inaudible acoustic signals. Leveraging a combination of acoustic features, deep learning techniques, and a custom loss function, the system transforms acoustic reflections from the human body into precise 3D pose estimations. Ground truth data were recorded using the Microsoft Azure Kinect DK, a depth-sensing camera used for evaluation. Data were collected from 22 participants performing five fast-movement cardio exercises in both home and lab environments, yielding over 11 hours of synchronized acoustic and ground truth data. EchoMotion achieved a low Mean Absolute Error (MAE) of 0.59 mm, demonstrating superior accuracy for fast movement exercises compared to the reported MAE range of 2.8 mm to 96 mm in SOTA works, which also focus on slow movements. Our system is non-invasive, cost-effective, respects privacy, and is capable of performing in various acoustic conditions, making it an ideal tool for home-based exercise monitoring and feedback. EchoMotion’s ability to analyze the exercise pose estimations provides valuable insights for users, trainers, and clinicians, enhancing the quality of remote exercise programs.
Acoustic sensing, pose estimation, homebased monitoring
1772-1776
Mosuily, Mohammed
9bd9045b-dff3-4545-b9ab-22960ca5c92f
Chauhan, Jagmohan
831a12dc-6df9-40ea-8bb3-2c5da8882804
Mosuily, Mohammed
9bd9045b-dff3-4545-b9ab-22960ca5c92f
Chauhan, Jagmohan
831a12dc-6df9-40ea-8bb3-2c5da8882804

Mosuily, Mohammed and Chauhan, Jagmohan (2025) EchoMotion: enhancing exercise analysis with acoustic sensing. European Signal Processing Conference, Isola delle Femmine, Palermo, Italy. 08 Sep - 12 Oct 2025. pp. 1772-1776 .

Record type: Conference or Workshop Item (Paper)

Abstract

EchoMotion revolutionizes exercise monitoring with its innovative acoustic-based smart speaker system, designed to perform human pose estimation using inaudible acoustic signals. Leveraging a combination of acoustic features, deep learning techniques, and a custom loss function, the system transforms acoustic reflections from the human body into precise 3D pose estimations. Ground truth data were recorded using the Microsoft Azure Kinect DK, a depth-sensing camera used for evaluation. Data were collected from 22 participants performing five fast-movement cardio exercises in both home and lab environments, yielding over 11 hours of synchronized acoustic and ground truth data. EchoMotion achieved a low Mean Absolute Error (MAE) of 0.59 mm, demonstrating superior accuracy for fast movement exercises compared to the reported MAE range of 2.8 mm to 96 mm in SOTA works, which also focus on slow movements. Our system is non-invasive, cost-effective, respects privacy, and is capable of performing in various acoustic conditions, making it an ideal tool for home-based exercise monitoring and feedback. EchoMotion’s ability to analyze the exercise pose estimations provides valuable insights for users, trainers, and clinicians, enhancing the quality of remote exercise programs.

This record has no associated files available for download.

More information

Published date: 10 September 2025
Venue - Dates: European Signal Processing Conference, Isola delle Femmine, Palermo, Italy, 2025-09-08 - 2025-10-12
Keywords: Acoustic sensing, pose estimation, homebased monitoring

Identifiers

Local EPrints ID: 506783
URI: http://eprints.soton.ac.uk/id/eprint/506783
PURE UUID: e46b9c65-6e22-4935-9038-e4ff0e3968cc
ORCID for Mohammed Mosuily: ORCID iD orcid.org/0009-0003-6563-7440

Catalogue record

Date deposited: 18 Nov 2025 17:49
Last modified: 20 Nov 2025 03:01

Export record

Contributors

Author: Mohammed Mosuily ORCID iD
Author: Jagmohan Chauhan

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×