The University of Southampton
University of Southampton Institutional Repository

Analysis of complex-valued neural networks for audio source localisation

Analysis of complex-valued neural networks for audio source localisation
Analysis of complex-valued neural networks for audio source localisation
An increasing number of studies recently have dealt with novel methods for locating the source of a wide range of acoustic events. Applications such as teleconferencing, human-robot interaction, source separation or speech recognition can make use of the Direction of Arrival (DoA) of a sound source to improve their results. Since most of the newer localisation methods proposed recently make use of neural networks in their estimation of source position, the work will focus on comparing the use of complex-valued neural networks with real-valued networks for localising sound source in different scenarios. The data used for the comparison will be simulated using a number of geometrical microphone arrangements with the acoustic sources placed in the far-field of the microphone arrays. The simulated data used in combination with the chosen range of microphone arrangements will be used to investigate the localisation limits of the chosen algorithms. A particular objective of the work will be to evaluate any potential advantages in using complex-valued neural networks rather than real valued networks.
1478-6095
78-87
Curran Associates, Inc.
Paul, Vlad Stefan
a643f880-7e70-4ae0-a27b-4e77c3c451de
Nelson, Philip
5c6f5cc9-ea52-4fe2-9edf-05d696b0c1a9
Paul, Vlad Stefan
a643f880-7e70-4ae0-a27b-4e77c3c451de
Nelson, Philip
5c6f5cc9-ea52-4fe2-9edf-05d696b0c1a9

Paul, Vlad Stefan and Nelson, Philip (2022) Analysis of complex-valued neural networks for audio source localisation. In Reproduced Sound 2022: Auralisation and Personalisation-Beyond Reality. vol. 44, Curran Associates, Inc. pp. 78-87 .

Record type: Conference or Workshop Item (Paper)

Abstract

An increasing number of studies recently have dealt with novel methods for locating the source of a wide range of acoustic events. Applications such as teleconferencing, human-robot interaction, source separation or speech recognition can make use of the Direction of Arrival (DoA) of a sound source to improve their results. Since most of the newer localisation methods proposed recently make use of neural networks in their estimation of source position, the work will focus on comparing the use of complex-valued neural networks with real-valued networks for localising sound source in different scenarios. The data used for the comparison will be simulated using a number of geometrical microphone arrangements with the acoustic sources placed in the far-field of the microphone arrays. The simulated data used in combination with the chosen range of microphone arrangements will be used to investigate the localisation limits of the chosen algorithms. A particular objective of the work will be to evaluate any potential advantages in using complex-valued neural networks rather than real valued networks.

This record has no associated files available for download.

More information

Submitted date: 2022
Published date: 1 December 2022
Additional Information: Publisher Copyright:© 2022 Institute of Acoustics. All rights reserved.
Venue - Dates: Reproduced Sound 2022: Auralisation and Personalisation-Beyond Reality, , Bristol, United Kingdom, 2022-11-15 - 2022-11-17

Identifiers

Local EPrints ID: 476879
URI: http://eprints.soton.ac.uk/id/eprint/476879
ISSN: 1478-6095
PURE UUID: d5116c15-a7ec-4944-867d-f6389a918577
ORCID for Vlad Stefan Paul: ORCID iD orcid.org/0000-0002-5562-6102
ORCID for Philip Nelson: ORCID iD orcid.org/0000-0002-9563-3235

Catalogue record

Date deposited: 18 May 2023 16:56
Last modified: 17 Mar 2024 03:59

Export record

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×