The University of Southampton
University of Southampton Institutional Repository

A framework for intelligent metadata adaptation in object-based audio

A framework for intelligent metadata adaptation in object-based audio
A framework for intelligent metadata adaptation in object-based audio
Object-based audio can be used to customize, personalize, and optimize audio reproduction depending on the specific listening scenario. To investigate and exploit the benefits of object-based audio, a framework for intelligent metadata adaptation was developed. The framework uses detailed semantic metadata that describes the audio objects, the loudspeakers, and the room. It features an extensible software tool for real-time metadata adaptation that can incorporate knowledge derived from perceptual tests and/or feedback from perceptual meters to drive adaptation and facilitate optimal rendering. One use case for the system is demonstrated through a rule-set (derived from perceptual tests with experienced mix engineers) for automatic adaptation of object levels and positions when rendering 3D content to two- and five-channel systems.
Metadata, Semantics, Software testing, Three dimensional computer graphics, 3D content, Audio reproduction, Automatic adaptation, Channel systems, Object based, Perceptual test, Real time, Semantic metadata, Loudspeakers
349-359
Woodcock, James
5f2edb7e-7a42-4080-9379-7090bcc33234
Francombe, Jon
b214bf4d-72d7-4ad6-8b05-8a1b56dba6b5
Franck, A.
fa179b73-6a83-4c42-b300-81f1dfe9ef6d
Coleman, P.
7be43b5a-2298-4af5-b0b6-650015efe666
Hughes, R.
938ac53c-edfb-4b20-b99b-31e2cbd18768
Kim, H.
2c7c135c-f00b-4409-acb2-85b3a9e8225f
Liu, Q.
25b12223-59c0-4c54-915b-484635afa08a
Menzies, D.
0cc76abc-8a10-4b7f-96e5-56eceb0b2c5c
Simón Gálvez, M.F.
777da25f-86fc-4a22-8ff1-ac2cbbfe27ae
Tang, Y.
a0483c57-271f-4ff8-8a45-d476a7d3067e
Brookes, T.
ffd7dc65-ac9a-456e-bc9c-dfee06239835
Davies, W.J.
13a1264f-5940-4689-ba7e-d1c1ca7f313a
Fazenda, B.M.
7a9788fe-6f09-4c64-ba61-acebe302a742
Mason, R.
6984b44c-4077-4557-9264-9eeb9d60bb6a
Cox, T.J.
13f915fb-1615-4913-882b-b8bfb8c6f78a
Fazi, F.M.
e5aefc08-ab45-47c1-ad69-c3f12d07d807
Jackson, P.J.B.
6e3f3079-448f-42b1-a880-1a74af6991d8
Pike, C.
26d7ed04-eb9f-4f5d-a741-d0d0ca1f5dfc
Hilton, A.
12782a55-4c4d-4dfb-a690-62505f6665db
Woodcock, James
5f2edb7e-7a42-4080-9379-7090bcc33234
Francombe, Jon
b214bf4d-72d7-4ad6-8b05-8a1b56dba6b5
Franck, A.
fa179b73-6a83-4c42-b300-81f1dfe9ef6d
Coleman, P.
7be43b5a-2298-4af5-b0b6-650015efe666
Hughes, R.
938ac53c-edfb-4b20-b99b-31e2cbd18768
Kim, H.
2c7c135c-f00b-4409-acb2-85b3a9e8225f
Liu, Q.
25b12223-59c0-4c54-915b-484635afa08a
Menzies, D.
0cc76abc-8a10-4b7f-96e5-56eceb0b2c5c
Simón Gálvez, M.F.
777da25f-86fc-4a22-8ff1-ac2cbbfe27ae
Tang, Y.
a0483c57-271f-4ff8-8a45-d476a7d3067e
Brookes, T.
ffd7dc65-ac9a-456e-bc9c-dfee06239835
Davies, W.J.
13a1264f-5940-4689-ba7e-d1c1ca7f313a
Fazenda, B.M.
7a9788fe-6f09-4c64-ba61-acebe302a742
Mason, R.
6984b44c-4077-4557-9264-9eeb9d60bb6a
Cox, T.J.
13f915fb-1615-4913-882b-b8bfb8c6f78a
Fazi, F.M.
e5aefc08-ab45-47c1-ad69-c3f12d07d807
Jackson, P.J.B.
6e3f3079-448f-42b1-a880-1a74af6991d8
Pike, C.
26d7ed04-eb9f-4f5d-a741-d0d0ca1f5dfc
Hilton, A.
12782a55-4c4d-4dfb-a690-62505f6665db

Woodcock, James, Francombe, Jon, Franck, A., Coleman, P., Hughes, R., Kim, H., Liu, Q., Menzies, D., Simón Gálvez, M.F., Tang, Y., Brookes, T., Davies, W.J., Fazenda, B.M., Mason, R., Cox, T.J., Fazi, F.M., Jackson, P.J.B., Pike, C. and Hilton, A. (2018) A framework for intelligent metadata adaptation in object-based audio. Audio Engineering Society Conference on Spatial Reproduction, , Tokyo, Japan. 06 - 09 Aug 2018. pp. 349-359 .

Record type: Conference or Workshop Item (Paper)

Abstract

Object-based audio can be used to customize, personalize, and optimize audio reproduction depending on the specific listening scenario. To investigate and exploit the benefits of object-based audio, a framework for intelligent metadata adaptation was developed. The framework uses detailed semantic metadata that describes the audio objects, the loudspeakers, and the room. It features an extensible software tool for real-time metadata adaptation that can incorporate knowledge derived from perceptual tests and/or feedback from perceptual meters to drive adaptation and facilitate optimal rendering. One use case for the system is demonstrated through a rule-set (derived from perceptual tests with experienced mix engineers) for automatic adaptation of object levels and positions when rendering 3D content to two- and five-channel systems.

Text
AES_Tokyo2018_Metadata_Adaptation_temp - Version of Record
Restricted to Repository staff only
Request a copy

More information

Published date: August 2018
Venue - Dates: Audio Engineering Society Conference on Spatial Reproduction, , Tokyo, Japan, 2018-08-06 - 2018-08-09
Keywords: Metadata, Semantics, Software testing, Three dimensional computer graphics, 3D content, Audio reproduction, Automatic adaptation, Channel systems, Object based, Perceptual test, Real time, Semantic metadata, Loudspeakers

Identifiers

Local EPrints ID: 425680
URI: http://eprints.soton.ac.uk/id/eprint/425680
PURE UUID: 689bc2d8-8239-418b-b268-bdb0b5010fc9
ORCID for A. Franck: ORCID iD orcid.org/0000-0002-4707-6710
ORCID for H. Kim: ORCID iD orcid.org/0000-0003-4907-0491
ORCID for D. Menzies: ORCID iD orcid.org/0000-0003-1475-8798
ORCID for F.M. Fazi: ORCID iD orcid.org/0000-0003-4129-1433

Catalogue record

Date deposited: 31 Oct 2018 17:30
Last modified: 13 Jun 2024 01:46

Export record

Contributors

Author: James Woodcock
Author: Jon Francombe
Author: A. Franck ORCID iD
Author: P. Coleman
Author: R. Hughes
Author: H. Kim ORCID iD
Author: Q. Liu
Author: D. Menzies ORCID iD
Author: Y. Tang
Author: T. Brookes
Author: W.J. Davies
Author: B.M. Fazenda
Author: R. Mason
Author: T.J. Cox
Author: F.M. Fazi ORCID iD
Author: P.J.B. Jackson
Author: C. Pike
Author: A. Hilton

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×