Augmenting pre-trained language models with audio feature embedding for argumentation mining in political debates
Augmenting pre-trained language models with audio feature embedding for argumentation mining in political debates
The integration of multimodality in natural language processing (NLP) tasks seeks to exploit the complementary information contained in two or more modalities, such as text, audio and video. This paper investigates the integration of often under-researched audio features with text, using the task of argumentation mining (AM) as a case study. We take a previously reported dataset and present an audio-enhanced version (the Multimodal USElecDeb60To16 dataset). We report the performance of two text models based on BERT and GloVe embeddings, one audio model (based on CNN and Bi-LSTM) and multimodal combinations, on a dataset of 28,850 utterances. The results show that multimodal models do not outperform text-based models when using the full dataset. However, we show that audio features add value in fully supervised scenarios with limited data. We find that when data is scarce (e.g. with 10% of the original dataset) multimodal models yield improved performance, whereas text models based on BERT considerably decrease performance. Finally, we conduct a study with artificially generated voices and an ablation study to investigate the importance of different audio features in the audio models.
Mestre, Rafael
33721a01-ab1a-4f71-8b0e-abef8afc92f3
Middleton, Stuart E.
404b62ba-d77e-476b-9775-32645b04473f
Ryan, Matt
f07cd3e8-f3d9-4681-9091-84c2df07cd54
Gheasi, Masood
0e1a0af4-3f82-4498-a5e5-4f7f7618d68e
Norman, Timothy
663e522f-807c-4569-9201-dc141c8eb50d
Zhu, Jiatong
52569115-5d72-4fc0-8876-a66b991ed209
17 March 2023
Mestre, Rafael
33721a01-ab1a-4f71-8b0e-abef8afc92f3
Middleton, Stuart E.
404b62ba-d77e-476b-9775-32645b04473f
Ryan, Matt
f07cd3e8-f3d9-4681-9091-84c2df07cd54
Gheasi, Masood
0e1a0af4-3f82-4498-a5e5-4f7f7618d68e
Norman, Timothy
663e522f-807c-4569-9201-dc141c8eb50d
Zhu, Jiatong
52569115-5d72-4fc0-8876-a66b991ed209
Mestre, Rafael, Middleton, Stuart E., Ryan, Matt, Gheasi, Masood, Norman, Timothy and Zhu, Jiatong
(2023)
Augmenting pre-trained language models with audio feature embedding for argumentation mining in political debates.
In Findings of the 17th conference on European Chapter of the Association for Computational Linguistics (EACL).
Record type:
Conference or Workshop Item
(Paper)
Abstract
The integration of multimodality in natural language processing (NLP) tasks seeks to exploit the complementary information contained in two or more modalities, such as text, audio and video. This paper investigates the integration of often under-researched audio features with text, using the task of argumentation mining (AM) as a case study. We take a previously reported dataset and present an audio-enhanced version (the Multimodal USElecDeb60To16 dataset). We report the performance of two text models based on BERT and GloVe embeddings, one audio model (based on CNN and Bi-LSTM) and multimodal combinations, on a dataset of 28,850 utterances. The results show that multimodal models do not outperform text-based models when using the full dataset. However, we show that audio features add value in fully supervised scenarios with limited data. We find that when data is scarce (e.g. with 10% of the original dataset) multimodal models yield improved performance, whereas text models based on BERT considerably decrease performance. Finally, we conduct a study with artificially generated voices and an ablation study to investigate the importance of different audio features in the audio models.
Text
2023.findings-eacl.21
- Version of Record
Text
mestre_2023_MultimodalUSElecDeb60to16
More information
Published date: 17 March 2023
Identifiers
Local EPrints ID: 475962
URI: http://eprints.soton.ac.uk/id/eprint/475962
PURE UUID: ecbdf994-e027-4fd8-8476-fc98b6c2b383
Catalogue record
Date deposited: 03 Apr 2023 16:33
Last modified: 17 Mar 2024 04:06
Export record
Contributors
Author:
Jiatong Zhu
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics