Topics in contextualised attention embeddings
Topics in contextualised attention embeddings
Contextualised word vectors obtained via pre-trained language models encode a variety of knowledge that has already been exploited in applications. Complementary to these language models are probabilistic topic models that learn thematic patterns from the text. Recent work has demonstrated that conducting clustering on the word-level contextual representations from a language model emulates word clusters that are discovered in latent topics of words from Latent Dirichlet Allocation. The important question is how such topical word clusters are automatically formed, through clustering, in the language model when it has not been explicitly designed to model latent topics. To address this question, we design different probe experiments. Using BERT and DistilBERT, we find that the attention framework plays a key role in modelling such word topic clusters. We strongly believe that our work paves way for further research into the relationships between probabilistic topic models and pre-trained language models.
221-238
Talebpour, Mozhgan
c7a55ebd-06b8-4379-b44c-5cca9083cecd
Garcia Seco De Herrero, Alba
a8bda28e-2d29-4989-8fcc-f6873d6860d0
Jameel, Mohammad Shoaib
ae3c588e-4a59-43d9-af41-ea30d7caaf96
2023
Talebpour, Mozhgan
c7a55ebd-06b8-4379-b44c-5cca9083cecd
Garcia Seco De Herrero, Alba
a8bda28e-2d29-4989-8fcc-f6873d6860d0
Jameel, Mohammad Shoaib
ae3c588e-4a59-43d9-af41-ea30d7caaf96
Talebpour, Mozhgan, Garcia Seco De Herrero, Alba and Jameel, Mohammad Shoaib
(2023)
Topics in contextualised attention embeddings.
Kamps, Jaap, Goeuriot, Lorraine, Crestani, Fabio, Maistro, Maria, Joho, Hideo, Davis, Brian, Gurrin, Cathal, Caputo, Annalina and Kruschwitz, Udo
(eds.)
In Advances in Information Retrieval - 45th European Conference on Information Retrieval, ECIR 2023, Proceedings: Proceedings Part II.
vol. 13981 LNCS,
Springer Cham.
.
(doi:10.1007/978-3-031-28238-6_15).
Record type:
Conference or Workshop Item
(Paper)
Abstract
Contextualised word vectors obtained via pre-trained language models encode a variety of knowledge that has already been exploited in applications. Complementary to these language models are probabilistic topic models that learn thematic patterns from the text. Recent work has demonstrated that conducting clustering on the word-level contextual representations from a language model emulates word clusters that are discovered in latent topics of words from Latent Dirichlet Allocation. The important question is how such topical word clusters are automatically formed, through clustering, in the language model when it has not been explicitly designed to model latent topics. To address this question, we design different probe experiments. Using BERT and DistilBERT, we find that the attention framework plays a key role in modelling such word topic clusters. We strongly believe that our work paves way for further research into the relationships between probabilistic topic models and pre-trained language models.
Text
Topics in Contextualised Attention Embeddings
- Accepted Manuscript
More information
e-pub ahead of print date: 17 March 2023
Published date: 2023
Additional Information:
Publisher Copyright:
© 2023, The Author(s), under exclusive license to Springer Nature Switzerland AG.
Venue - Dates:
The 45th European Conference on Information Retrieval (ECIR'23), , Dublin, Ireland, 2023-04-02 - 2023-04-06
Identifiers
Local EPrints ID: 478782
URI: http://eprints.soton.ac.uk/id/eprint/478782
ISSN: 0302-9743
PURE UUID: 7331a9e3-4322-447a-a01f-1f613501f79d
Catalogue record
Date deposited: 10 Jul 2023 16:42
Last modified: 05 Jun 2024 18:40
Export record
Altmetrics
Contributors
Author:
Mozhgan Talebpour
Author:
Alba Garcia Seco De Herrero
Author:
Mohammad Shoaib Jameel
Editor:
Jaap Kamps
Editor:
Lorraine Goeuriot
Editor:
Fabio Crestani
Editor:
Maria Maistro
Editor:
Hideo Joho
Editor:
Brian Davis
Editor:
Cathal Gurrin
Editor:
Annalina Caputo
Editor:
Udo Kruschwitz
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics