Shrinking embeddings for hyper-relational knowledge graphs
Shrinking embeddings for hyper-relational knowledge graphs
Link prediction on knowledge graphs (KGs) has been extensively studied on binary relational KGs, wherein each fact is represented by a triple. A significant amount of important knowledge, however, is represented by hyperrelational facts where each fact is composed ofa primal triple and a set of qualifiers comprising a key-value pair that allows for expressing more complicated semantics. Although some recent works have proposed to embed hyper-relational KGs, these methods fail to capture essential inference patterns of hyperrelational facts such as qualifier monotonicity, qualifier implication, and qualifier mutual exclusion, limiting their generalization capability. To unlock this, we present ShrinkE, a geometric hyper-relational KG embedding method aiming to explicitly model these patterns. ShrinkE models the primal triple asa spatial-functional transformation from the head into a relation-specific box. Each qualifier “shrinks” the box to narrow down the possible answer set and, thus, realizes qualifier monotonicity. The spatial relationships between the qualifier boxes allow for modeling core inference patterns of qualifiers such as implication and mutual exclusion. Experimental results demonstrate ShrinkE’s superiority on three benchmarks of hyper-relational KGs.
Xiong, Bo
d8c3ce0a-07ac-43f8-bd67-f230c6cbc1ec
Nayyeri, Mojtaba
476e5009-e6fc-45e6-ac7f-c07fe0898632
Pan, Shirui
5defae6b-0217-4d18-8c99-8ef0dd7665ca
Staab, Steffen
bf48d51b-bd11-4d58-8e1c-4e6e03b30c49
Xiong, Bo
d8c3ce0a-07ac-43f8-bd67-f230c6cbc1ec
Nayyeri, Mojtaba
476e5009-e6fc-45e6-ac7f-c07fe0898632
Pan, Shirui
5defae6b-0217-4d18-8c99-8ef0dd7665ca
Staab, Steffen
bf48d51b-bd11-4d58-8e1c-4e6e03b30c49
Xiong, Bo, Nayyeri, Mojtaba, Pan, Shirui and Staab, Steffen
(2023)
Shrinking embeddings for hyper-relational knowledge graphs.
The 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023), , Toronto, Canada.
09 - 14 Jul 2023.
13 pp
.
(In Press)
Record type:
Conference or Workshop Item
(Paper)
Abstract
Link prediction on knowledge graphs (KGs) has been extensively studied on binary relational KGs, wherein each fact is represented by a triple. A significant amount of important knowledge, however, is represented by hyperrelational facts where each fact is composed ofa primal triple and a set of qualifiers comprising a key-value pair that allows for expressing more complicated semantics. Although some recent works have proposed to embed hyper-relational KGs, these methods fail to capture essential inference patterns of hyperrelational facts such as qualifier monotonicity, qualifier implication, and qualifier mutual exclusion, limiting their generalization capability. To unlock this, we present ShrinkE, a geometric hyper-relational KG embedding method aiming to explicitly model these patterns. ShrinkE models the primal triple asa spatial-functional transformation from the head into a relation-specific box. Each qualifier “shrinks” the box to narrow down the possible answer set and, thus, realizes qualifier monotonicity. The spatial relationships between the qualifier boxes allow for modeling core inference patterns of qualifiers such as implication and mutual exclusion. Experimental results demonstrate ShrinkE’s superiority on three benchmarks of hyper-relational KGs.
Text
ShrinkE
- Accepted Manuscript
More information
Accepted/In Press date: 1 June 2023
Venue - Dates:
The 61st Annual Meeting of the Association for Computational Linguistics (ACL 2023), , Toronto, Canada, 2023-07-09 - 2023-07-14
Identifiers
Local EPrints ID: 478321
URI: http://eprints.soton.ac.uk/id/eprint/478321
PURE UUID: 347f7793-51d7-4df1-a46a-6e2301e031f6
Catalogue record
Date deposited: 28 Jun 2023 16:31
Last modified: 17 Mar 2024 03:38
Export record
Contributors
Author:
Bo Xiong
Author:
Mojtaba Nayyeri
Author:
Shirui Pan
Author:
Steffen Staab
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics