2022
DOI: 10.1109/tpami.2021.3124805
|View full text |Cite
|
Sign up to set email alerts
|

Bringing Light Into the Dark: A Large-Scale Evaluation of Knowledge Graph Embedding Models Under a Unified Framework

Abstract: The heterogeneity in recently published knowledge graph embedding models' implementations, training, and evaluation has made fair and thorough comparisons difficult. To assess the reproducibility of previously published results, we re-implemented and evaluated 21 models in the PyKEEN software package. In this paper, we outline which results could be reproduced with their reported hyper-parameters, which could only be reproduced with alternate hyper-parameters, and which could not be reproduced at all, as well … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

0
41
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 58 publications
(41 citation statements)
references
References 32 publications
(82 reference statements)
0
41
0
Order By: Relevance
“…Tensor factorization and tensor networks [12] are wellknown factorization techniques used for knowledge graph embedding with semantic matching interactions [13]. In a tensor representation, each triplet of the knowledge graph is represented by the value, X ijk of a three dimensional tensor, X ∈ R n×n×m the triplet (i, j, k) exists, or can be an arbitrary non-negative value that measures the strength of the k-relation between the iand j-entities.…”
Section: Introductionmentioning
confidence: 99%
“…Tensor factorization and tensor networks [12] are wellknown factorization techniques used for knowledge graph embedding with semantic matching interactions [13]. In a tensor representation, each triplet of the knowledge graph is represented by the value, X ijk of a three dimensional tensor, X ∈ R n×n×m the triplet (i, j, k) exists, or can be an arbitrary non-negative value that measures the strength of the k-relation between the iand j-entities.…”
Section: Introductionmentioning
confidence: 99%
“…On the other hand, QE bypasses the need for a database or query engine and performs reasoning directly in a latent space by computing a similarity score between the query representation and entity representations 1 . A query representation is obtained by processing its equivalent logical formula where joins become intersections (∧), and variables are existentially quantified (∃).…”
Section: Introductionmentioning
confidence: 99%
“…Instead, CQD decomposes a query in a sequence of reasoning steps and performs a beam search in a latent space of KG embeddings models pre-trained on a simple 1-hop link prediction task. A particular novelty of this approach is that no end-to-end training on complex queries is required, and any trained embedding model of the existing abundance [1,16] can be employed as is.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Existing software packages that provide implementations for different KGEMs usually lack entire composability: model architectures (or interaction models), training approaches, loss functions, and the usage of explicit inverse relation cannot arbitrarily be combined. The full composability of KGEMs is fundamental for assessing the performance of KGEMs because it allows assessing single components individually on the model's performance instead of attributing a performance increase solely to the model architecture, which is misleading (Ruffinelli et al, 2020;Ali et al, 2020). Besides, often only limited functionalities are provided, e.g., a small number of KGEMs are supported, or functionalities such as HPO are missing.…”
Section: Introductionmentioning
confidence: 99%