Abstract

The task of entity alignment between knowledge graphs aims to find entities in two knowledge graphs that represent the same real-world entity. Recently, embedding-based models are proposed for this task. Such models are built on top of a knowledge graph embedding model that learns entity embeddings to capture the semantic similarity between entities in the same knowledge graph. We propose to learn embeddings that can capture the similarity between entities in different knowledge graphs. Our proposed model helps align entities from different knowledge graphs, and hence enables the integration of multiple knowledge graphs. Our model exploits large numbers of attribute triples existing in the knowledge graphs and generates attribute character embeddings. The attribute character embedding shifts the entity embeddings from two knowledge graphs into the same space by computing the similarity between entities based on their attributes. We use a transitivity rule to further enrich the number of attributes of an entity to enhance the attribute character embedding. Experiments using real-world knowledge bases show that our proposed model achieves consistent improvements over the baseline models by over 50% in terms of hits@1 on the entity alignment task.

1 Introduction

2 Related Work

2.1. String-Similarity-based Entity Alignment

  1. use string similarity as main alignment tool
    1. LIMES (Ngomo and Auer 2011)
    2. RDF-AI (Scharffe, Yanbin, and Zhou 2009)
    3. SILK (Volz et al., 2009)
  2. use graph similarity to improve performance
    1. LD-Mapper (Raimond, Sutton, and Sandler 2008)
    2. RuleMiner (Niu et al. 2012)
    3. HolisticEM (Pershina, Yakout, and Chakrabarti 2015)

2.2. Embedding-based Entity Alignment

[KG completion]

  1. translation based KGE models
    1. distributed representation that separate the relationship vector space from entity vector space: TransE (Bordes et al. 2013), TransH (Wang et al. 2014), TransR (Lin et al. 2015), and TransD (Ji et al. 2015), ...
    2. additional information along with the relationship triples to compute entity embeddings: DKRL (Xie et al. 2016), TEKE (Wang and Li 2016)
  2. non-translation based KGE models
    1. tensor-based factorization and representating relationships with matrices: RESCAL (Nickel, Tresp, and Kriegel 2012) and HolE (Nickel, Rosasco, and Poggio 2016)
    2. bilinear tensor operator to represent each relationship and jointly models head and tail entity embeddings: NTN (Socher et al. 2013)

[Entity alignment]

  1. Chen et al. (2017a; 2017b):