The search functionality is under construction.
The search functionality is under construction.

Open Access
Type-Enhanced Ensemble Triple Representation via Triple-Aware Attention for Cross-Lingual Entity Alignment

Zhishuo ZHANG, Chengxiang TAN, Xueyan ZHAO, Min YANG

  • Full Text Views

    190

  • Cite this
  • Free PDF (1.7MB)

Summary :

Entity alignment (EA) is a crucial task for integrating cross-lingual and cross-domain knowledge graphs (KGs), which aims to discover entities referring to the same real-world object from different KGs. Most existing embedding-based methods generate aligning entity representation by mining the relevance of triple elements, paying little attention to triple indivisibility and entity role diversity. In this paper, a novel framework named TTEA - Type-enhanced Ensemble Triple Representation via Triple-aware Attention for Cross-lingual Entity Alignment is proposed to overcome the above shortcomings from the perspective of ensemble triple representation considering triple specificity and diversity features of entity role. Specifically, the ensemble triple representation is derived by regarding relation as information carrier between semantic and type spaces, and hence the noise influence during spatial transformation and information propagation can be smoothly controlled via specificity-aware triple attention. Moreover, the role diversity of triple elements is modeled via triple-aware entity enhancement in TTEA for EA-oriented entity representation. Extensive experiments on three real-world cross-lingual datasets demonstrate that our framework makes comparative results.

Publication
IEICE TRANSACTIONS on Information Vol.E107-D No.9 pp.1182-1191
Publication Date
2024/09/01
Publicized
2024/05/22
Online ISSN
1745-1361
DOI
10.1587/transinf.2023EDP7234
Type of Manuscript
PAPER
Category
Artificial Intelligence, Data Mining

1.  Introduction

Knowledge Graphs (KGs) have been powerful platform for organizing plain text into regularized data formatted as (head entity,relation,tail entity), such as (Beijing, Capital_of, China) denoting that Beijing is the capital of China, in which Capital_of holds the directional relation from head entity Beijing to tail entity China. However, most of KGs are independently constructed in different languages or domains. Although these cross-lingual or cross-domain KGs such as DBpedia [1], YAGO [2] and ConceptNet [3] have been widely applied in many real-world scenarios, such as finance [4], medical care [5], [6], and artificial intelligence [7]-[9], the structural heterogeneity and rule specificity among these KGs make the knowledge cannot be effectively integrated. Entity alignment (EA) provides a crucial step for information fusion, which aims to discover equivalent entities from different KGs.

Recently, embedding-based methods have attracted wide attention, which embed entity and relation by encoding them into latent vector spaces and measure the embedding distance for EA [10]-[13]. There have been many efforts to obtain excellent representation of entity and relation for EA, which can be roughly divided into two categories according to perspectives: Trans-based methods and GNNs-based methods.

Trans-based methods treat element interaction as a translation process \(\boldsymbol{h+r \approx t}\) for a triple \(\boldsymbol{(h,r,t)}\). These methods [14]-[16] are effective and simple, but unable to form the complete representation of triple elements as the internal correlation is complex and indescribable. Besides, noise will be introduced inevitably in the mapping process between different latent spaces. GNNs-based methods fall into two categories, GCNs-based and GATs-based. The former usually reflect EA via neighbor alignment and topology structure [17]-[19], and the latter integrate the surrounding information to enhance embedding [20], [21]. Although these methods can effectively improve performance by fusing neighbor information, they usually focused on the element itself and rarely considered the specificity of ensemble triple and role diversity, which may hold rich semantic information. For details, in case the entity pair \(\left(e^{1}, e_{1}\right)\) includes two entities to be aligned in Fig. 1. Then EA-oriented entity representations of \(e^{1}\) in KG1 and \(e_{1}\) in KG2 are optimally generated during training. On the one hand, the triple specificity tries to illustrate entity semantic from a triple perspective rather than in a element view. In the training process, the triple \(\left(e^{1}, r1 ,e^{2}\right)\) and \(\left(e^{1}, r2 ,e^{3}\right)\) should play important roles for generating EA-oriented entity representation of \(e^{1}\), and the triples \(\left(e_{1}, R2 ,e_{3}\right)\) and \(\left(e_{2}, R1 ,e_{1}\right)\) have the responsibility for the EA-oriented representation of \(e_{1}\) as the same. On the other hand, the role diversity attempts to distinguish the influence of elements with different roles on triple. Specifically, for the entity \(e_{1}\) attaching to more than one triple in KG2. It is intuitive that the influence of \(e_{1}\) on triple \(\left(e_{1}, R2 ,e_{3}\right)\) is inconsistent with the influence of \(e_{1}\) on triple \(\left(e_{2}, R1 ,e_{1}\right)\), in which the entity \(e_{1}\) plays a role as head entity in the former while tail entity in the latter. Furthermore, there may be multiple relations holding different types between head entity and tail entity, as the entity pairs \(\left(e^{3}, e^{5}\right)\) in KG1 and \(\left(e_{3}, e_{5}\right)\) in KG2 show.

Fig. 1  A toy example of entity-role diversity and multiple relations between entities. The blue solid lines between entities in KG1 and KG2 refer to the alignments.

To address the above shortcomings, we propose TTEA - Type-enhanced Ensemble Triple Representation via Triple-aware Attention for Cross-lingual Entity Alignment in this paper with the intuitive assumption that relation is generic in semantic space and type space for a specific triple, which is capable of fully utilizing triple specificity and role diversity. Considering that triple elements are indivisible, TTEA introduces a type-enhanced ensemble triple representation module to capture semantic and type information while preserving triple specificity. In terms of multiple relations and roles in a triple, we design a triple-aware entity enhancement mechanism to obtain cycle co-enhanced head-aware and tail-aware entity embedding. RAGA [21] is the closest work with this paper, the difference is that TTEA tries to exploit ensemble triple specificity and role diversity of head and tail entities while RAGA tackle EA from element perspective only. Experimental results on three cross-lingual KGs prove that TTEA made comparative results with baselines1.

In summary, our main contributions are as follows:

\(\bullet\) We provide a novel perspective to regard relation as information carrier during spatial transformation, which is capable to effectively alleviate the noise introduced during mapping.

\(\bullet\) We propose a novel EA framework which sufficiently utilizes triple specificity and role diversity via ensemble triple representation and triple-aware entity enhancement.

\(\bullet\) Extensive experiments conducted on public datasets demonstrate that TTEA makes comparative performance among corresponding EA baseline methods.

2.  Related Work

Embedding-based methods have been the mainstream formula for EA in recent years. These methods can be roughly divided into two categories: Trans-based embedding methods and GNNs-based embedding methods. Trans-based embedding methods: A line of Trans-based works embed entity and relation in different latent spaces for different KGs [16], [22]-[25]. The second line of Trans-based works embed entity and relation from different KGs into a unified latent space via parameters sharing [26] and extending aligned relation [27]. Then these methods utilize TransE [28] and its variants [14], [15] to model internal correlations among triple elements as a linear equation to learn EA-oriented entity representation. However, this formula cannot comprehensively mining the complex interaction among triple components with the simple mode. GNNs-based embedding methods: GNNs-based embedding methods have obtained great improvement in EA. GCN-Align [10] is the first work to enhance entity embedding via GCNs, following which GCNs-based methods are extended to aggregate neighbor information. Some works use neighbor entities and relations alignment to tackle EA with the assumption that equivalent entities sharing approximate neighbors [29], and some other related efforts utilize topology structure to reinforce entity embedding via GCNs [12], [19], [30]-[32]. While GCN-based methods treat neighbor relation equally, which is unreasonable. Thus GATs-based methods are designed considering different neighbor entities contribute different importance. Some works adopt neighbor attention for entity embedding [18], [33]-[36], in which to our best knowledge, RAGA [21] achieves state-of-the-art results by modeling correlation between entity and relation in embedding-based methods. And others utilize cross-KG attention to spread neighbor information of aligned pairs [20], [37], [38]. Existing GNNs-based methods have effectively improved the performance of EA, but not considered diversity of entities roles and the multi-level representation of ensemble triple. Our work in this paper is also a GNNs-based method from the perspective of ensemble triple and element roles diversity.

Besides, there are some techniques for EA. Some works tried to mine EA-oriented message contained in entity name, such as [39], [40] explored the effect of different levels of entity name on EA, [41] proposed a tree-view label propagation mechanism. And external resources such as entity attributes, images, and other string information are utilized [17], [19], [42], [42]-[45]. However, on the one hand, the external sources are not available or inauthentic in some real scenarios such as user supervision cross social medias. On the other hand, structural heterogeneity among different KGs is the main issue being addressed in this paper. Therefore, unlike previous works [39], [46], external sources are not utilized in TTEA. Instead, it relies on topology structure and neighbor information for solving heterogeneity at its core. What is more, to handle the insufficient seed pairs, Boostrapping learning [47] and iterative training strategy [23] are introduced. Specifically, bi-directional iterative training strategy [48] are widely applied recently [11], [13], [49], [50], which is also adopted in TTEA for improving performance.

3.  Problem Formulation

A KG could be formalized as \(KG = (E, R, T)\), where \(E\) and \(R\) are the sets of entity and relation respectively, \(T \subset E \times R \times E\) is the set of relational triple. Given two cross-lingual KGs, \(KG1=(E1, R1, T1)\) and \(KG2=(E2, R2, T2)\), the task of EA is defined as identifying entity pairs composed of entities from \(KG1\) and \(KG2\) based on a set of seed entity pairs in this paper, in which each entity pair comprises two entities, originating from distinct KGs, that refer to the identical real-world object. The seed entity pairs are known aligned and can be denoted as \(S=\{(e_1, e_2)|e_1\in E1, e_2 \in E2\}\), where \(e_1\) and \(e_2\) are equivalent objects from \(KG1\) and \(KG2\), respectively.

4.  TTEA Framework

We propose a framework TTEA based on type-enhanced ensemble triple representation and triple-aware entity enhancement via triple-aware attention mechanism. The overall architecture of TTEA is illustrated as Fig. 2, which mainly consists of four parts: Topology Structure Aggregation, Type-enhanced Ensemble Triple Representation, Triple-aware Entity Enhancement and Entity Alignment Strategy. Entity name-based embedding is enhanced via structural information for initialization in the first part, after that the ensemble triple representation with specificity is generated in Type-enhanced Ensemble Triple Representation part. Then, triple-aware representations of head and tail entities are respectively obtained and are circularly reinforced by each other in Triple-aware Entity Enhancement module. Finally, in the Entity Alignment Strategy part, the bi-direction iterative strategy is applied to enlarge seed pairs, meanwhile the entity embedding and the parameters are updated via back-propagation.

Fig. 2  An overall of TTEA framework, which consists of the following components: 1) topology structure aggregation for integrating neighbors, 2) type-enhanced ensemble triple representation for generating ensemble triple and fusing type information, and 3) triple-aware entity enhancement for obtaining EA-oriented entity representation.

4.1  Topology Structure Aggregation

We firstly expand relation as a combination of original-relation, reverse-relation and self-relation to fully describe topology structure in KGs. We also use the entity name-based embedding as the initialized representation as RAGA does, following which a two-layer GCNs with Highway Networks [51] are deployed to aggregate topological information while preserving entity primary semantic. The \(l\)-th Highway-GCN layer is formulated as:

\[\begin{align} & \boldsymbol{X}^{l+1}_{gcn}=\operatorname{ReLU}\left(\boldsymbol{\tilde{D}}^{-\frac{1}{2}} \boldsymbol{\tilde{A}} \boldsymbol{\tilde{D}}^{-\frac{1}{2}} \boldsymbol{X}^{l}\right) \tag{1} \\ & T\left(\boldsymbol{X}^{l}\right) = \sigma\left(\boldsymbol{X}^{l} \boldsymbol{W}^{l}+\boldsymbol{b}^{l}\right) \tag{2} \\ & \boldsymbol{X}^{l+1}=T\left(\boldsymbol{X}^{l}\right) \cdot \boldsymbol{X}^{l+1}_{gcn}+\left(1-T\left(\boldsymbol{X}^{l}\right)\right) \cdot \boldsymbol{X}^{l} \tag{3} \end{align}\]

where \(\boldsymbol{\tilde{A}} = \boldsymbol{A} + \boldsymbol{I}\), \(\boldsymbol{A}\) is the adjacency matrix of relation-expanded graph, \(I\) is the identity matrix, \(\boldsymbol{\tilde{D}}\) is the degree matrix of \(\boldsymbol{\tilde{A}}\) and \(\boldsymbol{X}^{l} \in \mathbb{R}^{n \times d_{e}}\) denotes the input entity embedding in \(l\)-th hidden layer, \(n\) is the number of entities in a KG, \(d_{e}\) is the dimension of entity embedding, \(\boldsymbol{X}^{l+1}_{gcn}\) is the output of \(l\)-th GCN layer, and \(\boldsymbol{X}^{l+1}\) is the output from the \(l\)-th highway network. \(\sigma(\cdot)\) is activation function, \(\cdot\) denotes the element-wise multiplication, \(\boldsymbol{W}^{l}\) and \(\boldsymbol{b}^{l}\) are the weight matrix and bias vector of the input embedding in \(l\)-th hidden layer.

4.2  Type-Enhanced Ensemble Triple Representation

The ensemble triple representation is generated in this module by mining the influence of element and type information on triple in semantic space and type space via ensemble triple attention mechanism.

4.2.1  Ensemble Triple Representation in Semantic Space

To describe relation’s wholeness and multi-relation features, the global relation \(\boldsymbol{\bar{r}} \in \mathbb{R}^{2d_e}\) is computed as the average concatenation of head and tail entities with same relation.

\[\begin{equation*} \boldsymbol{\bar{r}}_{r}=\frac{1}{\left|T_{r}\right|} \sum_{(e_{i}, r, e_{j}) \in T_{r}} \left(\boldsymbol{x}_{i} \oplus \boldsymbol{x}_{j}\right) \tag{4} \end{equation*}\]

where \(\boldsymbol{x}_{i}\) is the embedding of \(e_{i}\), \(\oplus\) denotes the concatenation operation, \(T_{r}\) is the set of triple with a specific relation \(r\) in KG.

Moreover, we utilize triple specificity to alleviate the redundancy and noise during element interaction. The local relation of a specific triple \((e_{i}, r, e_{j})\) is firstly defined as \(\boldsymbol{\tilde{r}}_{irj} = \boldsymbol{x}_{i} \oplus \boldsymbol{x}_{j}\), based on which the triple specificity is denoted as \(\boldsymbol{\tilde{R}}_{irj} = \boldsymbol{x}_{i} \oplus(\boldsymbol{\tilde{r}}_{irj}\boldsymbol{W}_{e}^{l}) \oplus \boldsymbol{x}_{j}\), where \(\boldsymbol{W}_{e}^{l} \in \mathbb{R}^{2d_e \times d_r}\) is the mapping matrix, \(d_r\) is the relation dimension.

Then, we design a three-view ensemble triple attention mechanism to legitimately incorporate global triple features, which includes head-aware attention, tail-aware attention and relation-aware attention. It is noteworthy that we use the overall relation \(\boldsymbol{\tilde{r}}_{r} = \boldsymbol{W}^{g}_{e}\boldsymbol{\bar{r}}_{r} + \boldsymbol{W}^{l}_{e}\boldsymbol{\tilde{r}}_{irj}\) for relation-aware triple attention mechanism, where \(\boldsymbol{W}^{g}_{e}, \boldsymbol{W}^{l}_{e} \in \mathbb{R}^{2d_e \times d_r}\) are mapping matrices for global and local relation respectively. Specifically, the head-aware semantic triple representation \(\boldsymbol{\bar{x}}_{h r}\) is obtained via head-aware triple attention:

\[\begin{align} & c_{i r}=a^{T}\left(\boldsymbol{x}_{i} \boldsymbol{W}_{h} \oplus\boldsymbol{\tilde{R}}_{irj} \boldsymbol{W}^{s}_{t}\right) \tag{5} \\ & \alpha_{i r}=\frac{\exp \left(\operatorname{LReLU}\left(c_{i r}\right)\right)}{\sum_{\left(e_{i^{\prime}}, r, e_{j^{\prime}} \right) \in T_{r}} \exp \left(\operatorname{LReLU}\left(c_{i^{\prime} r}\right)\right)} \tag{6} \\ & \boldsymbol{\bar{x}}_{h r}=\operatorname{LReLU}\left(\sum_{T_r}\left(\alpha_{i r} \boldsymbol{x}_{i} \boldsymbol{W}_{h}\right)\right) \tag{7} \end{align}\]

where \(a \in \mathbb{R}^{2d_r \times 1}\) is a one-dimension vector to map the multi-dimension input into a scalar. \(\boldsymbol{W}_{h} \in \mathbb{R}^{d_e \times d_r}\) and \(\boldsymbol{W}^{s}_{t} \in \mathbb{R}^{(2d_e + d_r) \times d_r}\) are linear transition matrices for head entity and ensemble triple in semantic space. \(LReLU\) is the Leaky_Relu activation function, \(\alpha_{i r}\) is the attention score from head entity \(\boldsymbol{x}_i\) to ensemble triple representation. Then the tail-aware semantic triple representation \(\boldsymbol{\bar{x}}_{t r}\) and relation-aware semantic triple representation \(\boldsymbol{\bar{R}}_{r}\) can be obtained in the same way.

Then, the fused ensemble triple representation in semantic space is computed as combining \(\boldsymbol{\bar{x}}_{h r}\), \(\boldsymbol{\bar{x}}_{t r}\) and \(\boldsymbol{\bar{R}}_{r}\), which is added with the primary specificity as the final triple representation \(\boldsymbol{S}_{irj}\) in semantic space for \((e_{i}, r, e_{j}) \in T\) to fully retain the triple semantic specificity:

\[\begin{equation*} \boldsymbol{S}_{irj}=\boldsymbol{\bar{x}}_{h r}+\boldsymbol{\bar{x}}_{tr}+\boldsymbol{\bar{R}}_{r}+(\boldsymbol{\tilde{R}}_{irj}\boldsymbol{W}^{s}_{t}) \tag{8} \end{equation*}\]

4.2.2  Type Space-Enhanced Triple Representation

In this module, we adopt nonlinear mapping to generate type embedding \(\boldsymbol{X}^{t} \in \mathbb{R}^{n \times d_t}\) from semantic embedding \(\boldsymbol{X} \in \mathbb{R}^{n \times d_e}\), where \(d_t\) is the type dimension:

\[\begin{equation*} \boldsymbol{X}^{t} = \operatorname{tanh}({\boldsymbol{XW}} + b) \tag{9} \end{equation*}\]

To effectively characterize triple, we regard the elements of type triple as a whole to incorporate type information considering the type inseparability. For a triple \((e_i,r,e_j)\), the global relation representation \(\boldsymbol{\bar{r}}^{t}_{r} \in \mathbb{R}^{2d_t}\) is computed as averaging the concatenation of entity pair with the same relation \(r\), which is added to the local relation \(\boldsymbol{\tilde{r}}^{t}_{irj} = \boldsymbol{x_{i}} \oplus \boldsymbol{x_{j}}\) for generating type triple \(\boldsymbol{\tilde{R}}^{t}_{irj}\) as Eq. (10)-(12), where \(\boldsymbol{W}^{r}_t \in \mathbb{R}^{2d_t \times d_r}\) is the transition matrix:

\[\begin{align} & \boldsymbol{\bar{r}}^{t}_{r}=\frac{1}{\left|T_{r}\right|} \sum_{(e_i,r,e_j) \in T_{r}}\left(\boldsymbol{x}^{t}_{i} \oplus \boldsymbol{x}^{t}_{j}\right) \tag{10} \\ & \boldsymbol{r}^{t}_{irj}=\boldsymbol{\bar{r}}^{t}_{r}+\boldsymbol{\tilde{r}}^{t}_{irj} \tag{11} \\ & \boldsymbol{\tilde{R}}^{t}_{irj} = \boldsymbol{x}^{t}_{i} \oplus (\boldsymbol{r}^{t}_{irj}\boldsymbol{W}^{r}_t) \oplus \boldsymbol{x}^{t}_{j} \tag{12} \end{align}\]

Then a Semantic-Type mutual attention is designed, in which the enhanced type-space triple representation \(\boldsymbol{\bar{T}}_{r}\) can be obtained:

\[\begin{align} & \alpha_{t r} = \frac{\exp \left( \operatorname{LReLU} \left( a^{T} \left( \boldsymbol{S}_{irj} \oplus \boldsymbol{\tilde{R}}^{t}_{irj} \boldsymbol{W}_{t} \right)\right)\right) }{ \sum\limits_{T_r} \exp \left( \operatorname{LReLU} \left( a^{T} \left( \boldsymbol{S}_{i^{\prime}rj^{\prime}} \oplus \boldsymbol{\tilde{R}}^{t}_{i^{\prime}rj^{\prime}} \boldsymbol{W}_{t} \right)\right)\right) } \tag{13} \\ & \boldsymbol{\bar{T}}_{r}=\operatorname{ReLU}\left(\sum_{(e_i,r,e_j) \in T_{r}}\left(\alpha_{t r} \boldsymbol{S}_{irj}\right)\right) \tag{14} \end{align}\]

where \(\boldsymbol{W}_{t} \in \mathbb{R}^{(2d_t + d_r) \times d_r}\) are the trainable parameter for type triple. And the enhanced global representation \(\boldsymbol{\bar{S}}_{r}\) can be generated in the same way.

Finally, the type-enhanced ensemble triple representation \(\boldsymbol{T}_{ijr} \in \mathbb{R}^{(d_r + d_t)}\) is obtained while preserving primary type features.

\[\begin{align} & \boldsymbol{T}^{\prime}_{irj} = \boldsymbol{S}_{irj} + \boldsymbol{\bar{S}}_{r} + (\boldsymbol{\tilde{R}}^{t}_{irj}\boldsymbol{W}_{t}) + \boldsymbol{\bar{T}}_{r} \tag{15} \\ & \boldsymbol{T}_{irj} = \boldsymbol{T}^{\prime}_{irj} \oplus \boldsymbol{\bar{r}}^{t}_{r} \tag{16} \end{align}\]

4.3  Triple-Aware Entity Enhancement
4.3.1  Cycle Co-Enhanced Entity Pair Representation Based on Triple Attention

An entity may play different roles as head or tail in different triples and the influences of a head entity \(e_i\) and a tail entity \(e_j\) on triple \((e_i, r, e_j)\) are entirely different. In this module, head entity and tail entity are respectively generated via triple attention and get reinforced circularly.

Firstly, the head-oriented entity representation is enforced via triple attention, which can be formulated as Eq. (17)-(18).

\[\begin{align} & \alpha_{t h} = \frac{ \exp \left( \operatorname{LReLU} \left( a^{T} \left( \boldsymbol{T}_{ijr}\boldsymbol{W}^{c}_{h} \oplus \boldsymbol{x}_{i} \right)\right)\right) }{ \sum\limits_{ \left( e_i,r^{\prime},e_j^{\prime} \right) \in T} \exp \left( \operatorname{LReLU} \left( a^{T} \left( \boldsymbol{T}_{ij^{\prime}r^{\prime}}\boldsymbol{W}^{c}_{h} \oplus \boldsymbol{x}_{i} \right)\right)\right) } \tag{17} \\ & \boldsymbol{x}_{i} = \boldsymbol{x}_{i} + \operatorname{ReLU} \left( \sum_{\left(e_i,r,e_j\right) \in T}\left(\alpha_{t h} \boldsymbol{T}_{ijr} \boldsymbol{W}^{c}_{h}\right) \right) \tag{18} \end{align}\]

After that, the tail-oriented entity representation is carried out as the same of head-oriented procedure, which can be formulated as Eq. (19)-(20).

\[\begin{align} & \alpha_{t t} = \frac{ \exp \left( \operatorname{LReLU} \left( a^{T} \left( \boldsymbol{T}_{ijr} \boldsymbol{W}^{c}_{t} \oplus \boldsymbol{x}_{j} \right)\right)\right) }{ \sum\limits_{ \left( e_i^{\prime},r^{\prime},e_j \right) \in T} \exp \left( \operatorname{LReLU} \left( a^{T} \left( \boldsymbol{T}_{i^{\prime}jr^{\prime}} \boldsymbol{W}^{c}_{t} \oplus \boldsymbol{x}_{j} \right)\right)\right) } \tag{19} \\ & \boldsymbol{x}_{j} = \boldsymbol{x}_{j} + \operatorname{ReLU}\left(\sum_{\left(e_i,r,e_j\right) \in T}\left(\alpha_{t t} \boldsymbol{T}_{ijr} \boldsymbol{W}^{c}_{t}\right) \right) \tag{20} \end{align}\]

where \(\boldsymbol{W}^{c}_{h}, \boldsymbol{W}^{c}_{t} \in \mathbb{R}^{(d_r + d_t) \times d_e}\) are the trainable weight parameters for triples.

Furthermore, it should be noted that the cycle enhancement between head and tail is accomplished by regarding the head-oriented entity representation as the participant of tail-oriented attention computation and vice versa. Specifically, the head-oriented entity representation generated by Eq. (18) will be the participant of the tail-oriented attention computation in Eq. (19). Similarly, the tail-oriented entity representation generated by Eq. (20) will be the participant of the head-oriented attention computation in Eq. (17) in the next cycle.

4.3.2  Neighbor Re-Aggregation

In the last part of TTEA, we apply a GAT layer to re-aggregate neighbor information with modeled representation and the final entity representation \(\boldsymbol{x}_{i}^{f}\) is generated for EA.

\[\begin{align} & \alpha_{i j}=\frac{\exp \left(\operatorname{LReLU }\left(a^{T}\left(\boldsymbol{x}_{i} \oplus \boldsymbol{x}_{j}\right)\right)\right)}{\sum\limits_{e_k \in N_{i}} \exp \left(\operatorname{LReLU}\left(a^{T}\left(\boldsymbol{x}_{i} \oplus \boldsymbol{x}_{k} \right)\right)\right)} \tag{21} \\ & \boldsymbol{x}_{i}^{f}=\boldsymbol{x}_{i} \oplus \left( \operatorname{ReLU}\left(\sum_{e_j \in N_{i}}\left(\alpha_{i j} \boldsymbol{x}_{j}\right)\right)\right) \tag{22} \end{align}\]

where \(N_{i}\) is the set of neighbor entities of \(e_i\).

4.4  Entity Alignment Strategy

To evaluate the confidence of entity pair being equivalent, the Manhattan distance is used to measure the semantic discrepancy between entity representations from different KGs for illustrating the similarity of entities as Eq. (23), based on which the metric in Chapter 5.5 is computed and the margin-based loss \(L\) is defined as Eq. (24). Moreover, we deploy a bi-direction iterative strategy following MRAGA [48] to expand training seed pairs based on negative-sample method, during which only the bi-directional aligned entity pairs can be regarded as new seed entity pairs in next training epoch.

\[\begin{align} & {dis}\left(e_{i}, e_{j}\right)=\left\|\boldsymbol{x}_{i}^{f}-\boldsymbol{x}_{j}^{f}\right\|_{1} \tag{23} \\ & L= \sum_{\left(e_{i}, e_{j}\right) \in S}\max\left({dis}\left(e_{i}, e_{j}\right)-{dis}\left(e_{i}^{\prime}, e_{j}^{\prime}\right)+\lambda, 0\right) \tag{24} \end{align}\]

where \((e_{i}, e_{j})\) is a pre-aligned entity pair in training set \(S\), \((e_i^{\prime}, e_j^{\prime})\) is the negative sample generated by randomly replacing \(e_{i}\) or \(e_{j}\) with their \(k\)-nearest neighbors, \(\lambda\) is the margin hyper-parameter.

The overall training process of TTEA framework can be described as Algorithm 1.

5.  Experimental Setup

5.1  DataSets

In order to make the reliable and fair comparison, we evaluate TTEA on three real-world multi-lingual datasets from a simplified DBP15K described in Table 1, which are adopt by almost all the baselines and are constructed by removing lots of isolated entities and relations from primary DBP15K [52]. Each dataset in the simplified version contains 15000 aligned entity pairs.

Table 1  Statistical data of simplified DBP15K.

5.2  Baselines

To comprehensively evaluate our approach, we choose some embedding-based methods as baselines to make comparison. These approaches are divided into three groups: Trans-based methods, GNNs-based methods and Semi-supervised based methods:

- Trans-based methods: MTransE [22], JAPE [52], BootEA [47], TransEdge [16], RpAlign [27].

- GNNs-based methods: (1) GCN-based methods: GCN-Align [10], HMAN [43], HGCN [30], MCEA [50]. (2) GAT-based methods: NAEA [36], RDGCN [20], KAGNN [33], MRGA [35], SHEA [37], RAGA-l [21].

- Semi-supervised methods: MRAEA [48], RREA-semi [11], RAGA-semi [21], MCEA-semi [50].

It should be noted that methods requiring additional information such as RAEA [42] and RNM [19] are not considered as baselines for fairness.

5.3  Model Variants

To make valid evaluation on different components in our framework, we implement three variants of TTEA to verify their effectiveness:

(1) wo-E: a simplified TTEA version without Ensemble Triple Attention.

(2) wo-T: a simplified TTEA version without Type Space-Enhanced module.

(3) wo-C: a simplified TTEA version without Cycle co-Enhanced module.

5.4  Implementation Details

We get the cross-lingual initial entity embedding from RAGA [21], which adopted translation technique to mapping Glove [53] embedding into different languages. For a fair comparison with the baselines, we use a 30% proportion of alignment seeds for training and the rest for testing as baselines do. The grid search strategy is used to manually tune the hyper-parameters. And we used the Adam optimizer [54] to train our model. The training epoch \(E\) is set to 80, and we make evaluation on test dataset every 5 epochs. The depth \(l\) of Highway-GCNs is 2, both the relation dimension \(d_r\) and type dimension \(d_t\) are 100. And the number of epochs \(p\) for updating negative samples is 5, the number of nearest negative samples \(k\) is 5. The margin hyper-parameter \(\lambda\) is 3.0.

5.5  Metrics

By convention, we report the Hits@1, Hits@10 and MRR results, which are widely used for the evaluation of alignment effectiveness. Specifically, Hits@k measures the percentage of correct alignment ranked at top k which can be ranked by Eq. (23), and MRR is the average of the reciprocal ranks of results, which can be computed as Eq. (25). Higher Hits@k and MRR scores indicate better performance.

\[\begin{equation*} {MRR}=\frac{1}{\left|N\right|} \sum_{1 \leq i \leq {\left|N\right|}}\frac{1}{rank_i} \tag{25} \end{equation*}\]

6.  Results and Analysis

6.1  Overall Performance

The overall EA results are shown in Table 2, which are either implemented with the source codes or provided by original papers. To make a comprehensive comparison with baselines from different perspectives, experiments are employed in two settings: (1) -base: the training seed entity pairs are not expanded during training process, (2) -semi: the training seed entity pairs will be dynamically expanded with a bi-directional iterative training strategy. The solid lines separate Trans-based methods, GNNs-based methods and Semi supervised-based methods and dot line makes a distinction between GCNs-based methods and GATs-based methods and TTEA variants are under the dashed line.

Table 2  Overall performance of entity alignment

For Trans-based methods, TransEdge and RpAlign outperform MTransE, JAPE, and BootEA with their unique representation for triple elements. In detail, RpAlign achieves better Hits@1 for its relation prediction and self-training mechanism, while TransEdge gets more excellent Hits@10 and MRR on ZH_EN and JA_EN via contextualizing relation representation in terms of specific head-tail entity pair. For GCNs-based methods, GCN-Align gets the worst results as shallow utilization of relation triple while MCEA outperforms others for extending the convolution region of long-tail entities. Furthermore, NAEA, RDGCN, KAGNN, MRGA, SHEA, RAGA-l all adopt GATs to obtain fine-grained representation, which get excellent performance without doubt. Among them, RAGA-l achieves the best results, which generate relation from entity via attention mechanism and then aggregate relation to entity. Compared with baselines, TTEA performs best in all evaluation metrics on three datasets with the consideration of triple specificity and the role diversity.

6.2  Ablation Analysis
6.2.1  Effect of TTEA Components

The results of model variants with base setting, i.e. (TTEA-base(wo-E), TTEA-base(wo-T) and TTEA-base(wo-C) in Table 2) and semi setting, i.e. (TTEA-semi(wo-E), TTEA-semi(wo-T) and TTEA-semi(wo-C) in Table 3) both show that while ensemble triple attention, Type Space-enhanced module and Cycle co-Enhanced module in TTEA all make a improvement, Type Space-Enhanced module has a more significant effect. Moreover, three modes of Cycle co-Enhanced module with different cycle orders are compared in Table 4 to explore appropriate cycle form.

mode1: the mode of co-enhanced process with the head-tail order.

mode2: the mode of cycle co-enhanced process with the head-tail-head order.

mode3: the mode of cycle co-enhanced process with the head-tail-head-tail order.

Table 3  Comparison of variants with semi setting.

Table 4  Comparison of different modes of cycle co-enhanced module.

We can see from Table 4 that the mode2 adopted in TTEA is more effective on ZH_EN and FR_EN for reasonable cycle process while the mode3 gets the almost same performance on JA_EN.

6.2.2  Impact of GCN Depth

To explore the impact of different Highway-GCNs depth \(l\), we compare TTEA variants with different depths with \(l = 1\), \(l = 2\) and \(l = 3\) in Table 5. The results show that the TTEA variant with a two-layer Highway-GCNs obtains the greatest superiority on ZH_EN and JA_EN for their complex structure, while a one-layered Highway-GCNs get the best performance on FR_EN for its entity semantic reliance.

Table 5  Comparison of different depths of Highway-GCNs.

6.2.3  Impact of Relation and Type Dimensions

There are two dimensional hyper-parameters: relation dimension \(d_r\) and type dimension \(d_t\) in TTEA. We respectively evaluate TTEA on six different relation and type dimensions as 50, 100, 150, 200, 250 and 300 to explore dimensional impacts. The results in Fig. 3 show that different relation and type dimensions have approximate performance, which indicate that dimensions have little influence on TTEA. Especially, the best results can be obtained on ZH_EN when \(d_r=d_t=100\).

Fig. 3  EA performance with different relation and type dimensions from 50 to 300.

6.2.4  Impact of Seed Entity Pairs

To explore the impact of different training seeds, we compare RAGA-l and RAGA-semi with TTEA-base and TTEA-semi by varying the proportion of training seeds from 25% to 50% with a step size of 5%. The results in Fig. 4, Fig. 5 and Fig. 6 respectively depict Hits@1, Hits@10 and MRR with different seeds proportions. It is showed that TTEA is better than RAGA on both base and semi-supervised local alignment methods two modes for all metrics of three datasets. And as training seeds increase, the Hit@1, Hits@10 and MRR curves of TTEA-base on three datasets are steeper than RAGA-l, which draw the better performance and potentiality.

Fig. 4  Hits@1 with different training seed pairs.

Fig. 5  Hits@10 with different training seed pairs.

Fig. 6  MRR with different training seed pairs.

7.  Conclusion

In this paper, to address insufficient utilization of triple specificity and the diversity of entity role, we present a novel framework TTEA - Type-enhanced Ensemble Triple Representation via Triple-aware attention for Cross-lingual Entity Alignment. By modeling role features and relational interaction between semantic space and type space, TTEA is capable to incorporate ensemble triple specificity and learn cycle co-enhanced head and tail representations. Compared with baselines, our model achieves comparative performance on three real-world cross-lingual datasets.

Acknowledgments

This work was supported by the National Natural Science Foundation of China [grant number 72274138].

References

[1] C. Bizer, J. Lehmann, G. Kobilarov, S. Auer, C. Becker, R. Cyganiak, and S. Hellmann, “Dbpedia-a crystallization point for the web of data,” Journal of web semantics, vol.7, no.3, pp.154-165, 2009.
CrossRef

[2] F.M. Suchanek, G. Kasneci, and G. Weikum, “Yago: A large ontology from wikipedia and wordnet,” Journal of Web Semantics, vol.6, no.3, pp.203-217, 2008.
CrossRef

[3] R. Speer, J. Chin, and C. Havasi, “Conceptnet 5.5: An open multilingual graph of general knowledge,” vol.31, no.1, 2017.
CrossRef

[4] J. Li, G. Yan, S. Wang, and J. Bao, “Financial network representation learning for time series investment behavior,” 2019 IEEE 10th International Conference on Software Engineering and Service Science (ICSESS), pp.277-281, IEEE, 2019.
CrossRef

[5] Y. Song, L. Cai, K. Zhang, H. Zan, T. Liu, and X. Ren, “Construction of chinese pediatric medical knowledge graph,” Joint International Semantic Technology Conference, vol.1157, pp.213-220, Springer, 2019.
CrossRef

[6] H. Sun, J. Xiao, W. Zhu, Y. He, S. Zhang, X. Xu, L. Hou, J. Li, Y. Ni, G. Xie, “Medical knowledge graph to enhance fraud, waste, and abuse detection on claim data: model development and performance evaluation,” JMIR Medical Informatics, vol.8, no.7, p.e17653, 2020.
CrossRef

[7] J. Han, B. Cheng, and X. Wang, “Open domain question answering based on text enhanced knowledge graph with hyperedge infusion,” Findings of the Association for Computational Linguistics: EMNLP 2020, Online, pp.1475-1481, Association for Computational Linguistics, Nov. 2020.
CrossRef

[8] X. Yang, Z. Huan, Y. Zhai, and T. Lin, “Research of personalized recommendation technology based on knowledge graphs,” Applied Sciences, vol.11, no.15, p.7104, 2021.
CrossRef

[9] M.A. Pellegrino, M. Santoro, V. Scarano, and C. Spagnuolo, “Automatic skill generation for knowledge graph question answering,” European Semantic Web Conference, vol.12739, pp.38-43, Springer, 2021.
CrossRef

[10] Z. Wang, Q. Lv, X. Lan, and Y. Zhang, “Cross-lingual knowledge graph alignment via graph convolutional networks,” Proc. 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, Association for Computational Linguistics, Oct.-Nov. 2018.
CrossRef

[11] X. Mao, W. Wang, H. Xu, Y. Wu, and M. Lan, “Relational reflection entity alignment,” CIKM ’20, Proc. 29th ACM International Conference on Information & Knowledge Management, Association for Computing Machinery, 2020.
CrossRef

[12] Y. Peng, J. Zhang, C. Zhou, and J. Xu, “Embedding-based entity alignment using relation structural similarity,” 2020 IEEE International Conference on Knowledge Graph (ICKG), pp.123-130, 2020.
CrossRef

[13] Y. Zhang, T. Gao, J. Lu, Z. Cheng, and G. Xiao, “Adaptive entity alignment for cross-lingual knowledge graph,” International Conference on Knowledge Science, Engineering and Management, vol.12816, pp.474-487, 2021.
CrossRef

[14] Z. Wang, J. Zhang, J. Feng, and Z. Chen, “Knowledge graph embedding by translating on hyperplanes,” Proc. AAAI conference on artificial intelligence, vol.28, no.1, 2014.
CrossRef

[15] Y. Lin, Z. Liu, M. Sun, Y. Liu, and X. Zhu, “Learning entity and relation embeddings for knowledge graph completion,” Twenty-ninth AAAI conference on artificial intelligence, vol.29, no.1, 2015.
CrossRef

[16] Z. Sun, J. Huang, W. Hu, M. Chen, L. Guo, and Y. Qu, “Transedge: Translating relation-contextualized embeddings for knowledge graphs,” International Semantic Web Conference, vol.11778, pp.612-629, 2019.
CrossRef

[17] J. Gao, X. Liu, Y. Chen, and F. Xiong, “Mhgcn: Multiview highway graph convolutional network for cross-lingual entity alignment,” Tsinghua Science and Technology, vol.27, no.4, pp.719-728, 2022.
CrossRef

[18] Z. Xie, R. Zhu, K. Zhao, J. Liu, G. Zhou, and J.X. Huang, “Dual gated graph attention networks with dynamic iterative training for cross-lingual entity alignment,” New York, NY, USA, Association for Computing Machinery, vol.40, no.3, pp.1-30, 2021.
CrossRef

[19] Y. Zhu, H. Liu, Z. Wu, and Y. Du, “Relation-aware neighborhood matching model for entity alignment,” Proc. AAAI Conference on Artificial Intelligence, vol.35, no.5, pp.4749-4756, 2021.
CrossRef

[20] Y. Wu, X. Liu, Y. Feng, Z. Wang, R. Yan, and D. Zhao, “Relation-aware entity alignment for heterogeneous knowledge graphs,” Proc. Twenty-Eighth International Joint Conference on Artificial Intelligence, pp.5278-5284, Aug. 2019.
CrossRef

[21] R. Zhu, M. Ma, and P. Wang, “Raga: Relation-aware graph attention networks for global entity alignment,” arXiv:2103.00791 [cs], March 2021.

[22] M. Chen, Y. Tian, M. Yang, and C. Zaniolo, “Multilingual knowledge graph embeddings for cross-lingual knowledge alignment,” Proc. Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI 2017, Melbourne, Australia, pp.1511-1517, ijcai.org, 2017.
CrossRef

[23] H. Zhu, R. Xie, Z. Liu, and M. Sun, “Iterative entity alignment via joint knowledge embeddings,” Proc. Twenty-Sixth International Joint Conference on Artificial Intelligence, IJCAI 2017, Melbourne, Australia, pp.4258-4264, ijcai.org, 2017.
CrossRef

[24] X. Song, H. Zhang, and L. Bai, “Entity alignment between knowledge graphs using entity type matching,” International Conference on Knowledge Science, Engineering and Management, pp.578-589, 2021.
CrossRef

[25] Y. Xiang, Z. Zhang, J. Chen, X. Chen, Z. Lin, and Y. Zheng, “Ontoea: Ontology-guided entity alignment via joint knowledge graph embedding,” Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online, pp.1117-1128, Association for Computational Linguistics, Aug. 2021.
CrossRef

[26] S. Kang, L. Ji, Z. Li, X. Hao, and Y. Ding, “Iterative cross-lingual entity alignment based on transc,” IEICE Trans. Inf. & Syst., vol.E103-D, no.5, pp.1002-1005, 2020.
CrossRef

[27] H. Huang, C. Li, X. Peng, L. He, S. Guo, H. Peng, L. Wang, and J. Li, “Cross-knowledge-graph entity alignment via relation prediction,” Knowledge-Based Systems, vol.240, p.107813, 2022.
CrossRef

[28] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, and O. Yakhnenko, “Translating embeddings for modeling multi-relational data,” Advances in Neural Information Processing Systems, ed. C. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K. Weinberger, Curran Associates, Inc., 2013.

[29] Y. Wu, X. Liu, Y. Feng, Z. Wang, and D. Zhao, “Neighborhood matching network for entity alignment,” Proc. 58th Annual Meeting of the Association for Computational Linguistics, Online, pp.6477-6487, Association for Computational Linguistics, 2020.
CrossRef

[30] Y. Wu, X. Liu, Y. Feng, Z. Wang, and D. Zhao, “Jointly learning entity and relation representations for entity alignment,” arXiv preprint arXiv:1909.09317, 2019.

[31] X. Li, Y. Zhang, and C. Xing, “Jointly learning knowledge embedding and neighborhood consensus with relational knowledge distillation for entity alignment,” arXiv preprint arXiv:2201.11249, 2022.

[32] Z. Sun, C. Wang, W. Hu, M. Chen, J. Dai, W. Zhang, and Y. Qu, “Knowledge graph alignment network with gated multi-hop neighborhood aggregation,” Proc. AAAI Conference on Artificial Intelligence, vol.34, no.1, pp.222-229, 2020.
CrossRef

[33] Z. Huang, X. Li, Y. Ye, B. Zhang, G. Xu, and W. Gan, “Multi-view knowledge graph fusion via knowledge-aware attentional graph neural network,” Applied Intelligence, vol.53, no.4, pp.3652-3671, 2022.
CrossRef

[34] S. Jiang, T. Nie, D. Shen, Y. Kou, and G. Yu, “Entity alignment of knowledge graph by joint graph attention and translation representation,” Web Information Systems and Applications, ed. C. Xing, X. Fu, Y. Zhang, G. Zhang, and C. Borjigin, vol.12999, pp.347-358, Springer International Publishing, Cham, 2021.
CrossRef

[35] L. Ding, W. Yuan, K. Meng, and G. Liu, “A multi -role graph attention network for knowledge graph alignment,” 2021 International Joint Conference on Neural Networks (IJCNN), pp.1-8, 2021.
CrossRef

[36] Q. Zhu, X. Zhou, J. Wu, J. Tan, and L. Guo, “Neighborhood-aware attentional representation for multilingual knowledge graphs,” Proc. Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China pp.1943-1949, International Joint Conferences on Artificial Intelligence Organization, Aug. 2019.
CrossRef

[37] Z. Yan, R. Peng, Y. Wang, and W. Li, “Soft-self and hard-cross graph attention network for knowledge graph entity alignment,” Knowledge-Based Systems, vol.231, p.107415, 2021.
CrossRef

[38] C. Li, Y. Cao, L. Hou, J. Shi, J. Li, and T.-S. Chua, “Semi-supervised entity alignment via joint knowledge embedding model and cross-graph model,” Proc. 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, pp.2723-2732, Association for Computational Linguistics, Nov. 2019.
CrossRef

[39] X. Mao, W. Wang, Y. Wu, and M. Lan, “From alignment to assignment: Frustratingly simple unsupervised entity alignment,” Proc. 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, pp.2843-2853, Association for Computational Linguistics, 2021.
CrossRef

[40] C. Ge, X. Liu, L. Chen, B. Zheng, and Y. Gao, “Make it easy: An effective end-to-end entity alignment framework,” SIGIR ’21: The 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, Canada, pp.777-786, ACM, 2021.
CrossRef

[41] X. Mao, W. Wang, Y. Wu, and M. Lan, “Lightea: A scalable, robust, and interpretable entity alignment framework via three-view label propagation,” Proc. 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, pp.825-838, Association for Computational Linguistics, 2022.
CrossRef

[42] B. Zhu, T. Bao, L. Liu, J. Han, J. Wang, and T. Peng, “Cross-lingual knowledge graph entity alignment based on relation awareness and attribute involvement,” Applied Intelligence, vol.53, no.6, pp.6159-6177, July 2022.
CrossRef

[43] H. Yang, Y. Zou, P. Shi, W. Lu, J. Lin, and X. Sun, “Aligning cross-lingual entities with multi-aspect information,” Proc. 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, pp.4430-4440, Association for Computational Linguistics, 2019.
CrossRef

[44] B.D. Trisedya, J. Qi, and R. Zhang, “Entity alignment between knowledge graphs using attribute embeddings,” Proc. AAAI Conference on Artificial Intelligence, vol.33, no.1, pp.297-304, 2019.
CrossRef

[45] K. Yang, S. Liu, J. Zhao, Y. Wang, and B. Xie, “Cotsae: Co-training of structure and attribute embeddings for entity alignment,” Proc. AAAI Conference on Artificial Intelligence, vol.34, no.3, pp.3025-3032, April 2020.
CrossRef

[46] Z. Chen, J. Chen, W. Zhang, L. Guo, Y. Fang, Y. Huang, Y. Zhang, Y. Geng, J.Z. Pan, W. Song, and H. Chen, “Meaformer: Multi-modal entity alignment transformer for meta modality hybrid,” Proc. 31st ACM International Conference on Multimedia, MM 2023, Ottawa, ON, Canada, pp.3317-3327, ACM, 2023.
CrossRef

[47] Z. Sun, W. Hu, Q. Zhang, and Y. Qu, “Bootstrapping entity alignment with knowledge graph embedding.,” IJCAI, pp.4396-4402, 2018.
CrossRef

[48] X. Mao, W. Wang, H. Xu, M. Lan, and Y. Wu, “Mraea: An efficient and robust entity alignment approach for cross-lingual knowledge graph,” New York, NY, USA, Association for Computing Machinery, pp.420-428, 2020.
CrossRef

[49] H.T. Trung, T. Van Vinh, N.T. Tam, H. Yin, M. Weidlich, and N.Q. Viet Hung, “Adaptive network alignment with unsupervised and multi-order convolutional networks,” 2020 IEEE 36th International Conference on Data Engineering (ICDE), pp.85-96, 2020.
CrossRef

[50] D. Qi, S. Chen, X. Sun, R. Luan, and D. Tong, “A multiscale convolutional gragh network using only structural information for entity alignment,” Applied Intelligence, vol.53, no.7, pp.7455-7465, 2023.
CrossRef

[51] R.K. Srivastava, K. Greff, and J. Schmidhuber, “Highway networks,” arXiv:1505.00387 [cs], Nov. 2015.

[52] Z. Sun, W. Hu, and C. Li, “Cross-lingual entity alignment via joint attribute-preserving embedding,” The Semantic Web - ISWC 2017, ed. C. d’Amato, M. Fernandez, V. Tamma, F. Lecue, P. Cudré-Mauroux, J. Sequeda, C. Lange, and J. Heflin, Cham, pp.628-644, Springer International Publishing, 2017.
CrossRef

[53] J. Pennington, R. Socher, and C.D. Manning, “Glove: Global vectors for word representation,” Proc. 2014 conference on empirical methods in natural language processing (EMNLP), pp.1532-1543, 2014.
CrossRef

[54] D.P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, Conference Track Proceedings, ed. Y. Bengio and Y. LeCun, 2015.

Footnotes

1. Available source codes in https://github.com/CodesForNlp/TTEA.

Authors

Zhishuo ZHANG
  Tongji University

received the B.S. degree in School of Computer Science and Engineering, Tongji University, Shanghai, China, in 2019. He is currently pursuing his Ph.D. degree in Tongji University of Computer Science and Engineering, Shanghai, China.

Chengxiang TAN
  Tongji University

received the Ph.D. degree in engineering from Northwestern Polytechnic University, China, in 1994. He is currently a Professor of computer science with Tongji University.

Xueyan ZHAO
  Tongji University

received the B.S. degree in School of Computer Science and Engineering, Tongji University, Shanghai, China, in 2017. He is currently pursuing his Ph.D. degree in Tongji University of Computer Science and Engineering, Shanghai, China.

Min YANG
  Tongji University

received the M.S. degree in School of Information and Communication Engineering, Dalian University for Nationalities, 2021. She is currently pursuing her Ph.D. degree in Tongji University of Computer Science and Engineering, Shanghai, China.

Keyword