Skip to main content
PLOS One logoLink to PLOS One
. 2021 Oct 28;16(10):e0258410. doi: 10.1371/journal.pone.0258410

MNI: An enhanced multi-task neighborhood interaction model for recommendation on knowledge graph

Xintao Ma 1,2, Liyan Dong 1,2, Yuequn Wang 1,2, Yongli Li 3, Hao Zhang 1,2,*
Editor: Qi Zhao4
PMCID: PMC8553089  PMID: 34710122

Abstract

To alleviate the data sparsity and cold start problems for collaborative filtering in recommendation systems, side information is usually leveraged by researchers to improve the recommendation performance. The utility of knowledge graph regards the side information as part of the graph structure and gives an explanation for recommendation results. In this paper, we propose an enhanced multi-task neighborhood interaction (MNI) model for recommendation on knowledge graphs. MNI explores not only the user-item interaction but also the neighbor-neighbor interactions, capturing a more sophisticated local structure. Besides, the entities and relations are also semantically embedded. And with the cross&compress unit, items in the recommendation system and entities in the knowledge graph can share latent features, and thus high-order interactions can be investigated. Through extensive experiments on real-world datasets, we demonstrate that MNI outperforms some of the state-of-the-art baselines both for CTR prediction and top-N recommendation.

1 Introduction

Nowadays, we are in the era of data explosion as the internet develops rapidly, which raises a lot of obstacles for users to find their interested information. Recommendation systems aim to tackle this problem by assisting users in exploring among massive data and excavating appealing information tailored for each user. Traditional recommendation systems suffer from the problem of data sparsity and cold-start. One approach is collaborative filtering (CF) [1], which cross-compares the users’ historical interest and then recommends the items according to their common preferences. However, the recommendation precision relies highly on the sparsity of user-item interactions. Therefore, researchers tend to integrate side information that includes user portraits [2], item attributes [3], text reviews [4] during recommendation.

One popular strategy is to combine the knowledge graph with the recommendation system. A Knowledge graph is a heterogeneous graph, wherein entities and relations are represented by different types of nodes and edges in the graph. Besides, side information can be mapped into a knowledge graph with entities and semantical relations [5, 6]. As shown in Fig 1, the movie “Avatar” and “Blood Diamond” are recommended to the user “Bob”. The users, movies, actors, directors, and other attributes are mapped into entities in the knowledge graph. One important contribution of the knowledge graph is that it facilities the recommendation with explainability. For instance, we recommend “Catch me if you can” to Bob because Bob likes “Inception” which shares the same actor “Leonardo DiCaprio”. In many tasks, the knowledge graph can serve not only the structure information but also the rich semantics of user-item interactions so that the behavior of users can be better captured. Those tasks such as link prediction [7], natural language processing (NLP) [8], network similarity [9], etc., are fulfilled by knowledge graph embedding (KGE), which projects the entities and relations into a low dimensional space while preserving the important information such as structural and semantical information.

Fig 1. An illustration of knowledge graph enhanced recommendation system.

Fig 1

The relations between items are annotated with different semantic meanings and the knowledge graph provides detailed facts about movies and their attributes.

A typical knowledge graph embedding method is translational distance models such as TransE [10] and its extensions [1113], which by a translation vector they embed the entities and relations into a vector space and evaluate the distance between the vectors. Though those methods are trivial to implement, but some of them lack the semantical understanding of entities and relations. Besides, some researchers are inspired by the development of deep learning algorithms such as graph neural networks. The deep-learning algorithms are widely used in many research fields, such as social networks [14], bioinformatics [15, 16], web advertising [8], etc. For example, based on the idea of graph convolution networks (GCNs) [17, 18], Wang et al. [19] proposed KGCN, which updates the entity representation by its neighbors, then propagates the information to another part of the knowledge graph. Also, inspired by graph attention network [20], KGAT [21] was proposed to model a high-order relation with attention mechanism in knowledge graphs. These methods take into consideration graph structural information as to improve the recommendation precision. Furthermore, Qu et al. [22] proposed knowledge-enhanced neighbor interaction (KNI), which fixes the problem of early summarization problem in existing graph-based models. The early summarization problem is caused by the fact that the models usually compress the neighbor information into the nodes before prediction, making the local structure of those neighbors implicit in utilizing. KNI compresses the user-item interactions and high-order neighbors to increase the graph local connectivity. However, KNI flaws in treating all the relations as one semantic “link exists or not”, neglecting the semantical and fruitful facts of relations.

Another trend of knowledge graph embedding is incorporating multi-task learning. Wang et al. proposed multi-task feature learning for knowledge graph and recommendation (MKR) [23], which in one way separates the learning tasks of recommendation system (RS) and the knowledge graph embedding, and in the other way inter-relates the two tasks by the relationship between the items in RS and the entities in KG. This method completes the user-item interaction with entities in the knowledge graph, which improves the recommendation precision. However, it neglects some useful information by the scoring matrix.

To address the limitations of the existing methods mentioned above, we propose an enhanced multi-task neighborhood interaction (MNI) model for recommendation on knowledge graph, which utilizes the multi-task learning and also extends the user-item interaction to their neighbors. To be specific, we adopt the idea of KNI to compress the knowledge graph into a local structure where both users’ and items’ neighbors are collected for prediction tasks. This process avoids the early summarization problem. Then in order to distinguish the different semantic information of relations, we integrate them into item entities and interact with the items in the local structure we build.

In summary, our main contributions of this paper are listed as follows:

  • We combine the multi-task learning with enhanced user and item neighbors for preference propagation, which entitles the relations with different semantics and also enhances the recommendation system with neighbor interactions.

  • We propose MNI, a framework that reconstructs the knowledge graph to a local neighbor interaction structure, and utilizes the connection between recommendation system and knowledge graph with mutual enhancement to improve the recommendation precision.

  • We evaluate our framework with extensive experiments on real-world KG datasets. The results show that our framework achieves better performance compared with other state-of-the-art algorithms.

The rest of this paper is organized as follows: we first introduced some related knowledge graph embedding methods in Section 2. Then we illustrate some notations and explain our framework in Section 3. Next, we demonstrate the experiments with the evaluation metrics and analyze the results in Section 4. Finally, we conclude our work and show some future work in Section 5.

2 Related work

Many recommendation techniques have been put forward to solve the data sparsity and cold-start problems in a recommendation system, where side information comes into researchers’ sights. One effective way to utilize the side information is factorization machine (FM) [24] and FFM [25]. The idea is to embed those side information into feature vectors, and then integrate them together with the user and item vectors to train together. However, the time of parameter learning and the prediction is linear. Later, DeepFM [26] is proposed that combines the idea of FM and neural networks. The FM module is used to extract the low-order latent features, and the neural network is used to capture the high-order features.

Another effective method is to build the user-item interaction graphs and their side information into nodes and edges, whereby various graph representation learning methods can be applied to learn the low-dimensional embedding to graph vertices. These embeddings preserve the information of graph topology, node similarity and others. Graph convolutional network (GCN) proposed by Kipf et al. [17] adopts transductive learning to learn all the node embedding. Then GraphSAGE [18] is proposed for inductive node embedding. By incorporating node features, GraphSAGE learns the whole graph topology by integrating each node with its neighbors. Then PGE [27] optimized the sampling process by node clustering to assign bias to different neighbors, and then aggregated neighbors information. Besides, graph attention network (GAT) applied the idea of attention mechanism [20], whereby different neighbors are assigned with different weights for information propagation.

In particular, knowledge graphs are widely used to incorporating side information and user-item interactions into graphs. There are three categories regarding KG-based recommendation systems: embedding-based, path-based, and propagation-based methods [27]. Firstly, embedding-based methods directly use the side information to enrich the representation of users or items. Zhang et al. proposed CKE [28] to utilize different types of side information such as textual and visual knowledge into the embedding process. Therefore, CKE combines the structural information and side information of items to improve the recommendation precision. Later CKFG [29] is proposed to constructs a user-item knowledge graph, so that user behavior can also be learned together with the item side information. Besides, Wang et al. proposed SHINE [14], which leveraged the auto-encoder to aggregate the users’ social network and their profile with the target item. Wang et.al also proposed MKR [23], which separates the recommendation module and the KGE module. The recommendation module is used to learn the latent representation of users and items, and the KGE module is used to learn the semantic relations and item-related entities. Then those two models are trained through a cross-unit.

Secondly, path-based methods aim to take the connectivity similarity of users or items to enhance the recommendation. Yu et al. proposed Hete-MF [30] and HeteRec [31], which extract meta-paths and the similarity of the paths to learn the representation of users and items. Then Zhao et al. came up with FMG [32] by replacing the meta-path by the meta-graph, which has richer connectivity information. Recently, some studies are carried out to learn the explicit representation of meta-paths. MCRec [33] is designed to learn the interaction context representation from the meta-paths. HetNERec [34] is proposed from another aspect, by constructing the co-occurrence networks to discover the implicit relationship.

The third category is propagation-based methods, which leverage the idea of GCN of information propagation [35]. The first work is Ripplenet by Wang et al. [36], which introduces the concept of ripple set and propagates the user’s preference to enrich the user’s representation. Contrarily, KGCN [19] is designed by enriching the representation of the candidate item with the embedding of entities and their neighbors in KG. Besides, Wang et al. proposed KGAT [21], aiming to model the high-order relations between users and items using attention-based information propagation. Recently, KNI is designed by Qu et al. [22], which incorporates the neighbors of items and also the neighbors of users. Thus the refinement of user and item embedding are pulled together. Besides, Wang et al. proposed Ripp-MKR [37], a deep framework that combined the main idea of Ripplenet and MKR. The framework enriches the recommendation system with users’ historical clicked-items, and combined with knowledge graph information.

However, our framework differs from the above literature in that: we apply the neighbor interaction of users and items with enhanced multi-task learning, so that the refinement of item representation is the mutual interaction of user-item-neighbor pair and item-entity pair. Unlike Ripp-MKR, our framework focuses on the neighbor-neighbor interactions, which is not limited to user-item interactions. Upon the reconstruction based on the neighborhood interactions of the graph, it can fully discover the implicit relations between items in order to capture the whole structure of the recommendation system. We also apply attention mechanism to focus on the important semantics on edges to improve recommendation precision. The details are explained in Section 3.

3 Proposed work

Our framework MNI is shown in Fig 2. It contains three parts: the recommendation module, knowledge graph embedding (KGE) module, and the cross&compress unit. The recommendation module is inspired by the work KNI [22]. The first step is to reconstruct the graph into neighbor-interaction graph, and the process is shown in Fig 3. With the reconstructed graph Gkig, yielding the initial vectors xu,v and their neighbors xp,q. We adopt GAT (multi-head attention) to learn the high-order information in the reconstructed graph as shown in the bottom part of Fig 2. Here, the blue blocks represent items, the orange blocks represent users, and the green ones represent entities. Thus, the embedding of users UL and items VL can be obtained. The knowledge graph embedding module shown on the upper part of Fig 2, uses a multi-layer to extract the features of head and relation from the triples, which can preserve the semantic of relations and also the graph structure information. The head entity vectors are represented as yellow blocks and the relation vectors are shown as grey blocks. Then with the cross&compress unit [23], a bridge is built between the recommendation module and the KGE module. The bridge exchanges and compensates for the latent information produced by the two modules through items VL and head entities rL. Afterwards, the predicted user-item interaction values y^uv and the predicted tail entities t^ are calculated and converged by the loss function. Therefore, our framework aggregates the high-order neighborhood information in the recommendation system and also the semantic relation information in the knowledge graph.

Fig 2. The framework of MNI.

Fig 2

The bottom part is the recommendation module where the knowledge-enhanced interaction model (KIG) is constructed, and high-order neighborhood is integrated by multi-head attention network. The middle part is the cross&compress unit, a bridge for exchanging latent features. The upper part shows the knowledge graph embedding that predicting the target tail entity by head and relation embedding.

Fig 3. The reconstruction from knowledge graph to knowledge-enhanced interaction model.

Fig 3

The orange circles are users, blue circles are items, and green circles are item attributes. Dash circles denote the user and its neighbors. The process from (b) to (c) demonstrate the high-order neighborhood information is propagated from 2-hop neighbors to 1-hop neighbors. (d) denote the final model, where the edges represent interaction among interactions.

In this section, we will first show some necessary notations and formulations in subsection 3.1. Then we explain our framework module by module, namely the cross&compress unit in subsection 3.2, the recommendation module in subsection 3.3, and the KGE module in subsection 3.4. Finally, we will discuss the learning algorithm in subsection 3.5.

3.1 Formulation

The knowledge graph recommendation system is normally formulated as follows. A recommendation system Grs contains the user set U and the item set V, represented as U = {u1, u2, …, um} and V = {v1, v2, …, vn}, where m, n are the number of users and items. Thus the user-item interaction can be represented as Y = {yuv|uU, vV}. yuv equals to 1 when an interaction between the user u and the item v is observed, such as behaviors like purchasing or rating; otherwise, yuv = 0. Besides, a knowledge graph Gkg is represented as massive entity-relation-entity triples (h, r, t), which correspond to head entity hE, relation rR, and tail entity tE, where E represents the entity sets and R is the relation set. For instance, (Inception, film.actor, Leonardo Dicaprio) is a triple that states the fact Leonardo is an actor of the film Inception. An entity can form multiple triples and associate with other entities in the KG. The entity Inception can also link to the entity Christopher Nolan with relation film.director.

Thus we can formulate our recommendation task as follows:

  • Input: recommendation system Grs and knowledge graph Gkg

  • Output: a prediction function that predicts the probability y^uv=F(u,v) that the user u interacts with the item v.

3.2 The cross&compress unit

The cross&compress unit is mainly designed to build a bridge between the recommendation module and the KGE module [22], through which the two models can share the latent features on each side. The communication is performed between the item and its related entities as shown in Fig 4.

Fig 4. The illustration of cross&compress unit.

Fig 4

The item and entity vectors are operated by cross function to generate the cross feature matrix.

We first introduce the cross feature matrix between the latent feature of item vlRd and the latent feature of entity elRd from layer l, which describe the cross function.

Cl=vlelT=vl1el1vl1eldvldel1vldeld (1)

where ClRd×d is the cross feature matrix of layer l, and d is the number of hidden layer. Additionally, the compress function is also performed by projecting the cross feature matrix into the next layer latent representation space:

vl+1=ClwlVV+ClTwlEV+blV (2)
el+1=ClwlEV+ClTwlEE+blE. (3)

where WliRd and bliRd are the learning parameters. We say that it is a compress function because this operation projects the matrix from Rd×d space back to Rd. To be used later, the cross&compress unit is denoted as the formula:

vl+1,el+1=C(vl,el). (4)

We use the suffix [v] or [e] to infer to the output separately. Thus the latent features of the item side module and its related entity side module can be exchanged and mutually strengthened. One thing to notice is that the cross&compress unit exists only in the low layer of the framework, because the feature transfer-ability decreases as the layer getting higher [23, 34].

3.3 The recommendation module

Most works aggregate the neighbor information before learning the interactions, which compresse the graph structure into two nodes and an edge for prediction. However, this prevents the methods exploring the local structure, which is called early summarization. In [18], the knowledge-enhanced neighborhood interaction (KNI) model is introduced to tackle this problem, where user-item interactions are further explored as neighbor-neighbor interaction and then integrated with high-order neighborhood information.

In this module, we first transform the graph G = GrsGkg to the knowledge-enhanced interaction graph Gkig.

Definition 1

Given the graph G = GrsGkg, where Grs contains the user set U and item set V, Gkg contains massive triples (h, r, t). Let Nu be the neighborhood of the user, and Nv be the neighborhood of the item. Thus we build Gkig = {(i, j)|iUNu, jVNv}.

We argue that in the newly constructed graph Gkig, interactions exist between user and item (u, v), user and item neighbor (u, q), user neighbor and item (p, v), user neighbor and item neighbor (p, q), where pNu and qNv. Those interactions are then all taken into consideration for prediction. Fig 3 shows the reconstruction of the knowledge-enhanced interaction graph. We can see in Fig 3d, the user neighborhoods and item neighborhoods are collected to compute the neighborhood interactions. We use a bi-attention network to use the information of neighborhood interactions:

ap,q=softmaxwTconcatxu,xp,xv,xq+b. (5)
y^uv=pNuqNvap,q<xp,xq>. (6)

where xu is the embedding of the user u, xp is the user neighbor node p, xv is the item v, xq is the item neighbor node q, <> is the inner product. We can see that Eq 1 is different from the attention function in [8, 28, 34], Eq 5 takes the user neighbor node and item neighbor node into account, and then in Eq 6, different interaction assigns different weight according to the attention parameter ai,j. This means that the early summation problem is solved by reconstructing the graph into Gkig, and all types of neighbor interaction are calculated and weighted for prediction.

The next step is integrating high-order neighbor information into the process of prediction. Studies show that high-order neighbor information can reveal latent taste of the user [3, 20, 38], which improves the precision of recommendation. Thus we use graph attention network (GAT) [20] to explore the high-order neighbor information and integrate it into the reconstructed graph. GAT is similar to graph convolution networks except that multi-head self-attention networks are calculated. It allocates different weights to different neighbors, so that the influence of the important neighbors is emphasize and absorbed by the whole graph. Here we use 2-layer GAT:

xp1=σjNpap,j1w1xj+b1. (7)
xu2=σ(iNuau,i1w2xp1+b2). (8)

where ai,jl is the attention score of node i to node j in the first layer attention network, xp1 and xu2 are the outputs of the first and second attention layers, w and b are the learning parameters weights and bias, and we use Relu for the activation function σ(). ai,jl is calculated by:

ai,jl=exp(LeakyRelu(walTconcatxil-1,xjl-1+bal))kNiexp(LeakyRelu(walTconcatxil-1,xkl-1+bal)) (9)

where wal and bal are the learning parameters of the attention network.

Thus, for any node i, we can calculate the embedding xil using Eqs 79. And then we replace any feature vectors in Eq 6 with the calculated xilGkig, so that the reconstructed graph contains the high-order neighborhood information, shown as Fig 3. The process can be explained from Fig 3b and 3c that the information of 2-hop neighbors are propagated to the 1-hop neighbors.

Besides, in order to cope with large graphs, we adopt sampling to use only a fixed-number neighbors for attention mechanism.

Ni~=sampleNi,k. (10)

Here we use neighbor sampling (NS) [18], which randomly samples k neighbors of each node, and thus controls the complexity of the neighborhood information. Other methods can also be used for sophistical sampling, such as random walks [3941].

Therefore, we can get the embedding for users and items ul and xul and xvl integrating the high-order neighbor information and the interactions among different types of neighbors. In order to align the vectors and extract the feature of users and items, we put them into another fully connected neural network ML with L layers, Then for an item, we use L cross&compress units to combine the feature of the item and the related entities.

UL=MLxul. (11)
VL=Ee~Sv[CL(v,e)[ML(xvl)]]. (12)

where M(x) = σ(Wx + b) is the fully-connected network, and S(v) is the related entities of item v.

3.4 The KGE module

In this module, the entities and relations are embedded in a way that the graph structure is maintained and the semantic of relations as well [42]. Since in the recommendation module, the high-order neighborhood information is integrated only by the semantic of existing or not, the information of semantic relation is vacant at some level. Thus, we need the KGE module to preserve the semantic relation in the KG. We use a deep semantic matching model for KGE embedding.

hL=Ee~Sh[CL(v,h)[e]]. (13)

where CL is the cross&compress unit to output the latent feature of head entity h of layer L, S(h) is the related item of the entity h.

rL=ML(r) (14)
t^=MK(hLrL) (15)

where the raw vectors of h and r can be feature vectors including ID, types, or textual descriptions. And then their latent features are concatenated to predict the tail entity.

Finally, the score of predicting the tail entity is calculated by the score function fKG shown as follows:

score(h,r,t)=fKG(t,t^) (16)

where we use fKG=σ(tT,t^).

3.5 Learning algorithm

The loss function of our framework contains three parts:

L=LRS+LKGE+LREG (17)

The first loss is the recommendation loss is calculated as:

LRS=-yu,v=1logy^uv-yu,v=0log1-y^uv (18)

And for the KGE loss, the same as the recommendation loss, we want to increase the score for all positive prediction.

LKGE=-λ1(h,r,tGkgscoreh,r,t-h,r,tGkgscoreh,r,t) (19)

The last loss term is regularization term that prevents overfitting:

LREG=λ2W22 (20)

In conclusion, the whole learning algorithm is shown in Algorithm 1, which we use negative sampling for the training, shown as follows:

Algorithm 1. The learning algorithm of MNI

Input: Interaction matrix Y, knowledge graph G.

Output: Prediction F(u, v|Θ, Y, G)

1: Initialize the parameters

2: Reconstruct the graph to knowledge-enhanced interaction graph Gkig

3: For i = 1,…, max iter do

4:  For t steps do

5:   Sample minibatch of positive and negative interactions from Gkig

6:   Sample e~S(v) for each item in the minibatch.

7:   Calculate the gradients and update the parameters on the minibatch according to Eqs 112, 17 and 20

8:  end for

9:   Sample minibatch of true and false triples from Gkg

10:   Sample v~S(h) for each entity in the minibatch.

11:   Calculate the gradients and update the paramters on the minibatch according to Eqs 1316, 19 and 20

12:end for

The training process starts with the reconstruction of the graph into Gkig. Then the process is divided into two procedures: from line 4–7 the recommendation module is trained with the cross&compress unit; line 8–11 represents the training of KGE module. The gradient of loss is calculated as multi-task training with respect to model parameters Θ. One thing to notice is that the recommendation module is trained for t times, we will discuss the choice of t later in the experiment session.

4 Experiments

In this section, we will show the performance of our framework on three real-world datasets, and compare the results with state-of-the-art baselines, then take parameter sensitivity experiments.

4.1 Dataset

We implement three common real-world datasets: Movielen-1M, Book-Crossing, Last.FM, which are widely used in the field of movies, books, and music. The details are shown as follows:

  • Movielens-1M [43] is a stable benchmark dataset from Grouplens Research. The data was collected over various time periods, which contains the user profile, movie attributes and ratings. The ratings are from 1 to 5. Here we set positive threshold as 4.

  • Book-Crossing [44] contains ratings from 0 to 10 of books in the Book-Crossing community, which also contains some attributes of the user and the book.

  • Last-FM [45] includes the musician listening information from over 2000 users from Last-fm online music system, and also some demographics information of users, tags of tracks.

We use Microsoft Satori [46] to construct the knowledge graph for each dataset. For the three datasets Movielens-1M, Book-Crossing, and Last-FM, we first sample a subset of triples from the knowledge graph, where the relation names contain “movie” or “book” distinguish and also the confidence level is more than 0.9. Then with the subset, we gather all the IDs by matching the names with the tails of the triples, such as (head, film.name, tail). Then we match the IDs with the head and tail of all the KG triples, and extend the set of entities to 4 hops iteratively. One thing to notice is that the items without any matching entity are excluded in our experiments. The details of the datasets are listed in Table 1.

Table 1. The details of the three real-world datasets.

Datasets Movielens-1M Book-Crossing Last-FM
Users 6,036 17,860 1,872
Items 2,347 14,910 3,846
Interactions 753,772 139,746 42,346
Triplets 20,195 19,793 15,518

4.2 Baselines

We compare our results with some state-of-the-art algorithms, which are categorized in factorization machines, embedding-based methods, path-based methods, and propagation-based methods of knowledge graph embedding. They are listed as follows:

  • LibFM [47]: It is a widely used factorization machine model. We use TransR [11] to pretrain the graph and obtain the entity embedding as input to LibFM, where the dimension of TransR is 32. For the user and item input we use raw feature vectors. Here the dimension is {1,1,8} and the training epoch is 50.

  • Wide&Deep [48]: It is another feature-based FM model, which uses deep learning models and the shallow models. The input here is the same as LibFM as described above. The dimension of user, item, and entity is 64. And a two-layer channel is processed with the dimension of 100 and 50.

  • CKE [28]: It is a representative of embedding-based method for KG recommendation, which embeds the semantic side information by TransR such as texual, image, and graphs information. The dimension of entity is 32, and the user and item for the three datasets are 64, 128, 32.

  • DKN [8]: It is another representative embedding-based method for KG recommendation. DKN utilizes multiple channels to embed the entity as well as word, and then combines them in a CNN for prediction. Here, we use movie and book names as textual input for DKN. The dimension for word and entity embedding is 64, and the number of filters is 128 for window size 1, 2, 3.

  • PER [49]: It is from the category path-based method by treating KG as heterogeneous information networks and extracting the meta-path. We manually set the meta path pattern as [19].

  • KNI [22]: This is a typical propagation-based method by regarding the interaction happen between the neighbors of user and the neighbors of items. Here the hop number is 4 and hidden dimension is 128.

  • MKR [23]: It is the main base algorithm for our framework which mutual trains the user-item pair and entity-relation pair. We set the high-level layer as K = 1. The dimension for each data is 8, 8, 4, and the step is 3, 2, 2.

4.3 Experiment setup

We evaluate those models on 2 tasks, click-through rate (CTR) prediction and top-N recommendation. For CTR, we use the evaluation metrics Area Under Curve (AUC) and Accuracy (ACC). And for the recommendation task, we use the metrics of Precision@K and Recall@K, which select K items with the highest predicted probability for each user.

Besides, we use the hyper-parameter k = 1 as high-level layers, λ2 = 10‒6 as the regularization term parameter. For three datasets, we set t as 3, 2, 2. Also, we use hop number of 4 to build the high-order neighborhood. Also, we use 80% of each data for training and the remaining for testing. Within the 80% training set, we randomly choose 20% data for validation.

4.4 Results and discussions

In this subsection, we show the comparison results with other baselines, and then discuss over some parameter sensitivity experiments.

Overall comparison

The comparison results with the baselines regarding CTR prediction is shown in Table 2, and the top-K recommendation results compared with MKR and KNI are shown in Figs 57. We can conclude the following observations:

Table 2. Comparison results of AUC, ACC.

The result of LibFM, Wide&Deep, CKE, PER, DKN is taken from [23].

Model Movielens-1M Book-Crossing Last-FM
AUC ACC AUC ACC AUC ACC
LibFM 0.892 0.812 0.685 0.640 0.777 0.709
Wide&Deep 0.898 0.820 0.712 0.624 0.756 0.688
CKE 0.801 0.742 0.671 0.633 0.744 0.673
DKN 0.655 0.589 0.622 0.598 0.602 0.581
PER 0.710 0.664 0.623 0.588 0.633 0.596
MKR 0.917 0.843 0.734 0.704 0.797 0.752
KNI 0.944 0.872 0.772 0.706 0.823 0.774
MNI 0.951 0.879 0.781 0.707 0.825 0.777
Fig 5. Top K recommendation result for Movielens-1M regarding precision and recall.

Fig 5

Fig 7. Top K recommendation result for Last-FM regarding precision and recall.

Fig 7

  • Our framework performs the best compared with other baselines. Specifically, compared with the two algorithms MKR and KNI that our framework mainly based on, MNI increases the AUC and ACC on all three real-world datasets. On Book-Crossing, our framework increases the AUC by 1.17% compared to KNI and 10.9% compared to MKR. The reason is that our framework takes the local neighborhood structures into account, especially the neighbor interaction integrates all types of neighbors. Besides, by semantically embedding the entities in the knowledge graph and exchanging latent features with the related items full of neighbor information, the embedding is improved by mutual perfection. As for top-K recommendation, our framework also outperforms MKR and KNI, which is consistent with CTR prediction.

  • PER’s performance is not as ideal as others, because it is path-based method for KG recommendation, which relies highly on the property of meta-paths and requires much human expertise. PER requires much more effort on manually designing the meta-path.

  • KNI performs best among all other baselines, which highlights the importance of local neighbor structure [22]. However, KNI treats the interactions among the users, items, and neighbors as the same, which lacks the semantical meaning of KG relations that also plays an important part in user preference.

  • Movielens-1M has the best experiment results because the data is denser. On the other side, Book-Crossing has the worst data sparsity. However, we improve the recommendation results on all datasets, meaning that our framework can deal with data sparsity.

  • Propagation-based methods perform better than embedding-based methods and path-based methods. The reason is the propagation methods combine the advantages of those two other types of methods. The result normally contains high-order information and is also explainable. However, some of them are computationally costly as the graph grows large.

  • Besides, LibFM and Wide&Deep generally perform better than embedding-based methods, which shows the effectiveness of FM-based methods, especially dealing with sparse data. One reason may be that some of the FM-based models exploit high-order layers, which may explore the high-order information. On the other side, embedding-based methods focus on the explicit interaction between users and items.

  • For the top-K recommendation results, recall@K increases as K becomes large. One thing to notice is that even when k is very small, our framework still much better than the others, which indicates that our framework can capture the user preference.

Fig 6. Top K recommendation result for Book-Crossing regarding precision and recall.

Fig 6

Parameter variation

We evaluate our framework on the parameter sensitivity by varying the training step t from 1 to 10. The result is shown in Fig 8. We observe that our framework achieves the best when t = 3. The reason is that a high training frequency of the KGE module falsely directs the objective function. On the other hand, a small training frequency cannot combine the transferred knowledge from both sides of the graphs.

Fig 8. Training frequency on book-crossing.

Fig 8

Then we tune the hop number of neighbors included in the recommendation module from 1 to 4. The result is shown in Table 3. We can see that the best AUC results on Movielens-1m exist using 4 hop neighbors, and for datasets Book-Crossing and Last-FM, the best results occur at hop = 3. This may because that too much high-order neighbor information introduces noise during training, especially when low-order already includes those neighbors that are also covered by high-order propagation, which is a consistent result with other studies [21, 35].

Table 3. Experimental results on AUC of the effect of hop number.
Hop number Movielens-1M Book-Crossing Last-FM
1-hop 0.925 0.739 0.780
2-hop 0.937 0.756 0.791
3-hop 0.951 0.779 0.816
4-hop 0.949 0.781 0.825

5 Conclusion

In this paper, we propose MNI, which is a multi-task learning approach that incorporates the neighbor-neighbor interaction and the semantical embedding. In the first learning task, we reconstruct the knowledge graph to a knowledge-enhanced neighborhood interaction model, which contains user and item, user and item-neighbor, user-neighbor, and item, user-neighbor and item-neighbor interactions before prediction so that early summarization problem is taken care of. Besides, high-order neighborhood information is also integrated by graph attention networks. The other learning task is to embed the entities and relations semantically in the knowledge graph. Meanwhile, the item embedding and the head entity embedding exchange latent features through a cross&compress unit, so that knowledge can mutual flows to each side of the task. We conduct a massive experiments on three real-world datasets. The results show that our framework improves the recommendation quality and efficiency.

For future works, we plan to investigate other types of sampling in the recommendation module, such as sampling a subgraph [50]. Random walks can be adopted to generate a walking sequence with most frequent visited nodes, which can emphasize the important neighbors in the recommendation module. Also, another possible improvement is to alter KGE methods to extract more knowledge from the knowledge graph, such as considering implicit relations between items using propagation techniques [21].

Acknowledgments

We would like to thank Hao Zhang (Jilin University) for the insightful comments on the manuscript and his guidance and patience enlighten us not only on this paper but also our future. We gratefully acknowledge the valuable contribution by Liyan Dong(Jilin University) during preparing the paper.

Data Availability

The data underlying the results presented in the study are available from Movielens, Book-Crossing, and Last-FM. Movielens:https://grouplens.org/datasets/movielens/. The data is third party and we confirm that others would be able to access these data in the same manner as the authors. We confirm that the authors did not have any special access privileges that others would not have.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Koren Y., Bell R., Volinsky C. 2009. Matrix Factorization Techniques for Recommender Systems. Computer. 2009;42(8):30–7. [Google Scholar]
  • 2.M. Jamali and M. Ester. 2009. Trustwalker: a random walk model for combining trust-based and item-based recommendation. In Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, 2009, pp. 397–406.
  • 3.H. Wang, J. Wang, J. Wang, M. Zhao, W. Zhang, F. Zhang, et al. 2018. Graphgan: Graph representation learning with generative adversarial nets. 32nd AAAI Conference on Artificial Intelligence, AAAI 2018, (2018), pp. 2508–2515.
  • 4.Y. Xu, Y. Yang, J. Han, E. Wang, F. Zhuang, and H. Xiong. 2018. Exploiting the sentimental bias between ratings and reviews for enhancing recommendation. In 2018 IEEE International Conference on Data Mining (ICDM). IEEE, 2018, pp. 1356–1361.
  • 5.F. Zhang, N. J. Yuan, D. Lian, X. Xie, and W.-Y. Ma. 2016. Collaborative knowledge base embedding for recommender systems. In Proceedings of the 22Nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ser. KDD ‘16. New York, NY, USA: ACM, 2016, pp. 353–362
  • 6.Zhang Y., Ai Q., Chen X., and Wang P. 2018. Learning over knowledge-base embeddings for recommendation. Algorithms, 2018, 11(9): 137. https://arxiv.org/abs/1803.06540. [Google Scholar]
  • 7.Rossi A., Firmani D., Matinata A., Merialdo P., and Barbosa D. 2021. Knowledge Graph Embedding for Link Prediction: A Comparative Analysis. ACM transactions on knowledge discovery from data. 2021;15(2):1–49. [Google Scholar]
  • 8.H. Wang, F. Zhang, X. Xie, and M. Guo. 2018. DKN: Deep knowledge-aware network for news recommendation. In Proceedings of the 2018 World Wide Web Conference, ser. WWW ‘18. Republic and Canton of Geneva, Switzerland: International World Wide Web Conferences Steering Committee, 2018, pp. 1835–1844.
  • 9.Wang C., Han C., Zhao Q., Chen X. 2021. Circular RNAs and complex diseases: from experimental results to computational models. Briefings in Bioinformatics, 2021, doi: 10.1093/bib/bbab286 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Bordes A., Usunier N., Garcia-Duran A., Weston J., and Yakhnenko O. 2013. Translating embeddings for modeling multi-relational data. Advances in neural information processing systems, 2013, pp. 2787–2795 [Google Scholar]
  • 11.Y. Lin, Z. Liu, M. Sun, Y. Liu, and X. Zhu. 2015. Learning entity and relation embeddings for knowledge graph completion. In AAAI conference on artificial intelligence, 2015
  • 12.G. Ji, K. Liu, S. He, and J. Zhao. 2016. Knowledge graph completion with adaptive sparse transfer matrix. In Proc. 30th AAAI, 2016, pp. 985–991.
  • 13.Nie B. and Sun S. 2019. Knowledge graph embedding via reasoning over entities, relations, and text. Future Generation Computer Systems, vol. 91, pp. 426–433, 2019. [Google Scholar]
  • 14.Wang H., Zhang F., Hou M., Xie X., Guo M., and Liu Q. 2018. Shine: Signed heterogeneous information network embedding for sentiment link prediction. In WSDM, vol. 2018-/(2018), pp.592–600. [Google Scholar]
  • 15.Zhang L., Yang P., Feng H., Zhao Q., Liu H. 2021. Using Network Distance Analysis to Predict lncRNA–miRNA Interactions. Interdisciplinary Sciences: Computational Life Sciences, 2021, 13(3), 535–545. doi: 10.1007/s12539-021-00458-z [DOI] [PubMed] [Google Scholar]
  • 16.Zhang L., Liu T., Chen H., Zhao Q., Liu H. 2021. Predicting lncRNA–miRNA interactions based on interactome network and graphlet interaction. Genomics, 2021, 113(3): 874–880 doi: 10.1016/j.ygeno.2021.02.002 [DOI] [PubMed] [Google Scholar]
  • 17.T. Kipf and M. Welling. 2017. Semi-supervised classification with graph convolutional networks. ICLR. 2017.
  • 18.Hamilton W., Ying Z., and Leskovec J. 2017. Inductive representation learning on large graphs. Advances in Neural Information Processing Systems, vol. 2017-/(2017), pp. 1025–1035. [Google Scholar]
  • 19.H. Wang, M. Zhao, X. Xie, W. Li, M. Guo. 2019. Knowledge Graph Convolutional Networks for Recommender Systems. WWW 2019, (2019), pp. 3307–3313.
  • 20.P. Velickovic, G. Cucurull, A. Casanova, A. Romero, P. Lio, and Y. Bengio. 2018. Graph attention networks. ICLR 2018.
  • 21.X. Wang, X. He, Y. Cao, M. Liu, and T.-S. Chua. 2019. Kgat: Knowledge graph attention network for recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, ser. KDD ‘19. New York, NY, USA: ACM, 2019, pp. 950–958
  • 22.Y. Qu, T. Bai, W. Zhang, J. Nie, and J. Tang. 2019. An end-to-end neighborhood-based interaction model for knowledge-enhanced recommendation. arXiv preprint arXiv:1908.04032, 2019.
  • 23.H. Wang, F. Zhang, M. Zhao, W. Li, X. Xie, and M. Guo.2019. Multitask feature learning for knowledge graph enhanced recommendation. In The World Wide Web Conference, ser. WWW ‘19. New York, NY, USA: ACM, 2019, pp. 2000–2010.
  • 24.S. Rendle, Z. Gantner, C. Freudenthaler, and L. Schmidt-Thieme.2011. Fast context-aware recommendations with factorization machines. In SIGIR.635–644
  • 25.Y. Juan, Y. Zhuang, W. Chin, and C. Lin. 2016. Field-aware Factorization Machines for CTR Prediction. In Proceedings of the 10th ACM Conference on Recommender Systems (RecSys ’16). Association for Computing Machinery, New York, NY, USA, 43–50. 10.1145/2959100.2959134 [DOI]
  • 26.H. Guo, R. Tang, Y. Ye, Z. Li, X. He. 2017. DeepFM: A Factorization-Machine based Neural Network for CTR Prediction. In IJCAI, 2017, pp.1725-1731.
  • 27.Y. Hou, H. Chen, C. Li, J. Cheng, M. Yang. 2019. A representation learning framework for property graphs. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 3–7 August 2019; pp. 65–73.
  • 28.Q. Guo, F. Zhuang, C. Qin. 2020. A Survey on Knowledge Graph-Based Recommender Systems. IEEE Transactions on Knowledge and Data Engineering, (2020), pp. 1–1
  • 29.Ai Q., Azizi V., Chen X., and Zhang Y. 2018. Learning Heterogeneous Knowledge Base Embeddings for Explainable Recommendation. Algorithms 11, 9 (2018), 137. [Google Scholar]
  • 30.X. Yu, X. Ren, Q. Gu, Y. Sun, and J. Han. 2013. Collaborative filtering with entity similarity regularization in heterogeneous information networks. In IJCAI, vol. 27, 2013
  • 31.X. Yu, X. Ren, Y. Sun, B. Sturt, U. Khandelwal, Q. Gu, et al. 2013. Recommendation in heterogeneous information networks with implicit user feedback. In RecSys, 2013, pp. 347–350
  • 32.H. Zhao, Q. Yao, J. Li, Y. Song, and D. L. Lee. 2017. Meta-graph based recommendation fusion over heterogeneous information networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 2017, pp. 635–644.
  • 33.B. Hu, C. Shi, W. X. Zhao, and P. S. Yu. 2018. Leveraging Metapath based Context for Top- N Recommendation with A Neural Co-Attention Model. In SIGKDD, pp.1531–1540.
  • 34.Zhao Z., Zhang X., Zhou H., Li C., Gong M., Wang Y. 2020. HetNERec: Heterogeneous network embedding based recommendation. Knowledge-based systems. 2020; vol.204, 106218 [Google Scholar]
  • 35.Yosinski J., Clune J., Bengio Y., and Lipson H. 2014. How transferable are features in deep neural networks?. Advances in Neural Information Processing Systems, vol. 4/no. January, (2014), pp. 3320–3328 [Google Scholar]
  • 36.H. Wang, F. Zhang, J. Wang, M. Zhao, W. Li, X. Xie et al. 2018. RippleNet: Propagating User Preferences on the Knowledge Graph for Recommender Systems. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management. (2018), pp. 417–426.
  • 37.Wang Y., Dong L., Li Y., Zhang H. 2021. Multitask feature learning approach for knowledge graph enhanced recommendations with RippleNet. PLoS ONE, 2021, 16(5):e0251162 doi: 10.1371/journal.pone.0251162 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.R. van den Berg, T. N. Kipf, and M. Welling. 2017. Graph Convolutional Matrix Completion. In KDD.
  • 39.R. Ying, R. He, K. Chen, P. Eksombatchai, W. L Hamilton, and J. Leskovec. 2018. Graph Convolutional Neural Networks for Web-Scale Recommender Systems. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, (2018), pp. 974–983.
  • 40.A. Grover and J. Leskovec. 2016. Node2vec: Scalable feature learning for networks. In SIGKDD, vol. 13-17-/(2016), pp. 855–864. [DOI] [PMC free article] [PubMed]
  • 41.Lao N, Cohen WW. Relational retrieval using a combination of path-constrained random walks. Machine learning. 2010;81(1):53–67 [Google Scholar]
  • 42.M. Nickel, L. Rosasco, T. A Poggio. 2016. Holographic Embeddings of Knowledge Graphs. In The 30th AAAI Conference on Artificial Intelligence, pp.1955–1961
  • 43.Movielens dataset, 1997, https://grouplens.org/datasets/movielens/.
  • 44.Book-crossing dataset, 2004, http://www2.informatik.Uni.freiburg.de/~cziegler/BX/.
  • 45.Last.fm online music system, 2002, http://www.last.fm/.
  • 46.Microsoft Satori, https://searchengineland.com/library/bing/bing-satori
  • 47.Rendle S. 2012. Factorization machines with libfm. ACM Transactions on Intelligent Systems and Technology vol. 3/no. 3, (2012), pp. 1–22 [Google Scholar]
  • 48.Cheng H., Koc L., Harmsen J., Shaked T., Chandra T., Aradhye H., et al. 2016. Wide & deep learning for recommender systems. ACM International Conference Proceeding Series, vol. 15-/(2016), pp., 7–10. [Google Scholar]
  • 49.X. Yu, X. Ren, Y. Sun, Q. Gu, B. Sturt, U. Khandelwal, et al. 2014. Personalized entity recommendation: A heterogeneous information network approach. In WSDM, (2014), pp. 283–292.
  • 50.Yousuf MI, Kim S. 2020. Guided sampling for large graphs. Data mining and knowledge discovery. 2020;34(4):905–48. [Google Scholar]

Decision Letter 0

Qi Zhao

7 Sep 2021

PONE-D-21-25428MNI: An enhanced multi-task neighborhood interaction model for recommendation on knowledge graphPLOS ONE

Dear Dr. Hao Zhang,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Oct. 6th. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Qi Zhao

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for your submission to PLOS ONE. Before we can proceed, we kindly ask you to address the following concerns:

Upon internal assessment of your manuscript, we found that it is similar to one of your previous works published in PLOS ONE: "Multitask Feature Learning Approach for Knowledge Graph Enhanced Recommendations with RippleNet".

PLOS policy specifies that if a submitted study is similar to previous work, as appears to be the case in this instance, authors should provide a sound scientific rationale for the submitted work and clearly outline how the new submission differs from the past work. The manuscript should also include, if appropriate, a discussion of the current study in context of previous work. We note that the previous work was not cited in your list of references, nor was it included in the section "Related Works". We would expect it to be included in both of these sections.

We also note that numerous citations in your list of references are conference proceedings. While we do allow for this, we would expect the majority of your statements to be supported by reference to peer-reviewed literature. As such, we suggest that you support your statements with a greater number of peer-reviewed citations.

Furthermore, you have used LastFM, MovieLens and Book-Crossing datasets; however, we ask you to provide the URLs for these sites, and to additionally provide the search terms or other details needed to extract the exact data used in the study. The manuscript should provide sufficient details such that any reader could readily replicate the results of your study. Without details on how you accessed the exact data, this is not possible.

We appreciate your attention to these queries and look forward to your response.

3. We suggest you thoroughly copyedit your manuscript for language usage, spelling, and grammar. If you do not know anyone who can help you do this, you may wish to consider employing a professional scientific editing service. 

Whilst you may use any professional scientific editing service of your choice, PLOS has partnered with both American Journal Experts (AJE) and Editage to provide discounted services to PLOS authors. Both organizations have experience helping authors meet PLOS guidelines and can provide language editing, translation, manuscript formatting, and figure formatting to ensure your manuscript meets our submission guidelines. To take advantage of our partnership with AJE, visit the AJE website (http://learn.aje.com/plos/) for a 15% discount off AJE services. To take advantage of our partnership with Editage, visit the Editage website (www.editage.com) and enter referral code PLOSEDIT for a 15% discount off Editage services.  If the PLOS editorial team finds any language issues in text that either AJE or Editage has edited, the service provider will re-edit the text for free.

Upon resubmission, please provide the following:

The name of the colleague or the details of the professional service that edited your manuscript

A copy of your manuscript showing your changes by either highlighting them or using track changes (uploaded as a *supporting information* file)

A clean copy of the edited manuscript (uploaded as the new *manuscript* file)”

4. Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This paper describes a recommendation system on knowledge graph, which deploys neighbor-neighbor interactions to explore high-order relations. Then the algorithm applies a cross & compress unit to combine the recommendation system and knowledge graph, whereby those two modules can interact and share latent features to improve the recommendation precision.

Pros:

The algorithm explains well understood and the experiment supports the algorithm assumption and shows improvement compared with other baselines.

Cons:

1.Figure 2 is the whole structure of the framework, however, needs more explanations in the paragraph.

2.There are some mistakes of English grammar in the abstract and the introduction part, needs to correct them. For example, Therefore, researchers tend to integrate side information which includes user… “which” should be changed to “that”

3.The future work should be discussed in detail.

4.The caption of Figure 1 is not self-explained. The caption should state which one is user, item, or attribute.

5.In the experiment, Figure 5-7 the yellow line is not clear enough to recognize. Also, the captions should be in more detail.

Reviewer #2: The algorithm MNI adopts the separation of recommendation system and knowledge graph, and use a bridge unit to share each latent feature. It also applies neighbor-neighbor reconstruction instead of user-item interaction, containing more information about user-user interaction, item-item interaction, and a deep understanding of the high-order relations.

However, some minor correction should be made:

1. Reference format is not consistent. The authors should double check and revise it.

2. The algorithm 1 on page 10 should be explained more in detail. It is fuzzy when just looking at the algorithm.

3. The blocks and colors in the figure 2 is not well explained. Are there any different meaning about different shapes and colors? The author needs to be clearer.

4. About the recommendation module, the author use 2-layer GAT. Could you explain GAT as for the reader is hard to follow.

5. Minor typos, e.g., Missing space between citation and text.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Zhen Liu

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Oct 28;16(10):e0258410. doi: 10.1371/journal.pone.0258410.r002

Author response to Decision Letter 0


12 Sep 2021

Dear editor,

I have uploaded the responses in three separate files, response to editors, response to reviewer 1 and 2.

Response to editor:

Q1: Style requirements

Response: We have changed the style and added the line number. The title page and the main body are adapted and hopefully consistent with your requirements. If there is something wrong, please contact us and we appreciate your time.

Q2: Similarity to one of your previous works published in PLOS ONE: "Multitask Feature Learning Approach for Knowledge Graph Enhanced Recommendations with RippleNet":

Response: Sorry for the fuzziness of the two works. Both works are improvements based MKR, however, in two different direction. This new submission differs from the previous one on three aspects:

We use neighbor-neighbor interactions to explore the recommendation system, which includes user-item interactions, item-item interactions, item-entity interactions, and user-item interactions as illustrated on Page 8 the recommendation module part. This algorithm leads to more accurate recommendation precision. However, in the previous work, it only consider the interactions between users and historical clicked items using Ripplenet.

We use a reconstruction step to reformulate the recommendation module to G_kig instead of normal recommendation system in the previous work.

We leverage attention mechanism to highlight the important edges between neighbors, in order to keep more meaningful semantics.

Thank you for your advices, we also added additional clarification and the comparison between them in the Related work with highlighted texts and also cited the previous work at [37] in the reference in the manuscript.

Q3: Data access:

Response: Thank you for your suggestion. We added URLs footprints to the data sets of LastFM, Movielens and Book-Crossing on Page 11. Those data can be directed downloaded from the websites and used to replicate our work. But we can also upload our data to the website if requested.

Q4: Peer-reviewed references:

Response: Thank you for your helpful advice to improve the quality of our paper. We add more journal peer-reviewed references and also change some citations’ form. The added reference number are: 9, 13, 14, 15, 16, 34, 37, 41, 50.

Q5: English revision:

Response: We have checked the language, grammar, and spelling by our colleagues, named Tianming Zhao and Jifen Tao. They have made changes to the grammar and the sentence fluency.

Response to reviewer 1:

Point 1: Figure 2 is the whole structure of the framework, however, needs more explanations in the paragraph 


Response 1: Thank you for your advice. We rewrite the first paragraph in section 3 on page 5. We add a detailed procedure for recommendation module and knowledge graph embedding module, pointing out the meaning of symbols and also the color of the blocks.

Point 2: There are some mistakes of English grammar in the abstract and the introduction part, needs to correct them. For example, Therefore, researchers tend to integrate side information which includes user… “which” should be changed to “that”

Response 2: Thank you for your suggestions. We found it very useful to improve our paper. We changed the English grammar carefully in the abstract and the introduction part.

Point 3: The future work should be discussed in detail.

Response 3: We have adopted your suggestion by extending the future works to random walks and implicit relations. We add some discussion to fulfil the future works and it is clearer to understand.

Point 4: The caption of Figure 1 is not self-explained. The caption should state which one is user, item, or attribute.

Response 4: Thank you for your suggestions. We add some explanation about the Figure 1.

Point 5: In the experiment, Figure 5-7 the yellow line is not clear enough to recognize. Also, the captions should be in more detail

Response 5: Thank you for your suggestions. We have changed the yellow line to orange to be clearer. Also, we add some modifications in the captions to be detaied.

Response to reviwer 2:

Point 1: Reference format is not consistent. The authors should double check and revise it

Response 1: Thank you for your suggestions. We have modify the reference format so that they are consistent.

Point 2: The algorithm 1 on page 10 should be explained more in detail. It is fuzzy when just looking at the algorithm.

Response 2: We are grateful for your suggestions. We have added some more explanations to Algorithm 1 so that it is clearer. We explain the whole procedure in steps: reconstruction, recommendation module, cross&compress unit, and KGE module.

Point 3: The blocks and colors in the figure 2 is not well explained. Are there any different meaning about different shapes and colors? The author needs to be clearer.

Response 4: Thank you very much for your comments. We have added some explanations about the colors and shapes in the figure 2 in the first paragraph under Section 3. Also, we modify some explanations about Figure 2 to be specific.

Point 4: About the recommendation module, the author use 2-layer GAT. Could you explain GAT as for the reader is hard to follow.

Response 5: We agree with the reviewers comments. We added some explanations about how the GAT works right before the phrase “2-layer GAT”.

Point 5: Minor typos, e.g., Missing space between citation and text

Response 5: We agree with the reviewers comments.We have double checked through the paper and change the space problem.

Attachment

Submitted filename: author to respond reviewer 2.docx

Decision Letter 1

Qi Zhao

27 Sep 2021

MNI: An enhanced multi-task neighborhood interaction model for recommendation on knowledge graph

PONE-D-21-25428R1

Dear Dr. Hao ZHANG,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Qi Zhao

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have expounded the opinions I put forward in the last round of review, and the data description in the manuscript can also technically support the author's conclusions. It is hoped that the authors will check the English expression before the official publication of the manuscript.

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Qi Zhao

19 Oct 2021

PONE-D-21-25428R1

MNI: An enhanced multi-task neighborhood interaction model for recommendation on knowledge graph

Dear Dr. Zhang:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Qi Zhao

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: author to respond reviewer 2.docx

    Data Availability Statement

    The data underlying the results presented in the study are available from Movielens, Book-Crossing, and Last-FM. Movielens:https://grouplens.org/datasets/movielens/. The data is third party and we confirm that others would be able to access these data in the same manner as the authors. We confirm that the authors did not have any special access privileges that others would not have.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES