论文下载地址: https://doi.org/10.1145/3485447.3512104
发表期刊:WWW
Publish time: 2022
作者及单位:
数据集: 正文中的介绍
代码:
其他:
其他人写的文章
简要概括创新点: 可以看做是对positive contrastive pairs的改进,negative还是 uniformly smaples
(1) 缝合了(i)structual space/neighbors (ii)semantic space/neighbors
(2)分别对应 structure-contrastive objective(用GNN提取,视为正对比对) + prototype-contrastive objective(用K-Means)
(3)用EM方法优化(因为这一过程无法进行端到端的优化,其实K-Means本就是EM算法的一个例子)
(4)是一个model-agnostic constative learning framework
(5)Loss 是3个loss的权重和,要用multi-task learning strategy
(6) 没用social, user-user靠 偶数步
- we propose a novel contrastive learning approach, named Neighborhood-enriched Contrastive Learning, named NCL, which explicitly incorporates the potential neighbors into contrastive pairs. (为了解决上述问题,我们提出了一种新的对比学习方法,称为邻域丰富对比学习(NCL),它将潜在的邻域明确地整合到对比对中。)
- Specifically, we introduce the neighbors of a user (or an item) from graph structure and semantic space respectively. (具体来说,我们分别从图结构和语义空间引入用户(或项目)的邻居。)
- For the structural neighbors on the interaction graph, we develop a novel structure-contrastive objective that regards users (or items) and their structural neighbors as positive contrastive pairs. (对于交互图上的结构邻居,我们提出了一种新的结构对比目标,将用户(或项目)及其结构邻居视为正对比对。)
- In implementation, the representations of users (or items) and neighbors correspond to the outputs of different GNN layers. (在实现中,用户(或项目)和邻居的表示对应于不同GNN层的输出。)
- Furthermore, to excavate the potential neighbor relation in semantic space, we assume that users with similar representations are within the semantic neighborhood, and incorporate these semantic neighbors into the prototype-contrastive objective. (此外,为了挖掘语义空间中潜在的邻域关系,我们假设具有相似表示的用户位于语义邻域内,并将这些语义邻域合并到原型对比目标中。)
- The proposed NCL can be optimized with EM algorithm and generalized to apply to graph collaborative filtering methods. (所提出的NCL可以用EM算法进行优化,并可推广应用于图协同过滤方法。)
- We propose a model-agnostic contrastive learning framework named NCL, which incorporates both structural and semantic neighbors for improving the neural graph collaborative filtering. (我们提出了一个模型不可知的对比学习框架NCL,它结合了结构和语义邻域来改进神经图协同过滤。)
- We propose to learn representative embeddings for both kinds of neighbors, such that the constative learning can be only performed between a node and the corresponding representative embeddings, which largely improves the algorithm efficiency. (我们建议学习这两种邻居的代表性嵌入,这样只能在一个节点和相应的代表性嵌入之间进行约束学习,这大大提高了算法的效率。)
(1) Recently, graph collaborative filtering methods have been proposed as an effective recommendation approach, which can capture users’ preference over items by modeling the user-item interaction graphs. (近年来,图协同过滤方法作为一种有效的推荐方法被提出,它通过对用户-项目交互图的建模来捕获用户对项目的偏好。)
(2) To tackle the above issue, we propose a novel contrastive learning approach, named Neighborhood-enriched Contrastive Learning, named NCL, which explicitly incorporates the potential neighbors into contrastive pairs. (为了解决上述问题,我们提出了一种新的对比学习方法,称为邻域丰富对比学习(NCL),它将潜在的邻域明确地整合到对比对中。)
• Information systems → Recommender systems.
Recommender System, Collaborative Filtering, Contrastive Learning, Graph Neural Network
(1) In the age of information explosion, recommender systems occupy an important position to discover users’ preferences and deliver online services efficiently [23]. (在信息爆炸的时代,推荐系统在发现用户偏好和高效提供在线服务方面占据着重要地位[23]。)
(2) Despite the remarkable success, existing neural graph collaborative filtering methods still suffer from two major issues. (尽管取得了显著的成功,但现有的神经图协同过滤方法仍然存在两个主要问题。)
(3) Besides direct user-item interactions, there exist multiple kinds of potential relations (e.g., user similarity) that are useful to the recommendation task, and we aim to design more effective constative learning approaches for leveraging such useful relations in neural graph collaborative filtering. (除了直接的用户项交互,还存在多种对推荐任务有用的潜在关系(例如,用户相似性),我们旨在设计更有效的约束学习方法,以便在神经图协同过滤中利用这些有用的关系。)
(4) To integrate and model the enriched neighborhood, we propose Neighborhood-enriched Contrastive Learning (NCL for short), a model-agnostic constative learning framework for the recommendation.
(5) To be specific,
(6) Our contributions can be summarized threefold: (我们的贡献可以概括为三个方面:)
(1) As the fundamental recommender system, collaborative filtering (CF) aims to recommend relevant items that users might be interested in based on the observed implicit feedback (e.g., expression, click and transaction). (作为最基本的推荐系统,协同过滤(CF)旨在根据观察到的隐性反馈(如表达、点击和交易)推荐用户可能感兴趣的相关项目。)
(2) In general, GNN-based collaborative filtering methods [9, 31, 32] produce informative representations for users and items based on the aggregation scheme, which can be formulated to two stages: (一般来说,基于GNN的协同过滤方法[9,31,32]基于聚合方案为用户和项目生成信息表示,可分为两个阶段:)
(1) As mentioned in Section 2, GNN-based methods produce user and item representations by applying the propagation and prediction function on the interaction graph G \mathcal{G} G. (如第2节所述,基于GNN的方法通过在交互图 G \mathcal{G} G上应用传播和预测函数来生成用户和项目表示。)
(2) After propagating with L L L layers, we adopt the weighted sum function as the readout function to combine the representations of all layers and obtain the final representations as follows:(在使用 L L L层进行传播后,我们采用加权和函数作为读出函数,组合所有层的表示,并获得如下最终表示:)
(3) With the final representations, we adopt inner product to predict how likely a user u u u would interact with items i i i: (在最终表述中,我们采用内积来预测用户 u u u与第 i i i项交互的可能性:)
(4) To capture the information from interactions directly, we adopt Bayesian Personalized Ranking (BPR) loss [22], which is a welldesigned ranking objective function for recommendation. (为了直接从交互中获取信息,我们采用了贝叶斯个性化排名(BPR)损失[22],这是一个精心设计的推荐排名目标函数。)
(5) By optimizing the BPR loss L B P R L_{BPR} LBPR, our proposed NCL can model the interactions between users and items. (通过优化BPR损失 L B P R L_{BPR} LBPR , 我们提出的NCL可以对用户和项目之间的交互进行建模。)
(1) Existing graph collaborative filtering models are mainly trained with the observed interactions (e.g., user-item pairs), while the potential relationships among users or items cannot be explicitly captured by learning from the observed data. (现有的图协同过滤模型主要是利用观察到的交互(如用户-项目对)进行训练,而用户或项目之间的潜在关系不能通过从观察到的数据中学习来明确捕获。)
(2) Considering that the interaction graph G \mathcal{G} G is a bipartite graph, information propagation with GNN-based model for even times on the graph naturally aggregates information of homogeneous structural neighbors which makes it convenient to extract the potential neighbors within users or items. (考虑到交互图 G \mathcal{G} G是一个二部图,基于GNN模型的偶数次信息传播自然地聚集了同质结构邻居的信息,这便于提取用户或项目中的潜在邻居。)
(3) In a similar way, the structure-contrastive loss of the item side L s t r u c i t e m L^{item}_{struc} Lstrucitem can be obtained as: (以类似的方式,项目侧的结构对比损失 L s t r u c i t e m L^{item}_{struc} Lstrucitem可通过以下方式获得:)
(4) And the complete structure-contrastive objective function is the weighted sum of the above two losses: (完整的结构对比目标函数是上述两种损失的加权和:)
(1) The structure-contrastive loss explicitly excavates the neighbors defined by the interaction graph. (结构对比损失明确挖掘由交互图定义的邻域。)
(2) Inspired by previous works [16], we can identify the semantic neighbors by learning the latent prototype for each user and item. (受之前工作[16]的启发,我们可以通过学习每个用户和项目的潜在原型来识别语义邻居)
(3) After that, the proposed prototype-contrastive learning objective is to minimize the following function based on InfoNCE [20]: (之后,提出的原型对比学习目标是基于InfoNCE最小化以下功能[20]:)
(4) The objective on the item side is identical: (项目侧的目标是相同的)
(5) The final prototype-contrastive objective is the weighted sum of user objective and item objective: (最终原型对比目标是用户目标和项目目标的加权和:)
(6) In this way, we explicitly incorporate the semantic neighbors of users/items into contrastive learning to alleviate the data sparsity. (通过这种方式,我们明确地将用户/项目的语义邻居纳入对比学习中,以缓解数据稀疏性)
(1) As Eq. (9) is hard to optimize, we obtain its Lower-Bound (LB) by Jensen’s inequality:
(2) In the E-step, e u e_u eu is fixed and Q ( c i ∣ e u ) Q (c_i | e_u) Q(ci∣eu) can be estimated by K-means algorithm over the embeddings of all users E E E. (可以通过K-均值算法在所有用户 E E E的嵌入上进行估计)
(3) In the M-step, the target function can be rewritten with Q ^ ( c i ∣ e u ) \hat{Q} (c_i | e_u) Q^(ci∣eu): (在M步中,目标函数可以写为)
(4) we can assume that the distrubution of users is isotropic Gaussian over all the clusters. So the function can be written as: (我们可以假设用户在所有集群上的分布是各向同性的高斯分布。所以这个函数可以写成)
(1) For graph collaborative filtering, the construction of neighborhood is more important than other collaborative filtering methods [36], since it is based on the graph structure. (对于图协同过滤,邻域的构建比其他协同过滤方法更重要[36],因为它基于图结构。)
(2) Furthermore, semantic neighbors have seldom been explored in GNNs for recommendation, while semantic neighbors are necessary to be considered for graph collaborative filtering due to the sparse, noisy interaction graphs. (此外,在GNNs中,语义邻居很少被用于推荐,而由于交互图稀疏、有噪声,在图协同过滤中需要考虑语义邻居。)
We compare the proposed method with the following baseline methods. (我们将提出的方法与以下基线方法进行比较)
Table 2 shows the performance comparison of the proposed NCL and other baseline methods on five datasets. From the table, we find several observations: (表2显示了提出的NCL和其他基线方法在五个数据集上的性能比较。从表中,我们发现了几个观察结果:)
(1) Compared to the traditional methods, such as BPRMF, GNN-based methods outperform as they encode the high-order information of bipartite graphs into representations. (与传统的方法(如BPRMF)相比,基于GNN的方法在将二部图的高阶信息编码为表示时表现更好。)
(2) For the self-supervised method, SGL [33] consistently outperforms other supervised methods on five datasets, which shows the effectiveness of contrastive learning for improving the recommendation performance. (对于自监督方法,SGL [33]在五个数据集上始终优于其他监督方法,这表明了对比学习对于提高推荐性能的有效性。)
(3) Finally, we can see that the proposed NCL consistently performs better than baselines. (最后,我们可以看到,提议的NCL始终比基线表现更好。)
(1) In NCL , the structural neighbors correspond to different layers of GNN. To investigate the impact of different structural neighbors, we select the nodes in one-, two-, and three-hop as the structural neighbors and test the effectiveness when incorporating them with contrastive learning. (在NCL中,结构邻居对应于GNN的不同层。为了研究不同结构邻居的影响,我们选择一跳、二跳和三跳中的节点作为结构邻居,并测试它们与对比学习结合时的有效性。)
(2) The results are shown in Table 3. We can find that the three variants of NCL all perform similar or better than LightGCN, which further indicates the effectiveness of the proposed hop-contrastive strategy. (结果如表3所示。我们可以发现,NCL的三个变体的性能都与LightGCN相似或更好,这进一步表明了所提出的hop对比策略的有效性。)
(1) As the proposed NCL architecture is model agnostic, we further test its performance with other GNN architectures. The results are reported in Table 4. (由于所提出的NCL体系结构是模型无关的,因此我们将用其他GNN体系结构进一步测试其性能。结果见表4。)
(2) From this table, we can observe that the proposed method can consistently improve the performance of NGCF, DGCF, and LightGCN, which further verifies the effectiveness of the proposed method. (从这个表中,我们可以观察到,所提出的方法可以持续改善NGCF、DGCF和LightGCN的性能,这进一步验证了所提出方法的有效性。)
(1) In this work, we propose a novel contrastive learning paradigm, named Neighborhood-enriched Contrastive Learning (NCL), to explicitly capture potential node relatedness into contrastive learning for graph collaborative filtering. (在这项工作中,我们提出了一种新的对比学习范式,称为邻域丰富的对比学习(NCL),以明确地捕捉潜在的节点关联性,用于图协同过滤的对比学习)
(2) As future work, we will extend our framework to other recommendation tasks, such as sequential recommendation. (作为未来的工作,我们将把我们的框架扩展到其他推荐任务,例如顺序推荐。)