多视图对比增强的异质图结构学习方法
作者:
作者单位:

作者简介:

邴睿(1994-),男,博士生,CCF学生会员,主要研究领域为图数据挖掘;王森章(1986-),男,博士,教授,博士生导师,香江学者,CCF高级会员,主要研究领域为时空数据挖掘,图数据挖掘,深度学习,城市计算.袁冠(1982-),男,博士,教授,博士生导师,CCF高级会员,主要研究领域为人工智能,大数据技术,机器学习,计算智能;乔少杰(1981-),男,博士,教授,博士生导师,CCF杰出会员,主要研究领域为时空数据库,城市计算,人工智能;孟凡荣(1962-),女,博士,教授,博士生导师,CCF高级会员,主要研究领域为数据库技术,数据挖掘;王志晓(1979-),男,博士,教授,博士生导师,CCF专业会员,主要研究领域为社交网络分析,数据挖掘.

通讯作者:

袁冠,E-mail:yuanguan@cumt.edu.cn

中图分类号:

基金项目:

国家自然科学基金(62272461,71774159,62272066,61871686);中国博士后科学基金(2021T140707);江苏省博士后科学基金(2021K565C)


Multi-view Contrastive Enhanced Heterogeneous Graph Structure Learning
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    异质图神经网络作为一种异质图表示学习的方法,可以有效地抽取异质图中的复杂结构与语义信息,在节点分类和连接预测任务上取得了优异的表现,为知识图谱的表示与分析提供了有力的支撑.现有的异质图由于存在一定的噪声交互或缺失部分交互,导致异质图神经网络在节点聚合、更新时融入错误的邻域特征信息,从而影响模型的整体性能.为解决该问题,提出了多视图对比增强的异质图结构学习模型.该模型首先利用元路径保持异质图中的语义信息,并通过计算每条元路径下节点之间特征相似度生成相似度图,将其与元路径图融合,实现对图结构的优化.通过将相似度图与元路径图作为不同视图进行多视图对比,实现无监督信息的情况下优化图结构,摆脱对监督信号的依赖.最后,为解决神经网络模型在训练初期学习能力不足、生成的图结构中往往存在错误交互的问题,设计了一个渐进式的图结构融合方法.通过将元路径图和相似度图递增地加权相加,改变图结构融合过程中相似度图所占的比例,在抑制了因模型学习能力弱引入过多的错误交互的同时,达到了用相似度图中的交互抑制原有干扰交互或补全缺失交互的目的,实现了对异质图结构的优化.选择节点分类与节点聚类作为图结构学习的验证任务,在4种真实异质网络数据集上的实验结果,也表明该异质图结构学习方法是可行且有效的.与最优对比模型相比,该模型在两种任务下的性能均有显著提升.

    Abstract:

    As a heterogeneous graph representation learning method, heterogeneous graph neural networks can effectively extract complex structural and semantic information from heterogeneous graphs, and have achieved excellent performance in node classification and connection prediction tasks, which provides strong support for the representation and analysis of knowledge graphs. Due to the existence of some noise interaction or missing interaction in the heterogeneous graph, the heterogeneous graph neural network incorporates erroneous neighbor features when nodes are aggregated and updated, thus affecting the overall performance of the model. In order to solve the above problems, this study proposes a heterogeneous graph structure learning model enhanced by multi-view contrastive. Firstly, the semantic information in the heterogeneous graph is maintained by using the meta path, and the similarity graph is generated by calculating the feature similarity between the nodes under each meta-path, which is fused with the meta-path graph to optimize the graph structure. By comparing the similarity graph and meta-path graph as different views, the graph structure is optimized without the supervision information, and the dependence on the supervision signal is eliminated. Finally, in order to solve the problem that the learning ability of neural network model is insufficient at the initial stage of training and there are often error interactions in the generated graph structure, this study designs a progressive graph structure fusion method. Through incremental weighted addition of meta-path graph and similarity graph, the weight of similarity graph is changed in the fusion process, it not only prevents erroneous interactions from being introduced in the initial stage of training, but also achieves the purpose of using the interaction in similarity graph to suppress interference interaction or complete missing interaction, thus the structure of heterogeneous graph is optimized. The node classification and node clustering are selected as the verification tasks of graph structure learning. The experimental results on four real heterogeneous graph datasets prove that the heterogeneous graph structure learning method proposed in this study is feasible and effective. Compared with the optimal comparison model, the performance of proposed model has been significantly improved under two evaluation metrics.

    参考文献
    相似文献
    引证文献
引用本文

邴睿,袁冠,孟凡荣,王森章,乔少杰,王志晓.多视图对比增强的异质图结构学习方法.软件学报,2023,34(10):4477-4500

复制
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2022-07-04
  • 最后修改日期:2022-08-18
  • 录用日期:
  • 在线发布日期: 2023-01-13
  • 出版日期: 2023-10-06
文章二维码
您是第位访问者
版权所有:中国科学院软件研究所 京ICP备05046678号-3
地址:北京市海淀区中关村南四街4号,邮政编码:100190
电话:010-62562563 传真:010-62562533 Email:jos@iscas.ac.cn
技术支持:北京勤云科技发展有限公司

京公网安备 11040202500063号