论文导读:关于联邦图神经网络的3篇文章
联邦图神经网络:概述、技术和挑战
论文信息
原文地址:https://arxiv.org/abs/2202.07256
摘要
With its powerful capability to deal with graph data widely found in practical applications, graph neural networks (GNNs) have received significant research attention. However, as societies become in-creasingly concerned with data privacy, GNNs face the need to adapt to this new normal. This has led to the rapid development of federated graph neural networks (FedGNNs) research in recent years.Although promising, this interdisciplinary field is highly challenging for interested researchers to enter into. The lack of an insightful survey on this topic only exacerbates this problem. In this paper, we bridge this gap by offering a comprehensive survey of this emerging field. We propose a unique 3-tiered taxonomy of the FedGNNs literature to provide a clear view into how GNNs work in the context of Federated Learning (FL). It puts existing works into perspective by analyzing how graph data manifest themselves in FL settings, how GNN training is performed under different FL system architectures and degrees of graph data overlapacross data silo, and how GNN aggregation is performed under various FL settings. Through discussions of the advantages and limitations of existingworks, we envision future research directions that can help build more robust, dynamic, efficient, and interpretable FedGNNs.
图神经网络( GNNs )凭借其强大的处理实际应用中广泛存在的图数据的能力,受到了广泛的研究关注。然而,随着社会越来越关注数据隐私,GNNs面临着适应这种新常态的需要。这导致了近年来联邦图神经网络( FedGNNs )研究的快速发展。虽然前景广阔,但这一跨学科领域感兴趣的研究者来说是极具挑战性的。对这一领域缺乏深入的调查只会加剧这一问题。在本文中,我们通过提供对这一新兴领域的全面调查来弥补这一差距。我们对FedGNNs文献提出了一个独特的3层分类法,以提供对GNNs如何在联邦学习( Federation Learning,FL )背景下工作的清晰视图。通过分析图数据在FL设置中的表现形式,不同FL系统架构和图数据跨数据仓重叠程度下如何进行GNN训练,以及不同FL设置下如何进行GNN聚合,对现有工作进行了展望。通过讨论现有工作的优势和局限性,我们展望了未来的研究方向,这些方向可以帮助构建更健壮、更动态、更高效和更易于解释的联邦图神经网络。
个性化的联邦图神经网络框架
论文信息
原文地址:https://doi.org/10.1038/s41467-022-30714-9
Graph neural network (GNN) is effective in modeling high-order interactions and has been widely used in various personalized applications such as recommendation. However, mainstream personalization methods rely on centralized GNN learning on global graphs, which have considerable privacy risks due to the privacy-sensitive nature of user data. Here, we present a federated GNN framework named FedPerGNN for both effective and privacy-preserving personalization. Through a privacy-preserving model update method, we can collaboratively train GNN models based on decentralized graphs inferred from local data. To further exploit graph information beyond local interactions, we introduce a privacy-preserving graph expansion protocol to incorporate high-order information under privacy protection. Experimental results on six datasets for personalization in different scenarios show that FedPerGNN achieves 4.0% ~ 9.6% lower errors than the state-of-the-art federated personalization methods under good privacy protection. FedPerGNN provides a promising direction to mining decentralized graph data in a privacy-preserving manner for responsible and intelligent personalization.
图神经网络( Graph Neural Network,GNN )在建模高阶交互方面十分有效,已被广泛应用于推荐等各种个性化应用中。然而,主流的个性化方法依赖于全局图上的集中式GNN学习,由于用户数据的隐私敏感特性,这些方法存在相当大的隐私风险。在这里,我们提出了一个名为FedPerGNN的联邦GNN框架,用于有效和隐私保护的个性化。通过一种保护隐私的模型更新方法,我们可以基于从本地数据推断的分散图协同训练GNN模型。为了进一步利用局部交互之外的图信息,引入了一个隐私保护的图扩展协议,以在隐私保护下纳入高阶信息。在6个不同场景下的个性化数据集上的实验结果表明,在良好的隐私保护下,FedPerGNN比现有的联合个性化方法降低了4.0 % ~ 9.6 %的误差。FedPerGNN以保护隐私的方式挖掘分散的图数据,为智能个性化提供了一个很有前途的方向。
主要内容
去中心化的联邦图神经网络
论文信息
原文地址:https://federated-learning.org/fl-ijcai-2021/FTL-IJCAI21_paper_20.pdf
摘要
Due to the increasing importance of user-side pri-vacy, federated graph neural networks are pro-posed recently as a reliable solution for privacy-preserving data analysis by enabling the jointtraining of a graph neural network from graph-structured data located at different clients. How-ever, existing federated graph neural networks arebased on a centralized server to orchestrate thetraining process, which is unacceptable in manyreal-world applications such as building financialrisk control models across competitive banks. Inthis paper, we propose a new Decentralized Feder-ated Graph Neural Network (D-FedGNN for short)which allows multiple participants to train a graphneural network model without a centralized server.Specifically, D-FedGNN uses a decentralized par-allel stochastic gradient descent algorithm DP-SGD to train the graph neural network model ina peer-to-peer network structure. To protect privacy during model aggregation, D-FedGNN intro-duces the Diffie-Hellman key exchange method toachieve secure model aggregation between clients.Both theoretical and empirical studies show that theproposed D-FedGNN model is capable of achiev-ing competitive results compared with traditionalcentralized federated graph neural networks in thetasks of classification and regression, as well as significantly reducing time cost of the communication of model parameters.
主要内容
从理论和实证两方面研究了D - FedGNN的性能。
原文链接:https://blog.csdn.net/weixin_43598687/article/details/126913574
往期推荐