Graph neural networks (GNN) have been widely deployed in real-world networked applications and systems due to their capability to handle graph-structured data. However, the growing awareness of data privacy severely challenges the traditional centralized model training paradigm, where a server holds all the graph information. Federated learning is an emerging collaborative computing paradigm that allows model training without data centralization. Existing federated GNN studies mainly focus on systems where clients hold distinctive graphs or sub-graphs. The practical node-level federated situation, where each client is only aware of its direct neighbors, has yet to be studied. In this paper, we propose the first federated GNN framework called Lumos that supports supervised and unsupervised learning with feature and degree protection on node-level federated graphs. We first design a tree constructor to improve the representation capability given the limited structural information. We further present a Monte Carlo Markov Chain-based algorithm to mitigate the workload imbalance caused by degree heterogeneity with theoretically-guaranteed performance. Based on the constructed tree for each client, a decentralized tree-based GNN trainer is proposed to support versatile training. Extensive experiments demonstrate that Lumos outperforms the baseline with significantly higher accuracy and greatly reduced communication cost and training time.
翻译:摘要:图神经网络因其处理图结构数据的能力,已被广泛应用于现实世界的网络化应用与系统中。然而,日益增强的数据隐私意识严重挑战了传统的集中式模型训练范式——即服务器持有全部图信息。联邦学习是一种新兴的协作计算范式,允许在不集中数据的情况下进行模型训练。现有联邦图神经网络研究主要集中于客户端持有不同图或子图的系统。而实际场景中的节点级联邦情境(即每个客户端仅知晓其直接邻居)尚未得到充分研究。本文首次提出名为Lumos的联邦图神经网络框架,该框架在节点级联邦图中支持具有特征与度数保护的有监督与无监督学习。我们首先设计了一棵树形构造器,以在有限结构信息下提升表征能力;随后提出一种基于蒙特卡洛马尔可夫链的算法,以缓解由度数异质性引起的工作负载不均衡问题,并具备理论保证的性能。基于为每个客户端构造的树形结构,我们进一步提出一种去中心化的树形图神经网络训练器,以支持多样化训练。大量实验表明,Lumos在显著提升准确率的同时,大幅降低了通信成本与训练时间,性能优于基线方法。