|
Frontiers of Information Technology & Electronic Engineering
ISSN 2095-9184 (print), ISSN 2095-9230 (online)
2025 Vol.26 No.8 P.1441-1453
E-CGL: an efficient continual graph learner
Abstract: Continual learning (CL) has emerged as a crucial paradigm for learning from sequential data while retaining previous knowledge. Continual graph learning (CGL), characterized by dynamically evolving graphs from streaming data, presents distinct challenges that demand efficient algorithms to prevent catastrophic forgetting. The first challenge stems from the interdependencies between different graph data, in which previous graphs influence new data distributions. The second challenge is handling large graphs in an efficient manner. To address these challenges, we propose an efficient continual graph learner (E-CGL) in this paper. We address the interdependence issue by demonstrating the effectiveness of replay strategies and introducing a combined sampling approach that considers both node importance and diversity. To improve efficiency, E-CGL leverages a simple yet effective multi-layer perceptron (MLP) model that shares weights with a graph neural network (GNN) during training, thereby accelerating computation by circumventing the expensive message-passing process. Our method achieves state-of-the-art results on four CGL datasets under two settings, while significantly lowering the catastrophic forgetting value to an average of -1.1%. Additionally, E-CGL achieves the training and inference speedup by an average of 15.83× and 4.89×, respectively, across four datasets. These results indicate that E-CGL not only effectively manages correlations between different graph data during continual training but also enhances efficiency in large-scale CGL.
Key words: Graph neural networks; Continual learning; Dynamic graphs; Continual graph learning; Graph acceleration
浙江大学计算机科学与技术学院数字媒体计算与设计实验室,中国杭州市,310027
摘要:连续学习已成为从序列数据中学习新知识并保留先前知识的关键范式。图连续学习(CGL)具有流式数据带来的动态演化图特征,其独特挑战要求高效算法以防止灾难性遗忘。首要挑战源于不同图数据间的相互依赖性--历史图数据会影响新数据的特征分布。第二个挑战在于如何高效地处理大规模图数据。为应对这些挑战,本文提出一种高效的图连续学习器(E-CGL)。通过验证回放策略的有效性,提出兼顾节点重要性与多样性的组合采样方法,成功解决图数据相互依赖问题。在效率提升方面,E-CGL采用一种简单而有效的多层感知机模型,该模型与图神经网络共享权重,在训练过程中解耦耗时的消息传递机制实现计算加速。本方法在两种实验设置下的4个数据集上取得先进的成果,同时将灾难性遗忘率显著降低至−1.1%的平均水平。此外,在4个数据集上,E-CGL的训练与推理速度分别提升了15.83倍和4.89倍。这些结果表明,E-CGL不仅在模型更新过程中有效保留图数据间的关联性,更在大规模图连续学习场景中显著提升效率。
关键词组:
References:
Open peer comments: Debate/Discuss/Question/Opinion
<1>
DOI:
10.1631/FITEE.2500162
CLC number:
TP181
Download Full Text:
Downloaded:
391
Download summary:
<Click Here>Downloaded:
204Clicked:
417
Cited:
0
On-line Access:
2025-06-04
Received:
2025-05-14
Revision Accepted:
2025-06-03
Crosschecked:
2025-09-04