|
Frontiers of Information Technology & Electronic Engineering
ISSN 2095-9184 (print), ISSN 2095-9230 (online)
2024 Vol.25 No.5 P.713-727
Communication efficiency optimization of federated learning for computing and network convergence of 6G networks
Abstract: Federated learning effectively addresses issues such as data privacy by collaborating across participating devices to train global models. However, factors such as network topology and computing power of devices can affect its training or communication process in complex network environments. Computing and network convergence (CNC) of sixth-generation (6G) networks, a new network architecture and paradigm with computing-measurable, perceptible, distributable, dispatchable, and manageable capabilities, can effectively support federated learning training and improve its communication efficiency. By guiding the participating devices’ training in federated learning based on business requirements, resource load, network conditions, and computing power of devices, CNC can reach this goal. In this paper, to improve the communication efficiency of federated learning in complex networks, we study the communication efficiency optimization methods of federated learning for CNC of 6G networks that give decisions on the training process for different network conditions and computing power of participating devices. The simulations address two architectures that exist for devices in federated learning and arrange devices to participate in training based on arithmetic power while achieving optimization of communication efficiency in the process of transferring model parameters. The results show that the methods we proposed can cope well with complex network situations, effectively balance the delay distribution of participating devices for local training, improve the communication efficiency during the transfer of model parameters, and improve the resource utilization in the network.
Key words: Computing and network convergence; Communication efficiency; Federated learning; Two architectures
1北京邮电大学无线信号处理与网络实验室,中国北京市,100876
2中国电信股份有限公司研究院,中国北京市,102209
3中国电信股份有限公司北京分公司,中国北京市,100011
摘要:联邦学习以参与设备之间协作训练全局模型的形式,有效地解决了数据隐私等问题。然而,在复杂的网络环境中,网络拓扑和设备算力等因素极其影响联邦学习的训练和通信过程。作为一种算力可测、可感知、可分配、可调度和可管理的新型网络架构和范式,6G中的算力网络恰好能有效支持联邦学习训练并提高其通信效率。根据业务需求、资源负载、网络条件和设备算力等信息,算力网络可以决策联邦学习的训练进而实现通信效率提高。为了提高复杂网络环境下联邦学习的通信效率,本文研究了其在6G算力网络中的通信效率优化方法,针对不同的网络条件和参与设备的算力作出训练过程的决策。仿真实验基于联邦学习中存在的两种架构,依据算力信息调度设备参与训练,并在传输模型参数的过程中实现通信效率的优化。仿真结果表明,本文提出的方法能够很好地应对复杂的网络情况,有效平衡参与设备的本地训练延迟差异,提高在传输模型参数时的通信效率,并提高网络中的资源利用率。
关键词组:
References:
Open peer comments: Debate/Discuss/Question/Opinion
<1>
DOI:
10.1631/FITEE.2300122
CLC number:
TP393
Download Full Text:
Downloaded:
1859
Download summary:
<Click Here>Downloaded:
274Clicked:
2083
Cited:
0
On-line Access:
2024-08-27
Received:
2023-10-17
Revision Accepted:
2024-05-08
Crosschecked:
2023-10-17