当前位置:网站首页>Semi-Decentralized Federated Learning for Cooperative D2D Local Model Aggregation
Semi-Decentralized Federated Learning for Cooperative D2D Local Model Aggregation
2022-08-05 02:19:00 【carrion】
Summary
Proposed Dual Time Scale Hybrid Federated Learning (TT-HF), a semi-decentralized learning architecture that combines the traditional device-to-server communication paradigm for federated learning with a device-to-device (D2D) communication is combined.
During each global aggregation interval, a device
(i) performs multiple iterations of stochastic gradient descent on its single dataset,
(ii) communicates via cooperative distributed D2D within the local cluster, aperiodicallyparticipate in the negotiation process of its model parameters.
Developed an adaptive control algorithm that can adjust the step size of TT-HF, the number of D2D communication rounds and the global aggregation period over time to achieve a sub-linear convergence speed of O(1/t) while minimizing theoptimize network resource utilization.
Traditional Federated Learning
In each iteration, each device trains a local model based on its own dataset, usually using (stochastic) gradient descent.The device then uploads its local model to the server, which aggregates it into a global model, usually using a weighted average, and synchronizes the device with this new model to initiate the next round of local training.
To reduce the cost of uplink and downlink transmission, a combination of local model training and periodic but infrequent global aggregation is proposed.
However, local datasets may exhibit significant heterogeneity in their statistical distributions, resulting in learning models that may be biased towards the local datasets, thus reducing the accuracy of the global model.
TT_HF
(i) involves a mix between device-to-device and device-to-server communication
(ii) contains two timescales for model training: stochastic gradient descent iterations and clustering on a single deviceMulti-round collaborative D2D communication within.
During global aggregation, only one device in the cluster needs to upload the cluster model to the server,
Specifically, during the local update interval of federated learning, a device can systematically share its model parameters with other devices in its vicinity, to form a distributed consensus among each cluster of edge devices.
At the end of each local training interval, assuming that each device's model now reflects the consistency of its cluster, the master can randomly select a device from each cluster forGlobal Aggregation.
Quantify the relationship between device-level stochastic gradient updates, cluster-level consensus processes, and network-level global aggregation.and use them to adjust the length of each local update and consensus cycle.
Results: A version of federated learning that optimizes global model convergence properties while minimizing uplink communication requirements in the system.
Recommended Literature: A Comprehensive Understanding of Federated Learning
[28] S. Abdulrahman, H. Tout, H. Ould-Slimane, A. Mourad, C. Talhi, and
M. Guizani, “A survey onfederated learning: The journey from central-
ized to distributed on-site learning and beyond,” IEEE Internet Things
J., vol. 8, no. 7, pp. 5476–5497, Apr. 2021.
[29] T. Li, A. K. Sahu, A. Talwalkar, and V. Smith, “Federated learning:
Challenges, methods, and future directions,” IEEE Signal Process. Mag.,
vol. 37, no. 3, pp. 50–60, May 2020.
In terms of wireless communication efficiency, several works have investigated the impact of performing multiple rounds of local gradient updates between successive global aggregations, including optimizing the aggregation period according to the total resource budget.
To further reduce the need for global aggregation, [31] proposed a hierarchical system model for federated learning, where edge servers are used for partial global aggregation.
[31] L. Liu, J. Zhang, S. Song, and K. B. Letaief, “Client-edge-cloud hierarchical federated learning,” in Proc. IEEE Int. Conf. Commun.
(ICC),vol. 2020, pp. 1–6.
Model quantization[] and sparsification techniques are also proposed.
This paper proposes a semi-decentralized architecture in which D2D communication is used to exchange model parameters between nodes together with global aggregation.
For data heterogeneity, in [34], the authors propose to upload a portion of the local dataset to the server, which is then used to augment global model training.
[34] N. Y oshida, T. Nishio, M. Morikura, K. Yamamoto, and R. Yo onetani,
“Hybrid-FL for wireless networks: Cooperative learning mechanism
using non-IID data,” in Proc. IEEE Int. Conf. Commun. (ICC), Jun. 2020,
pp. 1–7.
this articleUsing D2D communication to exchange between devicesmodel parameters, which alleviates such concerns.
The distributed collaborative learning among devices is introduced into the local update process, forming a new system architecture with D2D reinforcement learning.
Addresses both communication efficiency and data heterogeneity challenges.
In this case
(i) the device may perform multiple (stochastic) gradient iterations between global aggregations,
(ii) the global aggregation is aperiodic,
(iii) the deviceThe consensus process among them may happen aperiodically during each global aggregation.
Due to device mobility, the topology of each cluster (i.e. the number of nodes and their location within the cluster) may change over time, although we assume that compared to the time between two global aggregates,This evolution is slow.
Triangular inequality is used for the upper bound on the gradient...Smoothness Condition
TT-HF: Dual Time Scale Hybrid Federated Learning
1) Overview and rationale: consists of a series of local model training intervals between aperiodic global aggregations.During each interval, the device performs local stochastic gradient descent (SGD) iterations and aperiodically synchronizes its model parameters through a local consistency process within the cluster.
Looking at it later, there are too many formulas.....
边栏推荐
- 协作D2D局部模型聚合的半分散联合学习
- 01 [Foreword Basic Use Core Concepts]
- 直播回放含 PPT 下载|基于 Flink & DeepRec 构建 Online Deep Learning
- [Endnote] Word inserts a custom form of Endnote document format
- 【MySQL series】- Does LIKE query start with % will make the index invalid?
- Pisanix v0.2.0 发布|新增动态读写分离支持
- 常见的硬件延迟
- The 2022 EdgeX China Challenge will be grandly opened on August 3
- RAID disk array
- [ROS] (10) ROS Communication - Service Communication
猜你喜欢

Utilities ![Chapter 09 Use of Performance Analysis Tools [2. Index and Tuning] [MySQL Advanced]](/img/e7/61863c356e4c7bb03ba27865ee7be0.png)
Chapter 09 Use of Performance Analysis Tools [2. Index and Tuning] [MySQL Advanced]

程序员失眠时的数羊列表 | 每日趣闻

source program in assembly language

01 【前言 基础使用 核心概念】
![[Endnote] Word inserts a custom form of Endnote document format](/img/70/e8a6f15b80e9c53db369fd715e51df.png)
[Endnote] Word inserts a custom form of Endnote document format

直播回放含 PPT 下载|基于 Flink & DeepRec 构建 Online Deep Learning

“嘀哩哩,等灯等灯”,工厂安全生产的提示音

The 2022 EdgeX China Challenge will be grandly opened on August 3

一文看懂推荐系统:召回06:双塔模型——模型结构、训练方法,召回模型是后期融合特征,排序模型是前期融合特征
随机推荐
释放技术创新引擎,英特尔携手生态合作伙伴推动智慧零售蓬勃发展
Programmer's list of sheep counting when insomnia | Daily anecdote
iNFTnews | 对体育行业和球迷来说,NFT可以带来什么?
Greenplum数据库故障分析——能对数据库base文件夹进行软连接嘛?
Residential water problems
Access Characteristics of Constructor under Inheritance Relationship
Greenplum Database Fault Analysis - Can a Soft Connection Be Made to the Database Base Folder?
Oracle encapsulates restful interfaces into views
leetcode 15
"Dilili, wait for the lights, wait for the lights", the prompt sound for safe production in the factory
浅谈数据安全治理与隐私计算
【genius_platform软件平台开发】第七十六讲:vs预处理器定义的牛逼写法!!!!(其他组牛逼conding人员告知这么配置来取消宏定义)
Opening - Open a new .NET modern application development experience
后期学习计划
mysql树状结构查询问题
RAID disk array
SuperMap iDesktop.Net之布尔运算求交——修复含拓扑错误复杂模型
[Endnote] Word inserts a custom form of Endnote document format
进程在用户态和内核态的区别[独家解析]
Domain Driven Design - MDD