Communication-Efficient Federated Deep Learning with Asynchronous Model Updates and Temporally Weighted Aggregation

Reading time: 3 minute
...

📝 Original Paper Info

- Title: Communication-Efficient Federated Deep Learning with Asynchronous Model Update and Temporally Weighted Aggregation
- ArXiv ID: 1903.07424
- Date: 2020-08-31
- Authors: Yang Chen, Xiaoyan Sun, Yaochu Jin

📝 Abstract

Federated learning obtains a central model on the server by aggregating models trained locally on clients. As a result, federated learning does not require clients to upload their data to the server, thereby preserving the data privacy of the clients. One challenge in federated learning is to reduce the client-server communication since the end devices typically have very limited communication bandwidth. This paper presents an enhanced federated learning technique by proposing a synchronous learning strategy on the clients and a temporally weighted aggregation of the local models on the server. In the asynchronous learning strategy, different layers of the deep neural networks are categorized into shallow and deeps layers and the parameters of the deep layers are updated less frequently than those of the shallow layers. Furthermore, a temporally weighted aggregation strategy is introduced on the server to make use of the previously trained local models, thereby enhancing the accuracy and convergence of the central model. The proposed algorithm is empirically on two datasets with different deep neural networks. Our results demonstrate that the proposed asynchronous federated deep learning outperforms the baseline algorithm both in terms of communication cost and model accuracy.

💡 Summary & Analysis

This paper focuses on reducing the communication cost between clients and servers while training deep neural networks, proposing an asynchronous learning strategy for clients and temporally weighted aggregation for model updates at the server. The central challenge addressed is that mobile or IoT devices often have limited bandwidth, making traditional centralized training methods inefficient and costly in terms of communication.

The proposed solution involves categorizing layers in a neural network into shallow and deep layers; parameters of deeper layers are updated less frequently than those of shallower layers, reducing the number of updates required. On the server side, models trained locally on clients are aggregated using temporally weighted aggregation to improve model accuracy and convergence without requiring frequent communication.

The experimental results show significant improvements in both communication cost and model accuracy compared to baseline algorithms. This work is particularly important for applications involving mobile or IoT devices where bandwidth constraints are a major issue. It provides an effective way to train deep learning models while preserving data privacy on the client side, making it highly relevant for real-world scenarios.

📄 Full Paper Content (ArXiv Source)

[^1]: This work is supported by the National Natural Science Foundation of China with Grant No.61473298 and 61876184. (*Corresponding author: Yaochu Jin*)

📊 논문 시각자료 (Figures)

Figure 1



Figure 2



Figure 3



Figure 4



Figure 5



Figure 6



Figure 7



Figure 8



Figure 9



Figure 10



Figure 11



Figure 12



Figure 13



Figure 14



Figure 15



Figure 16



Figure 17



Figure 18



Figure 19



Figure 20



Figure 21



A Note of Gratitude

The copyright of this content belongs to the respective researchers. We deeply appreciate their hard work and contribution to the advancement of human civilization.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut