TY - GEN
T1 - Distributed deep neural network training on edge devices
AU - Benditkis, Daniel
AU - Avidor, Tomer
AU - Keren, Aviv
AU - Shoham, Neta
AU - Mor-Yosef, Liron
AU - Tal-Israel, Nadav
N1 - Publisher Copyright:
© 2019 Copyright is held by the owner/author(s).
PY - 2019/11/7
Y1 - 2019/11/7
N2 - Deep Neural Network (Deep Learning) models have been traditionally trained on dedicated servers, after collecting data from various edge devices and sending them to the server. In recent years new methodologies have emerged for training models in a distributed manner over edge devices, keeping the data on the devices themselves. This allows for better data privacy and reduces the training costs. One of the main challenges for such methodologies is reducing the communication costs to and mainly from the edge devices. In this work we compare the two main methodologies used for distributed edge training: Federated Learning and Large Batch Training. For each of the methodologies we examine their convergence rates, communication costs, and final model performance. In addition, we present two techniques for compressing the communication between the edge devices, and examine their suitability for each one of the training methodologies.
AB - Deep Neural Network (Deep Learning) models have been traditionally trained on dedicated servers, after collecting data from various edge devices and sending them to the server. In recent years new methodologies have emerged for training models in a distributed manner over edge devices, keeping the data on the devices themselves. This allows for better data privacy and reduces the training costs. One of the main challenges for such methodologies is reducing the communication costs to and mainly from the edge devices. In this work we compare the two main methodologies used for distributed edge training: Federated Learning and Large Batch Training. For each of the methodologies we examine their convergence rates, communication costs, and final model performance. In addition, we present two techniques for compressing the communication between the edge devices, and examine their suitability for each one of the training methodologies.
KW - Communication compression
KW - Deep learning
KW - Edge device
KW - Federated learning
KW - Large batch
KW - Neural network
UR - http://www.scopus.com/inward/record.url?scp=85076258134&partnerID=8YFLogxK
U2 - 10.1145/3318216.3363324
DO - 10.1145/3318216.3363324
M3 - ???researchoutput.researchoutputtypes.contributiontobookanthology.conference???
AN - SCOPUS:85076258134
T3 - Proceedings of the 4th ACM/IEEE Symposium on Edge Computing, SEC 2019
SP - 304
EP - 306
BT - Proceedings of the 4th ACM/IEEE Symposium on Edge Computing, SEC 2019
PB - Association for Computing Machinery, Inc
T2 - 4th ACM/IEEE Symposium on Edge Computing, SEC 2019
Y2 - 7 November 2019 through 9 November 2019
ER -